Skip to content

Developer Documentation

CI codecov Build Docker image Docs

Nevron Header

Learn how to build your first Nevron, an autonomous AI agent in Python.

This framework is designed to be a modular and extensible framework for building autonomous AI agents, which can perform tasks independently on their own.

Follow this documentation to learn how to create your first Nevron, build your own tools & workflows, and integrate with external services.

Documentation Versions

  • stable: Latest stable release (from the main branch)
  • dev: Development version (from the dev branch)

You can switch between versions using the version selector in the top navigation bar.

Overview

Nevron is an open-source framework that supports the development, deployment, and management of autonomous AI agents.

This framework is built on top of:

  • Python programming language
  • State-of-the-art LLM-powered intelligence
  • Modular architecture with planning, feedback, execution and memory components
  • Integration with external services (Telegram, Twitter, Discord, etc.)
  • Vector-based memory storage using Chroma or Qdrant
  • Local LLM support with Ollama

Core Features

  • Autonomous Decision Making: Nevron uses Q-learning algorithm for intelligent decision making
  • LLM Integration: Powered by a wide range of Large Language Models (e.g., OpenAI, Anthropic, xAI, DeepSeek, etc.)
  • Local LLM Support: Run models locally with Ollama integration
  • Modular Workflows: Predefined autonomous agent task execution patterns
    • Analyze signal workflow
    • Research news workflow
  • Memory Management: Vector storage for context retention
    • ChromaDB (default)
    • Qdrant (alternative)
  • External Integrations: Comprehensive set of tools for various platforms and services
  • Docker Deployment: Easy deployment with Docker Compose

Nevron vs. Other Frameworks

Nevron vs. Other Frameworks


Core Components

1. Planning Module

Handles decision-making using LLM of choice with the context of previous actions and their outcomes.

2. Memory Module

Manages agent's memory using vector storage for efficient context retrieval, which enables the agent to remember and recall previous interactions and events.

  • Multiple Vector Databases Support

  • Features

    • Vector embeddings via OpenAI's text-embedding-3-small model
    • Semantic similarity search
    • Metadata storage for context
    • Configurable collection management
  • Backend

    • Abstract memory backend interface
    • Modular backend architecture for optimal performance & customization
    • Async storage and retrieval operations

More about memory module can be found in the Memory section.

3. Feedback Module

Feedback module is responsible for processing action results and updating the context for future decisions in Planning Module.

  • Functions

    • Collects feedback from action execution
    • Evaluates action outcomes
    • Maintains feedback history and context
  • Integration

    • Direct integration with Planning Module
    • Performance metrics tracking

More about feedback module can be found in the Planning section.

4. Tools & Workflows

Nevron supports integrations with external services and APIs for extended functionality & integrations in different platforms.

For development purposes, Nevron comes with a comprehensive set of tools that can be used as a starting point for building your own tools and integrating more complex functionality to your AI agent.

Workflows

Nevron allows you to define custom logic of different tools into reusable workflows, which the agent can use.

Nevron comes with two pre-configured workflows:

  • Analyze signal: Processes and analyzes incoming signal data
  • Research news: Gathers and analyzes news using Perplexity API

Available Tools

Nevron includes a variety of tools for different purposes:

  • X (Twitter)

    • Post tweets
    • Media handling
    • Thread creation
  • Discord

    • Listen to incoming messages
    • Send messages to channels
  • Telegram

    • Send telegram messages
    • Bot integration
    • Channel/group support
    • HTML message formatting
  • WhatsApp

    • Get messages
    • Post messages
  • Tavily

    • Semantic search
    • Web search capabilities
  • Perplexity

    • Advanced search functionality
    • Research capabilities

And many more tools for various platforms and services. For a complete list, refer to the Tools section.

5. LLM Integration

Powers the agent's intelligence and natural language capabilities.

  • Supported Providers
    • OpenAI
    • gpt-4o for decision making
    • text-embedding-3-small for embeddings
    • Anthropic (Claude models)
    • xAI (Grok models)
    • DeepSeek
    • Qwen
    • Venice
    • Llama
    • Via API (api.llama-api.com)
    • Via OpenRouter
    • Via Fireworks
    • Locally with Ollama

More about LLM integration can be found in the LLM section.

6. Deployment Options

  • Docker Compose
  • Complete stack deployment
  • Service orchestration
  • Volume management
  • Network isolation

  • Local Development

  • Python 3.13 with Poetry
  • Customizable configuration
  • Easy debugging and extension

More about deployment can be found in the Deployment section.


Getting Started

For setup and development instructions, please refer to our Quickstart and set up your first AI agent in minutes.


Creators

Nevron was created by the Neurobro team. If you want to learn more about Nevron, our core values & the team behind it, please visit the About page.

Support by upvoting Nevron on Product Hunt.

Nevron - Open source AI agent framework in Python | Product Hunt

If you have any questions or need further assistance, please refer to the GitHub Discussions.