
Sage
Multi-Agent System Framework For Complex Tasks
Stars: 598

Sage is a production-ready, modular, and intelligent multi-agent orchestration framework for complex problem solving. It intelligently breaks down complex tasks into manageable subtasks through seamless agent collaboration. Sage provides Deep Research Mode for comprehensive analysis and Rapid Execution Mode for quick task completion. It offers features like intelligent task decomposition, agent orchestration, extensible tool system, dual execution modes, interactive web interface, advanced token tracking, rich configuration, developer-friendly APIs, and robust error recovery mechanisms. Sage supports custom workflows, multi-agent collaboration, custom agent development, agent flow orchestration, rule preferences system, message manager for smart token optimization, task manager for comprehensive state management, advanced file system operations, advanced tool system with plugin architecture, token usage & cost monitoring, and rich configuration system. It also includes real-time streaming & monitoring, advanced tool development, error handling & reliability, performance monitoring, MCP server integration, and security features.
README:
๐ A production-ready, modular, and intelligent multi-agent orchestration framework for complex problem solving
Sage is an advanced multi-agent system that intelligently breaks down complex tasks into manageable subtasks through seamless agent collaboration. Built with enterprise-grade reliability and extensibility in mind, it provides Deep Research Mode for comprehensive analysis and Rapid Execution Mode for quick task completion.
๐ง Intelligent Task Decomposition - Automatically breaks complex problems into manageable subtasks with dependency tracking
๐ Agent Orchestration - Seamless coordination between specialized agents with robust error handling
๐ ๏ธ Extensible Tool System - Plugin-based architecture with MCP server support and auto-discovery
โก Dual Execution Modes - Choose between thorough analysis or rapid execution based on your needs
๐ Interactive Web Interface - Modern React + FastAPI UI with real-time streaming visualization
๐ Advanced Token Tracking - Comprehensive usage statistics and cost monitoring across all agents
โ๏ธ Rich Configuration - Environment variables, config files, CLI options, and runtime updates
๐ง Developer Friendly - Clean APIs, comprehensive docs, examples, and extensive error handling
๐ฏ Production Ready - Robust error recovery, logging, retry mechanisms, and performance optimization
๐ Model | ๐ง API Identifier | ๐ Key Strengths | ๐ฏ Best Use Cases |
---|---|---|---|
๐ฅ DeepSeek-V3 | deepseek-chat |
Excellent complex reasoning | Deep analysis, Code generation |
๐ Qwen-3 |
qwen-turbo , qwen-plus
|
Outstanding bilingual capabilities | Multilingual tasks, Text processing |
๐ง GPT-4.1 |
gpt-4-turbo , gpt-4o
|
Premium performance for all tasks | Enterprise apps, Complex reasoning |
โก Claude-3.5 Sonnet | claude-3-5-sonnet-20241022 |
Exceptional reasoning abilities | Creative writing, Logic analysis |
๐ข Provider | ๐ Integration | ๐ Supported Models |
---|---|---|
OpenAI | Direct API | All GPT models |
OpenRouter | Unified API | 200+ models access |
Anthropic | Native support | Claude family |
Google AI | Official API | Gemini series |
DeepSeek | Native API | All DeepSeek models |
Alibaba Cloud | Direct integration | Qwen series |
Mistral AI | Full support | All Mistral models |
๐ก Note: While Sage is optimized for the models listed above, it's designed to work with any OpenAI-compatible API endpoint.
graph LR
U[User Input] --> AC[Agent Controller]
AC --> WF
AC --> RM
subgraph WF[Workflow]
A[Analysis Agent] --> B[Planning Agent] --> C[Execution Agent] --> D[Observation Agent] --> E[Summary Agent]
D -- "if not complete" --> B
C -- uses --> X[๐ ๏ธ Tool System]
end
E --> R[Result Display]
subgraph RM["Resource & State Management"]
F[TaskManager]
G[MessageManager]
H[Workspace]
end
Note: All workflow agents read/write state & context from Resource & State Management (right).
git clone https://github.com/ZHangZHengEric/Sage.git
cd Sage
# Install core dependencies
pip install -r requirements.txt
# Install dependencies for the FastAPI React demo
pip install -r examples/fastapi_react_demo/requirements.txt
Sage includes several powerful tool systems that require specific dependencies:
-
Core Framework:
openai
,pydantic
,python-dotenv
-
Tool System:
chardet
,docstring_parser
,requests
,httpx
-
MCP Support:
mcp
,fastmcp
-
Web Interface:
fastapi
,uvicorn
,websockets
-
Demo Applications:
streamlit
,gradio
All dependencies are automatically managed by the installation script.
Experience Sage through our beautiful web interface with real-time agent visualization. Supports DeepSeek-V3, OpenRouter, and OpenAI models.
Experience Sage through our cutting-edge web application featuring a modern React frontend with FastAPI backend:
Features:
- ๐ค Multi-Agent Collaboration - Visual workflow with decomposition, planning, execution, observation, and summary
- ๐ง Deep Thinking Mode - Expandable thought bubbles showing agent reasoning process
- ๐ Custom Workflow Management - Create, edit, and manage custom workflows with visual mind-map editor
- โก Response Interruption - Stop AI responses at any time with graceful cancellation handling
- ๐ FastAPI Backend - High-performance async API server with streaming support
- โ๏ธ React Frontend - Modern responsive UI with Ant Design components
- ๐ก Real-time Communication - WebSocket + SSE dual support for live updates
- ๐จ Beautiful Interface - Collapsible deep thinking bubbles with modern design
- ๐ง Tool Management - Automatic tool discovery and management
- ๐ก Rule Preferences - Personalized AI behavior configuration with custom rules and preferences
- ๐ฑ Responsive Design - Adapts to all screen sizes
- ๐ง TypeScript Support - Full type safety throughout
Quick Start: See FastAPI React Demo README for detailed setup instructions.
๐ฏ Try the Live Demo: Experience all features immediately at Live Demo โ
Demo Features:
- ๐ฌ Interactive Chat Interface - Chat with AI agents using custom workflows
- ๐ Workflow Configuration - Create and customize workflows with visual editor
- โก Response Interruption - Click stop button to interrupt AI responses at any time
- ๐ก Rule Preferences - Configure AI behavior with custom rules and preferences
- ๐ ๏ธ System Configuration - Adjust model settings, temperature, and other parameters
- ๐ Real-time Monitoring - Watch token usage and execution progress in real-time
Access the local application at http://localhost:8080
. For detailed setup instructions, see the FastAPI React Demo README.
Sage provides a powerful command-line interface for interactive AI agent conversations:
# Basic usage
python examples/sage_cli.py --api_key YOUR_API_KEY --model deepseek/deepseek-chat --base_url https://api.deepseek.com
# With advanced options
python examples/sage_cli.py \
--api_key YOUR_API_KEY \
--model deepseek/deepseek-chat \
--base_url https://api.deepseek.com \
--max_tokens 4096 \
--temperature 0.2 \
--workspace ./workspace
CLI Features:
- ๐ค Interactive Conversations: Natural language chat with AI agents
- ๐ง Tool Integration: Built-in MCP tools for file operations, web search, etc.
- ๐ง Deep Thinking Mode: Optional detailed reasoning process
- ๐ฅ Multi-Agent Support: Complex task handling with agent collaboration
- ๐จ Beautiful Interface: Colored message frames with different visual effects
- โก Streaming Output: Real-time AI responses for smooth interaction
๐ For detailed CLI usage, configuration, and examples, see Examples README
- Task Analysis Agent: Enhanced deep understanding with context awareness and unified system prompt management
- Task Decompose Agent: Intelligent task breakdown with dependency analysis, parallel execution planning, and TaskManager integration
- Planning Agent: Strategic decomposition with dependency management, optimal tool selection, and MessageManager optimization
- Executor Agent: Intelligent tool execution with error recovery, retry mechanisms, parallel processing, and result management
- Observation Agent: Advanced progress monitoring with completion detection, quality assessment, and TaskManager state tracking
- Summary Agent: Comprehensive result synthesis with structured output, actionable insights, and execution history analysis
- Task Router Agent: ๐ Intelligent task routing system that automatically directs tasks to the most suitable agents based on task type and complexity
- Task Rewrite Agent: Intelligent task reformulation and optimization for better execution
- Task Stage Summary Agent: Intermediate progress summarization and milestone tracking
- Query Suggest Agent: Smart query enhancement and suggestion generation
- Workflow Select Agent: Intelligent workflow selection and optimization
- Simple Agent: Lightweight agent for basic tasks and rapid prototyping with optimized tool handling logic
- Simple React Agent: Reactive agent with real-time response capabilities
- Common Agent: General-purpose agent for standard operations
- Message Manager: Smart message filtering and compression system for token optimization across all agents
- Task Manager: Structured task lifecycle management with state persistence and dependency tracking
- Visual Workflow Editor: Interactive drag-and-drop interface for creating custom workflows with mind-map visualization
- Predefined Templates: Ready-to-use workflows for research reports, product development, content creation, and more
- Smart Step Management: Hierarchical workflow structure with main steps and sub-steps for complex task organization
- Real-time Preview: Live visualization of workflow structure with automatic layout and connection rendering
- Workflow Stability: Deterministic execution paths with consistent results for production environments
- Template Sharing: Export/import workflow configurations and share across teams and projects
- Zoom & Pan Support: Navigate large workflows with mouse wheel zoom and drag-to-pan functionality
- Auto-fit Display: Intelligent viewport adjustment to show all workflow nodes at optimal scale
- AgentBase Framework: Abstract base class for creating custom agents with standardized interfaces
- Agent-to-Tool Conversion: Automatic conversion of agents to tool format for seamless integration
- Streaming Support: Built-in streaming capabilities for real-time agent responses
- Context Management: Unified session context and system message handling
- Plugin Architecture: Extensible plugin system for custom agent implementations
- Agent Registration: Dynamic agent discovery and registration from directories
- Sequential Execution: Define custom agent execution sequences with AgentFlow
- Session Management: Automatic session context initialization and cleanup
- Workflow Integration: Support for available_workflows parameter in agent flows
- Error Recovery: Robust error handling with session state preservation
- Interruption Support: Graceful handling of workflow interruptions
- Memory Management: Automatic cleanup to prevent memory leaks
- Personalized AI Behavior: Configure AI assistant behavior with custom rules and preferences
- Code Style Preferences: Define coding standards, naming conventions, and style guidelines
- Response Language Settings: Control language preferences and localization settings
- Detail Level Control: Adjust verbosity and explanation depth according to your needs
- Template Library: Quick-start templates for common preference patterns
- Real-time Management: Add, edit, enable/disable rules through intuitive web interface
- Context Integration: Rules automatically apply across all agent interactions
- Intelligent Filtering: Agent-specific message filtering strategies for optimal context management
- Automatic Compression: Smart message compression reducing token usage by 30-70%
- Session Isolation: Independent message managers per session preventing cross-contamination
- Agent-Specific Strategies: Customized filtering for each agent type (TaskDecompose, Planning, Executor, etc.)
- Real-time Statistics: Live compression metrics and optimization tracking
- State Persistence: Automatic saving and restoration of message manager state
- Task Lifecycle Management: Complete task creation, execution, and completion tracking
- Dependency Tracking: Smart dependency resolution and execution ordering
- State Persistence: Automatic task state saving to workspace files
- Progress Monitoring: Real-time task progress and completion status
- Session Integration: Seamless integration with AgentController for workflow management
- Structured Data: Rich task objects with metadata, timing, and result storage
- Smart Content Search: Multi-keyword search with context extraction and relevance scoring
- Encoding Detection: Automatic character encoding detection for international files
- Security Validation: Path traversal protection and dangerous file detection
- Metadata Extraction: Comprehensive file information including size, permissions, and timestamps
- Range Reading: Efficient partial file reading with line-based navigation
- Error Recovery: Robust error handling with detailed diagnostic information
- Plugin Architecture: Hot-reloadable tool development with automatic registration and versioning
- MCP Server Support: Seamless integration with Model Context Protocol servers and remote APIs, with added API key authentication for SSE MCP server connections
- Built-in MCP Servers: Pre-built servers for file operations, parsing, command execution, and web search
- Auto-Discovery: Intelligent tool detection from directories, modules, and remote endpoints
- Type Safety: Comprehensive parameter validation with schema enforcement and runtime checks
- Error Handling: Robust error recovery, timeout management, retry strategies, and detailed logging
- Performance Monitoring: Tool execution time tracking, bottleneck detection, and optimization suggestions
- Security Features: Path validation, dangerous file detection, and protected directory access control
- Real-time Tracking: Monitor token consumption across all agents and operations with MessageManager optimization
- Detailed Analytics: Input, output, cached, and reasoning token breakdown with compression statistics
- Cost Estimation: Calculate costs based on model pricing and usage patterns with savings tracking
- Performance Metrics: Track execution time, success rates, efficiency, and token reduction rates
- Smart Optimization: Automatic message filtering and compression reducing token usage by 30-70%
- Export Capabilities: CSV, JSON export for further analysis including optimization metrics
-
Web Interface: Configure rules through the modern React interface at
/rules
- Runtime Application: Rules automatically apply to all agent interactions
- Template System: Quick-start with predefined rule templates
- Export/Import: Share rule configurations across environments
-
Environment Variables:
SAGE_DEBUG
,OPENAI_API_KEY
,SAGE_MAX_LOOP_COUNT
, etc. - Config Files: YAML/JSON configuration with validation and hot-reload
- Runtime Updates: Dynamic configuration changes without restart
- CLI Options: Comprehensive command-line interface with help system
- Profile Management: Save and load configuration profiles
- Enable comprehensive task analysis and detailed decomposition
- Generate detailed summary with insights
- Full multi-agent pipeline execution
- Enable task analysis
- Generate summary
- Skip detailed decomposition phase
- Skip analysis phase
- Direct execution
- Minimize processing time
Watch your agents work in real-time with detailed progress tracking and performance metrics, supporting real-time statistics and monitoring capabilities.
Create sophisticated custom tools with full framework integration, including caching, validation, error handling, and advanced features.
Sage includes comprehensive error handling and recovery mechanisms with automatic retry, exponential backoff, and exception management.
Monitor and optimize your agent performance with detailed tracking, statistics analysis, and bottleneck identification.
Seamlessly integrate with Model Context Protocol servers, supporting automatic tool discovery and remote API calls.
Sage includes several production-ready MCP servers:
- Smart File Operations: Advanced file reading with line range support and encoding detection
- Security Controls: Path validation, dangerous file detection, and protected directory access
- Cloud Integration: Optional cloud upload capabilities
- Batch Processing: Multi-file operations with error handling
- Multi-Format Support: 20+ file formats including PDF, Word, Excel, PowerPoint, HTML, and more
- Intelligent Extraction: Smart text extraction with metadata preservation
- Web Content: URL parsing and HTML content extraction
- Batch Processing: Multiple file parsing with performance optimization
- Secure Execution: Safe command execution with timeout management
- Cross-Platform: Windows, macOS, and Linux support
- Error Handling: Comprehensive error capture and reporting
- Security Features: Command validation and execution sandboxing
- Serper Integration: High-quality web search results
- Content Extraction: Automatic content parsing from search results
- Rate Limiting: Built-in request throttling
- Result Formatting: Clean, structured search output
Sage supports three MCP connection types:
{
"mcpServers": {
"file_system": {
"command": "python",
"args": ["./mcp_servers/file_system/file_system.py"],
"connection_type": "stdio"
}
}
}
{
"mcpServers": {
"file_parser": {
"sse_url": "http://127.0.0.1:34001/sse",
"api_key": "your-api-key"
}
}
}
{
"mcpServers": {
"web_service": {
"streamable_http_url": "http://api.example.com/mcp",
"api_key": "your-api-key"
}
}
}
Connection Type | Use Case | Advantages | Best For |
---|---|---|---|
STDIO | Local processes | Low latency, secure | Development, local tools |
SSE | Remote servers | Real-time streaming | Cloud services, live data |
Streamable HTTP | Web APIs | HTTP compatibility | REST APIs, microservices |
- API Key Authentication: Secure access control for remote MCP servers
- Connection Validation: Automatic health checks and connection monitoring
- Error Recovery: Robust reconnection and failover mechanisms
- Rate Limiting: Built-in request throttling and quota management
from sagents.agent.agent_base import AgentBase
from sagents.context.session_context import SessionContext
from sagents.context.messages.message import MessageChunk, MessageRole, MessageType
class CustomResearchAgent(AgentBase):
"""Custom agent for specialized research tasks"""
def __init__(self, model, model_config):
super().__init__(model, model_config, system_prefix="Research Agent")
self.agent_description = "Specialized agent for in-depth research and analysis"
def run_stream(self, session_context: SessionContext, tool_manager=None, session_id=None):
"""Implement custom agent logic"""
# Access conversation history
messages = session_context.message_manager.get_messages_for_llm()
# Custom research logic here
research_prompt = "Conduct thorough research on the given topic..."
# Stream responses
for chunk in self._call_llm_streaming(
messages + [{"role": "user", "content": research_prompt}],
session_id=session_id,
step_name="research_analysis"
):
yield [chunk]
from sagents.agent_flow import AgentFlow
from sagents.agent.task_analysis_agent import TaskAnalysisAgent
from sagents.agent.task_planning_agent import PlanningAgent
from sagents.agent.task_executor_agent import ExecutorAgent
# Define custom agent sequence
custom_agents = [
TaskAnalysisAgent(model, model_config),
CustomResearchAgent(model, model_config),
PlanningAgent(model, model_config),
ExecutorAgent(model, model_config)
]
# Create agent flow
agent_flow = AgentFlow(custom_agents, workspace="./workspace")
# Execute with streaming
for message_chunks in agent_flow.run_stream(
input_messages=messages,
tool_manager=tool_manager,
session_id="custom-session",
system_context={
"project_type": "research",
"domain": "AI/ML"
}
):
# Process streaming results
for chunk in message_chunks:
print(f"{chunk.role}: {chunk.content}")
# Convert agent to tool for use in other workflows
research_tool = CustomResearchAgent(model, model_config).to_tool()
# Register with tool manager
tool_manager.register_tool(research_tool)
# Now available as a tool in other agent workflows
result = tool_manager.run_tool(
"CustomResearchAgent",
messages=messages,
session_id=session_id
)
- Quick Start Guide - Get up and running in 5 minutes
- Architecture Overview - Detailed system design
- API Reference - Complete API documentation
- Tool Development - Create custom tools
- Configuration Guide - Advanced configuration options
- Examples - Real-world usage examples
Sage is production-ready with enterprise features, supporting configuration management, logging, and monitoring capabilities.
Create, edit, and visualize custom workflows with our interactive mind-map editor, supporting automatic workflow selection and intelligent execution.
Visual Editor Features:
- ๐จ Mind-map visualization with hierarchical node layout
- ๐ฑ๏ธ Interactive editing - click to edit nodes directly
- ๐ Zoom & Pan - navigate large workflows with mouse controls
- ๐ Auto-fit display - intelligent viewport adjustment
- ๐พ Template system - save and reuse workflow configurations
Stop AI responses at any time with graceful cancellation and resource cleanup, with web interface support for stopping responses via button click.
Interruption Features:
- ๐ Immediate stopping - responses halt within 1-2 seconds
- ๐งน Resource cleanup - proper memory and connection management
- ๐พ State preservation - partial results are saved and accessible
- ๐ Resumable execution - continue from interruption point if needed
- Task Router Agent: ๐ Intelligent task routing system that automatically directs tasks to the most suitable agents based on task type and complexity
- Unified Tool Interface: Standardized tool calling interface using session_context instead of messages parameter for better consistency
- Enhanced Workflow Display: Improved workflow step visualization with detailed step descriptions and progress indicators
- Optimized Simple Agent: Enhanced tool handling logic that returns directly when tool count is minimal for better performance
- Simplified Configuration: Updated .gitignore with streamlined pycache configuration for cleaner project structure
- Advanced File Search: Enhanced file content search with multi-keyword support, context extraction, and relevance scoring
- Built-in MCP Servers: Four production-ready MCP servers for file operations, parsing, command execution, and web search
- Triple MCP Connection Support: STDIO, SSE, and Streamable HTTP connection types with API key authentication
- Extended Agent Ecosystem: 14 specialized agents including Task Router, Task Rewrite, Query Suggest, Workflow Select, and more
- Custom Agent Development: AgentBase framework for creating specialized agents with standardized interfaces
- Agent Flow Orchestration: Sequential agent execution with AgentFlow for custom workflow design
- Agent-to-Tool Conversion: Automatic conversion of agents to tools for seamless integration
- Interface Standardization: Unified tool calling patterns across all agents for better maintainability
- Performance Optimization: Improved file reading with range-based operations and metadata caching
- Error Recovery: Enhanced error handling with detailed diagnostic information and recovery strategies
- Type Safety: Comprehensive parameter validation with schema enforcement
- Memory Management: Optimized memory usage for large file operations
- Streaming Support: Real-time streaming capabilities for long-running operations
- Workflow Visualization: Enhanced step display with descriptive information and better user experience
- Tool Interface Consistency: Standardized tool calling interface across all agent types
- Workflow Step Display: Improved step description rendering and progress tracking
- Simple Agent Optimization: Fixed tool handling logic for scenarios with minimal tool requirements
- Configuration Management: Streamlined project configuration and dependency management
- Framework Stability: Enhanced overall system reliability and error recovery
This project is licensed under the MIT License - see the LICENSE file for details.
- OpenAI for the powerful language models
- DeepSeek for the exceptional V3 model
- Alibaba Cloud for the Qwen series
- The open-source community for inspiration and tools
- All contributors who help make Sage better
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for Sage
Similar Open Source Tools

Sage
Sage is a production-ready, modular, and intelligent multi-agent orchestration framework for complex problem solving. It intelligently breaks down complex tasks into manageable subtasks through seamless agent collaboration. Sage provides Deep Research Mode for comprehensive analysis and Rapid Execution Mode for quick task completion. It offers features like intelligent task decomposition, agent orchestration, extensible tool system, dual execution modes, interactive web interface, advanced token tracking, rich configuration, developer-friendly APIs, and robust error recovery mechanisms. Sage supports custom workflows, multi-agent collaboration, custom agent development, agent flow orchestration, rule preferences system, message manager for smart token optimization, task manager for comprehensive state management, advanced file system operations, advanced tool system with plugin architecture, token usage & cost monitoring, and rich configuration system. It also includes real-time streaming & monitoring, advanced tool development, error handling & reliability, performance monitoring, MCP server integration, and security features.

AGiXT
AGiXT is a dynamic Artificial Intelligence Automation Platform engineered to orchestrate efficient AI instruction management and task execution across a multitude of providers. Our solution infuses adaptive memory handling with a broad spectrum of commands to enhance AI's understanding and responsiveness, leading to improved task completion. The platform's smart features, like Smart Instruct and Smart Chat, seamlessly integrate web search, planning strategies, and conversation continuity, transforming the interaction between users and AI. By leveraging a powerful plugin system that includes web browsing and command execution, AGiXT stands as a versatile bridge between AI models and users. With an expanding roster of AI providers, code evaluation capabilities, comprehensive chain management, and platform interoperability, AGiXT is consistently evolving to drive a multitude of applications, affirming its place at the forefront of AI technology.

J.A.R.V.I.S.2.0
J.A.R.V.I.S. 2.0 is an AI-powered assistant designed for voice commands, capable of tasks like providing weather reports, summarizing news, sending emails, and more. It features voice activation, speech recognition, AI responses, and handles multiple tasks including email sending, weather reports, news reading, image generation, database functions, phone call automation, AI-based task execution, website & application automation, and knowledge-based interactions. The assistant also includes timeout handling, automatic input processing, and the ability to call multiple functions simultaneously. It requires Python 3.9 or later and specific API keys for weather, news, email, and AI access. The tool integrates Gemini AI for function execution and Ollama as a fallback mechanism. It utilizes a RAG-based knowledge system and ADB integration for phone automation. Future enhancements include deeper mobile integration, advanced AI-driven automation, improved NLP-based command execution, and multi-modal interactions.

AionUi
AionUi is a user interface library for building modern and responsive web applications. It provides a set of customizable components and styles to create visually appealing user interfaces. With AionUi, developers can easily design and implement interactive web interfaces that are both functional and aesthetically pleasing. The library is built using the latest web technologies and follows best practices for performance and accessibility. Whether you are working on a personal project or a professional application, AionUi can help you streamline the UI development process and deliver a seamless user experience.

persistent-ai-memory
Persistent AI Memory System is a comprehensive tool that offers persistent, searchable storage for AI assistants. It includes features like conversation tracking, MCP tool call logging, and intelligent scheduling. The system supports multiple databases, provides enhanced memory management, and offers various tools for memory operations, schedule management, and system health checks. It also integrates with various platforms like LM Studio, VS Code, Koboldcpp, Ollama, and more. The system is designed to be modular, platform-agnostic, and scalable, allowing users to handle large conversation histories efficiently.

ito
Ito is an intelligent voice assistant that provides seamless voice dictation to any application on your computer. It works in any app, offers global keyboard shortcuts, real-time transcription, and instant text insertion. It is smart and adaptive with features like custom dictionary, context awareness, multi-language support, and intelligent punctuation. Users can customize trigger keys, audio preferences, and privacy controls. It also offers data management features like a notes system, interaction history, cloud sync, and export capabilities. Ito is built as a modern Electron application with a multi-process architecture and utilizes technologies like React, TypeScript, Rust, gRPC, and AWS CDK.

paiml-mcp-agent-toolkit
PAIML MCP Agent Toolkit (PMAT) is a zero-configuration AI context generation system with extreme quality enforcement and Toyota Way standards. It allows users to analyze any codebase instantly through CLI, MCP, or HTTP interfaces. The toolkit provides features such as technical debt analysis, advanced monitoring, metrics aggregation, performance profiling, bottleneck detection, alert system, multi-format export, storage flexibility, and more. It also offers AI-powered intelligence for smart recommendations, polyglot analysis, repository showcase, and integration points. PMAT enforces quality standards like complexity โค20, zero SATD comments, test coverage >80%, no lint warnings, and synchronized documentation with commits. The toolkit follows Toyota Way development principles for iterative improvement, direct AST traversal, automated quality gates, and zero SATD policy.

evi-run
evi-run is a powerful, production-ready multi-agent AI system built on Python using the OpenAI Agents SDK. It offers instant deployment, ultimate flexibility, built-in analytics, Telegram integration, and scalable architecture. The system features memory management, knowledge integration, task scheduling, multi-agent orchestration, custom agent creation, deep research, web intelligence, document processing, image generation, DEX analytics, and Solana token swap. It supports flexible usage modes like private, free, and pay mode, with upcoming features including NSFW mode, task scheduler, and automatic limit orders. The technology stack includes Python 3.11, OpenAI Agents SDK, Telegram Bot API, PostgreSQL, Redis, and Docker & Docker Compose for deployment.

pluely
Pluely is a versatile and user-friendly tool for managing tasks and projects. It provides a simple interface for creating, organizing, and tracking tasks, making it easy to stay on top of your work. With features like task prioritization, due date reminders, and collaboration options, Pluely helps individuals and teams streamline their workflow and boost productivity. Whether you're a student juggling assignments, a professional managing multiple projects, or a team coordinating tasks, Pluely is the perfect solution to keep you organized and efficient.

ToolNeuron
ToolNeuron is a secure, offline AI ecosystem for Android devices that allows users to run private AI models and dynamic plugins fully offline, with hardware-grade encryption ensuring maximum privacy. It enables users to have an offline-first experience, add capabilities without app updates through pluggable tools, and ensures security by design with strict plugin validation and sandboxing.

claude-007-agents
Claude Code Agents is an open-source AI agent system designed to enhance development workflows by providing specialized AI agents for orchestration, resilience engineering, and organizational memory. These agents offer specialized expertise across technologies, AI system with organizational memory, and an agent orchestration system. The system includes features such as engineering excellence by design, advanced orchestration system, Task Master integration, live MCP integrations, professional-grade workflows, and organizational intelligence. It is suitable for solo developers, small teams, enterprise teams, and open-source projects. The system requires a one-time bootstrap setup for each project to analyze the tech stack, select optimal agents, create configuration files, set up Task Master integration, and validate system readiness.

llamafarm
LlamaFarm is a comprehensive AI framework that empowers users to build powerful AI applications locally, with full control over costs and deployment options. It provides modular components for RAG systems, vector databases, model management, prompt engineering, and fine-tuning. Users can create differentiated AI products without needing extensive ML expertise, using simple CLI commands and YAML configs. The framework supports local-first development, production-ready components, strategy-based configuration, and deployment anywhere from laptops to the cloud.

agentneo
AgentNeo is a Python package that provides functionalities for project, trace, dataset, experiment management. It allows users to authenticate, create projects, trace agents and LangGraph graphs, manage datasets, and run experiments with metrics. The tool aims to streamline AI project management and analysis by offering a comprehensive set of features.

mcp-memory-service
The MCP Memory Service is a universal memory service designed for AI assistants, providing semantic memory search and persistent storage. It works with various AI applications and offers fast local search using SQLite-vec and global distribution through Cloudflare. The service supports intelligent memory management, universal compatibility with AI tools, flexible storage options, and is production-ready with cross-platform support and secure connections. Users can store and recall memories, search by tags, check system health, and configure the service for Claude Desktop integration and environment variables.

opcode
opcode is a powerful desktop application built with Tauri 2 that serves as a command center for interacting with Claude Code. It offers a visual GUI for managing Claude Code sessions, creating custom agents, tracking usage, and more. Users can navigate projects, create specialized AI agents, monitor usage analytics, manage MCP servers, create session checkpoints, edit CLAUDE.md files, and more. The tool bridges the gap between command-line tools and visual experiences, making AI-assisted development more intuitive and productive.

presenton
Presenton is an open-source AI presentation generator and API that allows users to create professional presentations locally on their devices. It offers complete control over the presentation workflow, including custom templates, AI template generation, flexible generation options, and export capabilities. Users can use their own API keys for various models, integrate with Ollama for local model running, and connect to OpenAI-compatible endpoints. The tool supports multiple providers for text and image generation, runs locally without cloud dependencies, and can be deployed as a Docker container with GPU support.
For similar tasks

Azure-Analytics-and-AI-Engagement
The Azure-Analytics-and-AI-Engagement repository provides packaged Industry Scenario DREAM Demos with ARM templates (Containing a demo web application, Power BI reports, Synapse resources, AML Notebooks etc.) that can be deployed in a customerโs subscription using the CAPE tool within a matter of few hours. Partners can also deploy DREAM Demos in their own subscriptions using DPoC.

sorrentum
Sorrentum is an open-source project that aims to combine open-source development, startups, and brilliant students to build machine learning, AI, and Web3 / DeFi protocols geared towards finance and economics. The project provides opportunities for internships, research assistantships, and development grants, as well as the chance to work on cutting-edge problems, learn about startups, write academic papers, and get internships and full-time positions at companies working on Sorrentum applications.

tidb
TiDB is an open-source distributed SQL database that supports Hybrid Transactional and Analytical Processing (HTAP) workloads. It is MySQL compatible and features horizontal scalability, strong consistency, and high availability.

zep-python
Zep is an open-source platform for building and deploying large language model (LLM) applications. It provides a suite of tools and services that make it easy to integrate LLMs into your applications, including chat history memory, embedding, vector search, and data enrichment. Zep is designed to be scalable, reliable, and easy to use, making it a great choice for developers who want to build LLM-powered applications quickly and easily.

telemetry-airflow
This repository codifies the Airflow cluster that is deployed at workflow.telemetry.mozilla.org (behind SSO) and commonly referred to as "WTMO" or simply "Airflow". Some links relevant to users and developers of WTMO: * The `dags` directory in this repository contains some custom DAG definitions * Many of the DAGs registered with WTMO don't live in this repository, but are instead generated from ETL task definitions in bigquery-etl * The Data SRE team maintains a WTMO Developer Guide (behind SSO)

mojo
Mojo is a new programming language that bridges the gap between research and production by combining Python syntax and ecosystem with systems programming and metaprogramming features. Mojo is still young, but it is designed to become a superset of Python over time.

pandas-ai
PandasAI is a Python library that makes it easy to ask questions to your data in natural language. It helps you to explore, clean, and analyze your data using generative AI.

databend
Databend is an open-source cloud data warehouse that serves as a cost-effective alternative to Snowflake. With its focus on fast query execution and data ingestion, it's designed for complex analysis of the world's largest datasets.
For similar jobs

weave
Weave is a toolkit for developing Generative AI applications, built by Weights & Biases. With Weave, you can log and debug language model inputs, outputs, and traces; build rigorous, apples-to-apples evaluations for language model use cases; and organize all the information generated across the LLM workflow, from experimentation to evaluations to production. Weave aims to bring rigor, best-practices, and composability to the inherently experimental process of developing Generative AI software, without introducing cognitive overhead.

LLMStack
LLMStack is a no-code platform for building generative AI agents, workflows, and chatbots. It allows users to connect their own data, internal tools, and GPT-powered models without any coding experience. LLMStack can be deployed to the cloud or on-premise and can be accessed via HTTP API or triggered from Slack or Discord.

VisionCraft
The VisionCraft API is a free API for using over 100 different AI models. From images to sound.

kaito
Kaito is an operator that automates the AI/ML inference model deployment in a Kubernetes cluster. It manages large model files using container images, avoids tuning deployment parameters to fit GPU hardware by providing preset configurations, auto-provisions GPU nodes based on model requirements, and hosts large model images in the public Microsoft Container Registry (MCR) if the license allows. Using Kaito, the workflow of onboarding large AI inference models in Kubernetes is largely simplified.

PyRIT
PyRIT is an open access automation framework designed to empower security professionals and ML engineers to red team foundation models and their applications. It automates AI Red Teaming tasks to allow operators to focus on more complicated and time-consuming tasks and can also identify security harms such as misuse (e.g., malware generation, jailbreaking), and privacy harms (e.g., identity theft). The goal is to allow researchers to have a baseline of how well their model and entire inference pipeline is doing against different harm categories and to be able to compare that baseline to future iterations of their model. This allows them to have empirical data on how well their model is doing today, and detect any degradation of performance based on future improvements.

tabby
Tabby is a self-hosted AI coding assistant, offering an open-source and on-premises alternative to GitHub Copilot. It boasts several key features: * Self-contained, with no need for a DBMS or cloud service. * OpenAPI interface, easy to integrate with existing infrastructure (e.g Cloud IDE). * Supports consumer-grade GPUs.

spear
SPEAR (Simulator for Photorealistic Embodied AI Research) is a powerful tool for training embodied agents. It features 300 unique virtual indoor environments with 2,566 unique rooms and 17,234 unique objects that can be manipulated individually. Each environment is designed by a professional artist and features detailed geometry, photorealistic materials, and a unique floor plan and object layout. SPEAR is implemented as Unreal Engine assets and provides an OpenAI Gym interface for interacting with the environments via Python.

Magick
Magick is a groundbreaking visual AIDE (Artificial Intelligence Development Environment) for no-code data pipelines and multimodal agents. Magick can connect to other services and comes with nodes and templates well-suited for intelligent agents, chatbots, complex reasoning systems and realistic characters.