helix

helix

♾️ Helix is a private GenAI stack for building AI agents with declarative pipelines, knowledge (RAG), API bindings, and first-class testing.

Stars: 715

Visit
 screenshot

HelixML is a private GenAI platform that allows users to deploy the best of open AI in their own data center or VPC while retaining complete data security and control. It includes support for fine-tuning models with drag-and-drop functionality. HelixML brings the best of open source AI to businesses in an ergonomic and scalable way, optimizing the tradeoff between GPU memory and latency.

README:

logo

SaaSPrivate DeploymentDocsDiscord

HelixML - AI Agents on a Private GenAI Stack

👥 Discord

Deploy AI agents in your own data center or VPC and retain complete data security & control.

HelixML is an enterprise-grade platform for building and deploying AI agents with support for RAG (Retrieval-Augmented Generation), API calling, vision, and multi-provider LLM support. Build and deploy LLM applications by writing a simple helix.yaml configuration file.

Our intelligent GPU scheduler packs models efficiently into available GPU memory and dynamically loads and unloads models based on demand, optimizing resource utilization.

✨ Key Features

🤖 AI Agents

  • Easy-to-use Web UI for agent interaction and management
  • Session-based architecture with pause/resume capabilities
  • Multi-step reasoning with tool orchestration
  • Memory management for context-aware interactions
  • Support for multiple LLM providers (OpenAI, Anthropic, and local models)

AI Agents Interface

🛠️ Skills and Tools

  • REST API integration with OpenAPI schema support
  • MCP (Model Context Protocol) server compatibility
  • GPTScript integration for advanced scripting
  • OAuth token management for secure third-party access
  • Custom tool development with flexible SDK

Skills and Tools

📚 Knowledge Management

  • Built-in document ingestion (PDFs, Word, text files)
  • Web scraper for automatic content extraction
  • Multiple RAG backends: Typesense, Haystack, PGVector, LlamaIndex
  • Vector embeddings with PGVector for semantic search
  • Vision RAG support for multimodal content

Knowledge Base

Main use cases:

  • Upload and analyze corporate documents
  • Add website documentation URLs to create instant customer support agents
  • Build knowledge bases from multiple sources

🔍 Tracing and Observability

Context is everything. Agents can process tens of thousands of tokens per step—Helix provides complete visibility under the hood:

Tracing Interface

Tracing features:

  • View all agent execution steps
  • Inspect requests and responses to LLM providers, third-party APIs, and MCP servers
  • Real-time token usage tracking
  • Pricing and cost analysis
  • Performance metrics and debugging

🚀 Additional Features

  • Multi-tenancy with organization, team, and role-based access control
  • Scheduled tasks and cron jobs
  • Webhook triggers for event-driven workflows
  • Evaluation framework for testing and quality assurance
  • Payment integration with Stripe support
  • Notifications via Slack, Discord, and email
  • Keycloak authentication with OAuth and OIDC support

🏗️ Architecture

HelixML uses a microservices architecture with the following components:

┌─────────────────────────────────────────────────────────┐
│                      Frontend (React)                    │
│                     vite + TypeScript                    │
└────────────────────┬────────────────────────────────────┘
                     │
┌────────────────────▼────────────────────────────────────┐
│                  API / Control Plane (Go)                │
│  ┌──────────────┬──────────────┬──────────────────────┐ │
│  │   Agents     │  Knowledge   │   Auth & Sessions    │ │
│  │   Skills     │  RAG Pipeline│   Organizations      │ │
│  │   Tools      │  Vector DB   │   Usage Tracking     │ │
│  └──────────────┴──────────────┴──────────────────────┘ │
└─────────┬──────────────────────────────────┬───────────┘
          │                                  │
┌─────────▼──────────┐            ┌─────────▼──────────┐
│   PostgreSQL       │            │   GPU Runners      │
│   + PGVector       │            │   Model Scheduler  │
└────────────────────┘            └────────────────────┘
          │
┌─────────▼──────────────────────────────────────────────┐
│  Supporting Services: Keycloak, Typesense, Haystack,   │
│  GPTScript Runner, Chrome/Rod, Tika, SearXNG           │
└────────────────────────────────────────────────────────┘

Three-layer agent hierarchy:

  1. Session: Manages agent lifecycle and state
  2. Agent: Coordinates skills and handles LLM interactions
  3. Skills: Group related tools for specific capabilities
  4. Tools: Individual actions (API calls, functions, scripts)

💻 Tech Stack

Backend

  • Go 1.24.0 - Main backend language
  • PostgreSQL + PGVector - Data storage and vector embeddings
  • GORM - ORM for database operations
  • Gorilla Mux - HTTP routing
  • Keycloak - Identity and access management
  • NATS - Message queue
  • Zerolog - Structured logging

Frontend

  • React 18.3.1 - UI framework
  • TypeScript - Type-safe JavaScript
  • Material-UI (MUI) - Component library
  • MobX - State management
  • Vite - Build tool
  • Monaco Editor - Code editing

AI/ML

  • OpenAI SDK - GPT models integration
  • Anthropic SDK - Claude models integration
  • LangChain Go - LLM orchestration
  • GPTScript - Scripting capabilities
  • Typesense / Haystack / LlamaIndex - RAG backends

Infrastructure

  • Docker & Docker Compose - Containerization
  • Kubernetes + Helm - Orchestration
  • Flux - GitOps operator

🚀 Quick Start

Install on Docker

Use our quickstart installer:

curl -sL -O https://get.helixml.tech/install.sh
chmod +x install.sh
sudo ./install.sh

The installer will prompt you before making changes to your system. By default, the dashboard will be available on http://localhost:8080.

For setting up a deployment with a DNS name, see ./install.sh --help or read the detailed docs. We've documented easy TLS termination for you.

Next steps:

Install on Kubernetes

Use our Helm charts for production deployments:

🔧 Configuration

All server configuration is done via environment variables. You can find the complete list of configuration options in api/pkg/config/config.go.

Key environment variables:

  • OPENAI_API_KEY - OpenAI API credentials
  • ANTHROPIC_API_KEY - Anthropic API credentials
  • POSTGRES_* - Database connection settings
  • KEYCLOAK_* - Authentication settings
  • SERVER_URL - Public URL for the deployment
  • RUNNER_* - GPU runner configuration

See the configuration documentation for detailed setup instructions.

👨‍💻 Development

For local development, refer to the Helix local development guide.

Prerequisites:

  • Docker Desktop (or Docker + Docker Compose)
  • Go 1.24.0+
  • Node.js 18+
  • Make

Quick development setup:

# Clone the repository
git clone https://github.com/helixml/helix.git
cd helix

# Start supporting services
docker-compose up -d postgres keycloak

# Run the backend
cd api
go run . serve

# Run the frontend (in a new terminal)
cd frontend
npm install
npm run dev

See local-development.md for comprehensive setup instructions.

📖 Documentation

🤝 Contributing

We welcome contributions! Please see our Contributing Guide for details.

By contributing, you confirm that:

  • Your changes will fall under the same license
  • Your changes will be owned by HelixML, Inc.

📄 License

Helix is licensed under a similar license to Docker Desktop. You can run the source code (in this repo) for free for:

  • Personal Use: Individuals or people personally experimenting
  • Educational Use: Schools and universities
  • Small Business Use: Companies with under $10M annual revenue and less than 250 employees

If you fall outside of these terms, please use the Launchpad to purchase a license for large commercial use. Trial licenses are available for experimentation.

You are not allowed to use our code to build a product that competes with us.

Why these license clauses?

  • We generate revenue to support the development of Helix. We are an independent software company.
  • We don't want cloud providers to take our open source code and build a rebranded service on top of it.

If you would like to use some part of this code under a more permissive license, please get in touch.

🆘 Support

🌟 Star History

If you find Helix useful, please consider giving us a star on GitHub!


Built with ❤️ by HelixML, Inc.

For Tasks:

Click tags to check more tools for each tasks

For Jobs:

Alternative AI tools for helix

Similar Open Source Tools

For similar tasks

For similar jobs