tunacode

tunacode

🍣 TunaCode AI CLI coding agent with safe git branches, rich tools & multi-LLM support.

Stars: 81

Visit
 screenshot

TunaCode CLI is an AI-powered coding assistant that provides a command-line interface for developers to enhance their coding experience. It offers features like model selection, parallel execution for faster file operations, and various commands for code management. The tool aims to improve coding efficiency and provide a seamless coding environment for developers.

README:

TunaCode CLI

PyPI version Downloads Python 3.10+ License: MIT

AI-powered CLI coding assistant

TunaCode Example


Quick Install

# Option 1: One-line install (Linux/macOS)
wget -qO- https://raw.githubusercontent.com/alchemiststudiosDOTai/tunacode/master/scripts/install_linux.sh | bash

# Option 2: pip install
pip install tunacode-cli

For detailed installation and configuration instructions, see the Getting Started Guide.

Quickstart

# 1) Install
pip install tunacode-cli

# 2) Launch the CLI
tunacode --wizard   # guided setup (enter an API key, pick a model)

# 3) Try common commands in the REPL
/help        # see commands
/model       # explore models and set a default
/plan        # enter read-only Plan Mode

Tip: You can also skip the wizard and set everything via flags:

tunacode --model openai:gpt-4.1 --key sk-your-key

Development Installation

For contributors and developers who want to work on TunaCode:

# Clone the repository
git clone https://github.com/alchemiststudiosDOTai/tunacode.git
cd tunacode

# Quick setup (recommended)
./scripts/setup_dev_env.sh

# Or manual setup with UV (recommended)
uv venv
source .venv/bin/activate  # On Windows: .venv\Scripts\activate
uv pip install -e ".[dev]"

# Verify installation
tunacode --version

See the Hatch Build System Guide for detailed instructions on the development environment.

Configuration

Choose your AI provider and set your API key. For more details, see the Configuration Section in the Getting Started Guide. For local models (LM Studio, Ollama, etc.), see the Local Models Setup Guide.

New: Enhanced Model Selection

TunaCode now automatically saves your model selection for future sessions. When you choose a model using /model <provider:name>, it will be remembered across restarts.

If you encounter API key errors, you can manually create a configuration file that matches the current schema:

# Create the config file
cat > ~/.config/tunacode.json << 'EOF'
{
  "default_model": "openai:gpt-4.1",
  "env": {
    "OPENAI_API_KEY": "your-openai-api-key-here",
    "ANTHROPIC_API_KEY": "",
    "GEMINI_API_KEY": "",
    "OPENROUTER_API_KEY": ""
  },
  "settings": {
    "enable_streaming": true,
    "max_iterations": 40,
    "context_window_size": 200000
  },
  "mcpServers": {}
}
EOF

Replace the model and API key with your preferred provider and credentials. Examples:

  • openai:gpt-4.1 (requires OPENAI_API_KEY)
  • anthropic:claude-4-sonnet-20250522 (requires ANTHROPIC_API_KEY)
  • google:gemini-2.5-pro (requires GEMINI_API_KEY)

⚠️ Important Notice

I apologize for any recent issues with model selection and configuration. I'm actively working to fix these problems and improve the overall stability of TunaCode. Your patience and feedback are greatly appreciated as I work to make the tool more reliable.

Recommended Models

Based on extensive testing, these models provide the best performance:

  • google/gemini-2.5-pro - Excellent for complex reasoning
  • openai/gpt-4.1 - Strong general-purpose model
  • deepseek/deepseek-r1-0528 - Great for code generation
  • openai/gpt-4.1-mini - Fast and cost-effective
  • anthropic/claude-4-sonnet-20250522 - Superior context handling

Note: Formal evaluations coming soon. Any model can work, but these have shown the best results in practice.

Start Coding

tunacode

Basic Commands

Command Description
/help Show all commands
/model <provider:name> Switch model
/clear Clear message history
/compact Summarize conversation
/branch <name> Create Git branch
/yolo Skip confirmations
!<command> Run shell command
exit Exit TunaCode

Performance

TunaCode leverages parallel execution for read-only operations, achieving 3x faster file operations:

Parallel Execution Performance

Multiple file reads, directory listings, and searches execute concurrently using async I/O, making code exploration significantly faster.

Features in Development

  • Bug Fixes: Actively addressing issues - please report any bugs you encounter!

Note: While the tool is fully functional, we're focusing on stability and core features before optimizing for speed.

Safety First

⚠️ Important: TunaCode can modify your codebase. Always:

  • Use Git branches before making changes
  • Review file modifications before confirming
  • Keep backups of important work

Documentation

For a complete overview of the documentation, see the Documentation Hub.

User Documentation

  • Getting Started - How to install, configure, and use TunaCode.
  • Commands - A complete list of all available commands.

Developer Documentation

  • Architecture (planned) - The overall architecture of the TunaCode application.
  • Contributing (planned) - Guidelines for contributing to the project.
  • Tools (planned) - How to create and use custom tools.
  • Testing (planned) - Information on the testing philosophy and how to run tests.

Guides

Reference

  • Changelog (planned) - A history of changes to the application.
  • Roadmap (planned) - The future direction of the project.
  • Security (planned) - Information about the security of the application.

Links


MIT License - see LICENSE file

For Tasks:

Click tags to check more tools for each tasks

For Jobs:

Alternative AI tools for tunacode

Similar Open Source Tools

For similar tasks

For similar jobs