
gemini-coder
The free 2M context AI coding assistant
Stars: 67

README:

Copy folders and files for chatbots or initialize them hands-free!
Use the free Gemini API for FIM completions, file refactoring and applying AI-suggested changes.
Gemini Coder lets you conveniently copy folders and files for chatbots. With the Connector browser extension you can initalize them hands-free!
The extension uses the same context for built-in API features: Fill-In-the-Middle (FIM) completions and file refactoring. Hit apply changes to integrate AI responses with your codebase with just a single click.
- 100% free & open source: MIT License
- Versitale: Not limited to Gemini or AI Studio
- Private: Does not collect any usage data
- Local: Talks with the web browser via WebSockets
- One of a kind: Lets you use any website for context
- Lightweight: Unpacked build is just ~1MB
Other AI coding tools try to "guess" what context matters, often getting it wrong. Gemini Coder works differently:
- You select which folders and files provide relevant context
- You control what examples of coding conventions to include
- You know how much tokens is used in web chats and FIM/refactoring requests
The result? Unmatched in accuracy, speed and cost AI assistance.
Too many tokens fighting for attention may decrease performance due to being too "distracting", diffusing attention too broadly and decreasing a signal to noise ratio in the features. ~Andrej Karpathy
Gemini Coder works with many popular chatbots:
- AI Studio - fully supported (model, temperature, system instructions)
- Gemini
- ChatGPT
- Claude
- GitHub Copilot
- Grok
- DeepSeek
- Mistral
- HuggingChat
- Open WebUI (localhost)
- Open the new Gemini Coder view from the activity bar (sparkles icon).
- Select files/folders for the context.
- Click copy icon from the toolbar.
- (optional) Install browser integration for hands-free initializations.
- Get your API key from Google AI Studio.
- Open VS Code and navigate to settings.
- Search for "Gemini Coder" and paste your API key.
- Use Command Palette (Ctrl/Cmd + Shift + P) and type "FIM Completion".
- Bind the command to a keyboard shortcut by opening Keyboard Shortcuts (Ctrl/Cmd+K Ctrl/Cmd+S), searching for
Gemini Coder: FIM Completion
, clicking the + icon, and pressing your preferred key combination (e.g. Ctrl/Cmd+I).
-
Gemini Coder: FIM Completion
- Get fill-in-the-middle completion using default model. -
Gemini Coder: FIM Completion with...
- Get fill-in-the-middle completion with model selection. -
Gemini Coder: FIM Completion to Clipboard
- Copy FIM completion content to clipboard. -
Gemini Coder: Change Default FIM Model
- Change default AI model for FIM completions.
-
Gemini Coder: Refactor this File
- Apply changes based on refactoring instruction. -
Gemini Coder: Refactor this File with...
- Refactor with model selection. -
Gemini Coder: Refactor to Clipboard
- Copy refactoring content to clipboard. -
Gemini Coder: Change Default Refactoring Model
- Change default AI model for refactoring.
-
Gemini Coder: Apply Changes
- Apply changes suggested by AI using clipboard content. -
Gemini Coder: Apply Changes with...
- Apply changes with model selection. -
Gemini Coder: Apply Changes to Clipboard
- Copy apply changes content to clipboard. -
Gemini Coder: Change Default Apply Changes Model
- Change default AI model for applying changes.
-
Gemini Coder: Web Chat
- Enter instructions and open web chat hands-free. -
Gemini Coder: Chat to Clipboard
- Enter instructions and copy to clipboard.
-
Gemini Coder: Copy Context
- Copy selected files as XML context.
The extension supports OpenAI-API compatible model providers for API features.
"geminiCoder.providers": [
{
"name": "DeepSeek",
"endpointUrl": "https://api.deepseek.com/v1/chat/completions",
"bearerToken": "[API KEY]",
"model": "deepseek-chat",
"temperature": 0,
"instruction": ""
},
{
"name": "Mistral Large Latest",
"endpointUrl": "https://api.mistral.ai/v1/chat/completions",
"bearerToken": "[API KEY]",
"model": "mistral-large-latest",
"temperature": 0,
"instruction": ""
},
],
All contributions are welcome. Feel free to submit pull requests or create issues and discussions.
Copyright (c) 2025 Robert Piosik. MIT License.
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for gemini-coder
Similar Open Source Tools

aider-desk
AiderDesk is a desktop application that enhances coding workflow by leveraging AI capabilities. It offers an intuitive GUI, project management, IDE integration, MCP support, settings management, cost tracking, structured messages, visual file management, model switching, code diff viewer, one-click reverts, and easy sharing. Users can install it by downloading the latest release and running the executable. AiderDesk also supports Python version detection and auto update disabling. It includes features like multiple project management, context file management, model switching, chat mode selection, question answering, cost tracking, MCP server integration, and MCP support for external tools and context. Development setup involves cloning the repository, installing dependencies, running in development mode, and building executables for different platforms. Contributions from the community are welcome following specific guidelines.

WebAI-to-API
This project implements a web API that offers a unified interface to Google Gemini and Claude 3. It provides a self-hosted, lightweight, and scalable solution for accessing these AI models through a streaming API. The API supports both Claude and Gemini models, allowing users to interact with them in real-time. The project includes a user-friendly web UI for configuration and documentation, making it easy to get started and explore the capabilities of the API.

probe
Probe is an AI-friendly, fully local, semantic code search tool designed to power the next generation of AI coding assistants. It combines the speed of ripgrep with the code-aware parsing of tree-sitter to deliver precise results with complete code blocks, making it perfect for large codebases and AI-driven development workflows. Probe is fully local, keeping code on the user's machine without relying on external APIs. It supports multiple languages, offers various search options, and can be used in CLI mode, MCP server mode, AI chat mode, and web interface. The tool is designed to be flexible, fast, and accurate, providing developers and AI models with full context and relevant code blocks for efficient code exploration and understanding.

crawl4ai
Crawl4AI is a powerful and free web crawling service that extracts valuable data from websites and provides LLM-friendly output formats. It supports crawling multiple URLs simultaneously, replaces media tags with ALT, and is completely free to use and open-source. Users can integrate Crawl4AI into Python projects as a library or run it as a standalone local server. The tool allows users to crawl and extract data from specified URLs using different providers and models, with options to include raw HTML content, force fresh crawls, and extract meaningful text blocks. Configuration settings can be adjusted in the `crawler/config.py` file to customize providers, API keys, chunk processing, and word thresholds. Contributions to Crawl4AI are welcome from the open-source community to enhance its value for AI enthusiasts and developers.

R2R
R2R (RAG to Riches) is a fast and efficient framework for serving high-quality Retrieval-Augmented Generation (RAG) to end users. The framework is designed with customizable pipelines and a feature-rich FastAPI implementation, enabling developers to quickly deploy and scale RAG-based applications. R2R was conceived to bridge the gap between local LLM experimentation and scalable production solutions. **R2R is to LangChain/LlamaIndex what NextJS is to React**. A JavaScript client for R2R deployments can be found here. ### Key Features * **🚀 Deploy** : Instantly launch production-ready RAG pipelines with streaming capabilities. * **🧩 Customize** : Tailor your pipeline with intuitive configuration files. * **🔌 Extend** : Enhance your pipeline with custom code integrations. * **⚖️ Autoscale** : Scale your pipeline effortlessly in the cloud using SciPhi. * **🤖 OSS** : Benefit from a framework developed by the open-source community, designed to simplify RAG deployment.

search_with_ai
Build your own conversation-based search with AI, a simple implementation with Node.js & Vue3. Live Demo Features: * Built-in support for LLM: OpenAI, Google, Lepton, Ollama(Free) * Built-in support for search engine: Bing, Sogou, Google, SearXNG(Free) * Customizable pretty UI interface * Support dark mode * Support mobile display * Support local LLM with Ollama * Support i18n * Support Continue Q&A with contexts.

gpustack
GPUStack is an open-source GPU cluster manager designed for running large language models (LLMs). It supports a wide variety of hardware, scales with GPU inventory, offers lightweight Python package with minimal dependencies, provides OpenAI-compatible APIs, simplifies user and API key management, enables GPU metrics monitoring, and facilitates token usage and rate metrics tracking. The tool is suitable for managing GPU clusters efficiently and effectively.

quantalogic
QuantaLogic is a ReAct framework for building advanced AI agents that seamlessly integrates large language models with a robust tool system. It aims to bridge the gap between advanced AI models and practical implementation in business processes by enabling agents to understand, reason about, and execute complex tasks through natural language interaction. The framework includes features such as ReAct Framework, Universal LLM Support, Secure Tool System, Real-time Monitoring, Memory Management, and Enterprise Ready components.

Agentarium
Agentarium is a powerful Python framework for managing and orchestrating AI agents with ease. It provides a flexible and intuitive way to create, manage, and coordinate interactions between multiple AI agents in various environments. The framework offers advanced agent management, robust interaction management, a checkpoint system for saving and restoring agent states, data generation through agent interactions, performance optimization, flexible environment configuration, and an extensible architecture for customization.

simba
Simba is an open source, portable Knowledge Management System (KMS) designed to seamlessly integrate with any Retrieval-Augmented Generation (RAG) system. It features a modern UI and modular architecture, allowing developers to focus on building advanced AI solutions without the complexities of knowledge management. Simba offers a user-friendly interface to visualize and modify document chunks, supports various vector stores and embedding models, and simplifies knowledge management for developers. It is community-driven, extensible, and aims to enhance AI functionality by providing a seamless integration with RAG-based systems.

orra
Orra is a tool for building production-ready multi-agent applications that handle complex real-world interactions. It coordinates tasks across existing stack, agents, and tools run as services using intelligent reasoning. With features like smart pre-evaluated execution plans, domain grounding, durable execution, and automatic service health monitoring, Orra enables users to go fast with tools as services and revert state to handle failures. It provides real-time status tracking and webhook result delivery, making it ideal for developers looking to move beyond simple crews and agents.