
gemini-coder
The free 2M context AI coding assistant
Stars: 67

README:

Copy folders and files for chatbots or initialize them hands-free!
Use the free Gemini API for FIM completions, file refactoring and applying AI-suggested changes.
Gemini Coder lets you conveniently copy folders and files for chatbots. With the Connector browser extension you can initalize them hands-free!
The extension uses the same context for built-in API features: Fill-In-the-Middle (FIM) completions and file refactoring. Hit apply changes to integrate AI responses with your codebase with just a single click.
- 100% free & open source: MIT License
- Versitale: Not limited to Gemini or AI Studio
- Private: Does not collect any usage data
- Local: Talks with the web browser via WebSockets
- One of a kind: Lets you use any website for context
- Lightweight: Unpacked build is just ~1MB
Other AI coding tools try to "guess" what context matters, often getting it wrong. Gemini Coder works differently:
- You select which folders and files provide relevant context
- You control what examples of coding conventions to include
- You know how much tokens is used in web chats and FIM/refactoring requests
The result? Unmatched in accuracy, speed and cost AI assistance.
Too many tokens fighting for attention may decrease performance due to being too "distracting", diffusing attention too broadly and decreasing a signal to noise ratio in the features. ~Andrej Karpathy
Gemini Coder works with many popular chatbots:
- AI Studio - fully supported (model, temperature, system instructions)
- Gemini
- ChatGPT
- Claude
- GitHub Copilot
- Grok
- DeepSeek
- Mistral
- HuggingChat
- Open WebUI (localhost)
- Open the new Gemini Coder view from the activity bar (sparkles icon).
- Select files/folders for the context.
- Click copy icon from the toolbar.
- (optional) Install browser integration for hands-free initializations.
- Get your API key from Google AI Studio.
- Open VS Code and navigate to settings.
- Search for "Gemini Coder" and paste your API key.
- Use Command Palette (Ctrl/Cmd + Shift + P) and type "FIM Completion".
- Bind the command to a keyboard shortcut by opening Keyboard Shortcuts (Ctrl/Cmd+K Ctrl/Cmd+S), searching for
Gemini Coder: FIM Completion
, clicking the + icon, and pressing your preferred key combination (e.g. Ctrl/Cmd+I).
-
Gemini Coder: FIM Completion
- Get fill-in-the-middle completion using default model. -
Gemini Coder: FIM Completion with...
- Get fill-in-the-middle completion with model selection. -
Gemini Coder: FIM Completion to Clipboard
- Copy FIM completion content to clipboard. -
Gemini Coder: Change Default FIM Model
- Change default AI model for FIM completions.
-
Gemini Coder: Refactor this File
- Apply changes based on refactoring instruction. -
Gemini Coder: Refactor this File with...
- Refactor with model selection. -
Gemini Coder: Refactor to Clipboard
- Copy refactoring content to clipboard. -
Gemini Coder: Change Default Refactoring Model
- Change default AI model for refactoring.
-
Gemini Coder: Apply Changes
- Apply changes suggested by AI using clipboard content. -
Gemini Coder: Apply Changes with...
- Apply changes with model selection. -
Gemini Coder: Apply Changes to Clipboard
- Copy apply changes content to clipboard. -
Gemini Coder: Change Default Apply Changes Model
- Change default AI model for applying changes.
-
Gemini Coder: Web Chat
- Enter instructions and open web chat hands-free. -
Gemini Coder: Chat to Clipboard
- Enter instructions and copy to clipboard.
-
Gemini Coder: Copy Context
- Copy selected files as XML context.
The extension supports OpenAI-API compatible model providers for API features.
"geminiCoder.providers": [
{
"name": "DeepSeek",
"endpointUrl": "https://api.deepseek.com/v1/chat/completions",
"bearerToken": "[API KEY]",
"model": "deepseek-chat",
"temperature": 0,
"instruction": ""
},
{
"name": "Mistral Large Latest",
"endpointUrl": "https://api.mistral.ai/v1/chat/completions",
"bearerToken": "[API KEY]",
"model": "mistral-large-latest",
"temperature": 0,
"instruction": ""
},
],
All contributions are welcome. Feel free to submit pull requests or create issues and discussions.
Copyright (c) 2025 Robert Piosik. MIT License.
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for gemini-coder
Similar Open Source Tools

aider-desk
AiderDesk is a desktop application that enhances coding workflow by leveraging AI capabilities. It offers an intuitive GUI, project management, IDE integration, MCP support, settings management, cost tracking, structured messages, visual file management, model switching, code diff viewer, one-click reverts, and easy sharing. Users can install it by downloading the latest release and running the executable. AiderDesk also supports Python version detection and auto update disabling. It includes features like multiple project management, context file management, model switching, chat mode selection, question answering, cost tracking, MCP server integration, and MCP support for external tools and context. Development setup involves cloning the repository, installing dependencies, running in development mode, and building executables for different platforms. Contributions from the community are welcome following specific guidelines.

WebAI-to-API
This project implements a web API that offers a unified interface to Google Gemini and Claude 3. It provides a self-hosted, lightweight, and scalable solution for accessing these AI models through a streaming API. The API supports both Claude and Gemini models, allowing users to interact with them in real-time. The project includes a user-friendly web UI for configuration and documentation, making it easy to get started and explore the capabilities of the API.

probe
Probe is an AI-friendly, fully local, semantic code search tool designed to power the next generation of AI coding assistants. It combines the speed of ripgrep with the code-aware parsing of tree-sitter to deliver precise results with complete code blocks, making it perfect for large codebases and AI-driven development workflows. Probe is fully local, keeping code on the user's machine without relying on external APIs. It supports multiple languages, offers various search options, and can be used in CLI mode, MCP server mode, AI chat mode, and web interface. The tool is designed to be flexible, fast, and accurate, providing developers and AI models with full context and relevant code blocks for efficient code exploration and understanding.

crawl4ai
Crawl4AI is a powerful and free web crawling service that extracts valuable data from websites and provides LLM-friendly output formats. It supports crawling multiple URLs simultaneously, replaces media tags with ALT, and is completely free to use and open-source. Users can integrate Crawl4AI into Python projects as a library or run it as a standalone local server. The tool allows users to crawl and extract data from specified URLs using different providers and models, with options to include raw HTML content, force fresh crawls, and extract meaningful text blocks. Configuration settings can be adjusted in the `crawler/config.py` file to customize providers, API keys, chunk processing, and word thresholds. Contributions to Crawl4AI are welcome from the open-source community to enhance its value for AI enthusiasts and developers.

R2R
R2R (RAG to Riches) is a fast and efficient framework for serving high-quality Retrieval-Augmented Generation (RAG) to end users. The framework is designed with customizable pipelines and a feature-rich FastAPI implementation, enabling developers to quickly deploy and scale RAG-based applications. R2R was conceived to bridge the gap between local LLM experimentation and scalable production solutions. **R2R is to LangChain/LlamaIndex what NextJS is to React**. A JavaScript client for R2R deployments can be found here. ### Key Features * **🚀 Deploy** : Instantly launch production-ready RAG pipelines with streaming capabilities. * **🧩 Customize** : Tailor your pipeline with intuitive configuration files. * **🔌 Extend** : Enhance your pipeline with custom code integrations. * **⚖️ Autoscale** : Scale your pipeline effortlessly in the cloud using SciPhi. * **🤖 OSS** : Benefit from a framework developed by the open-source community, designed to simplify RAG deployment.

search_with_ai
Build your own conversation-based search with AI, a simple implementation with Node.js & Vue3. Live Demo Features: * Built-in support for LLM: OpenAI, Google, Lepton, Ollama(Free) * Built-in support for search engine: Bing, Sogou, Google, SearXNG(Free) * Customizable pretty UI interface * Support dark mode * Support mobile display * Support local LLM with Ollama * Support i18n * Support Continue Q&A with contexts.

gpustack
GPUStack is an open-source GPU cluster manager designed for running large language models (LLMs). It supports a wide variety of hardware, scales with GPU inventory, offers lightweight Python package with minimal dependencies, provides OpenAI-compatible APIs, simplifies user and API key management, enables GPU metrics monitoring, and facilitates token usage and rate metrics tracking. The tool is suitable for managing GPU clusters efficiently and effectively.

Agentarium
Agentarium is a powerful Python framework for managing and orchestrating AI agents with ease. It provides a flexible and intuitive way to create, manage, and coordinate interactions between multiple AI agents in various environments. The framework offers advanced agent management, robust interaction management, a checkpoint system for saving and restoring agent states, data generation through agent interactions, performance optimization, flexible environment configuration, and an extensible architecture for customization.

simba
Simba is an open source, portable Knowledge Management System (KMS) designed to seamlessly integrate with any Retrieval-Augmented Generation (RAG) system. It features a modern UI and modular architecture, allowing developers to focus on building advanced AI solutions without the complexities of knowledge management. Simba offers a user-friendly interface to visualize and modify document chunks, supports various vector stores and embedding models, and simplifies knowledge management for developers. It is community-driven, extensible, and aims to enhance AI functionality by providing a seamless integration with RAG-based systems.

VITA
VITA is an open-source interactive omni multimodal Large Language Model (LLM) capable of processing video, image, text, and audio inputs simultaneously. It stands out with features like Omni Multimodal Understanding, Non-awakening Interaction, and Audio Interrupt Interaction. VITA can respond to user queries without a wake-up word, track and filter external queries in real-time, and handle various query inputs effectively. The model utilizes state tokens and a duplex scheme to enhance the multimodal interactive experience.

pilottai
PilottAI is a Python framework for building autonomous multi-agent systems with advanced orchestration capabilities. It provides enterprise-ready features for building scalable AI applications. The framework includes hierarchical agent systems, production-ready features like asynchronous processing and fault tolerance, advanced memory management with semantic storage, and integrations with multiple LLM providers and custom tools. PilottAI offers specialized agents for various tasks such as customer service, document processing, email handling, knowledge acquisition, marketing, research analysis, sales, social media, and web search. The framework also provides documentation, example use cases, and advanced features like memory management, load balancing, and fault tolerance.

comfyui-web-viewer
The ComfyUI Web Viewer by vrch.ai is a real-time AI-generated interactive art framework that integrates realtime streaming into ComfyUI workflows. It supports keyboard control nodes, OSC control nodes, sound input nodes, and more, accessible from any device with a web browser. It enables real-time interaction with AI-generated content, ideal for interactive visual projects and enhancing ComfyUI workflows with efficient content management and display.

CrewAI-Studio
CrewAI Studio is an application with a user-friendly interface for interacting with CrewAI, offering support for multiple platforms and various backend providers. It allows users to run crews in the background, export single-page apps, and use custom tools for APIs and file writing. The roadmap includes features like better import/export, human input, chat functionality, automatic crew creation, and multiuser environment support.

swift-ocr-llm-powered-pdf-to-markdown
Swift OCR is a powerful tool for extracting text from PDF files using OpenAI's GPT-4 Turbo with Vision model. It offers flexible input options, advanced OCR processing, performance optimizations, structured output, robust error handling, and scalable architecture. The tool ensures accurate text extraction, resilience against failures, and efficient handling of multiple requests.