talon-ai-tools

talon-ai-tools

Query LLMs and AI tools with voice commands

Stars: 66

Visit
 screenshot

Control large language models and AI tools through voice commands using the Talon Voice dictation engine. This tool is designed to help users quickly edit text, code by voice, reduce keyboard use for those with health issues, and speed up workflow by using AI commands across the desktop. It prompts and extends tools like Github Copilot and OpenAI API for text and image generation. Users can set up the tool by downloading the repo, obtaining an OpenAI API key, and customizing the endpoint URL for preferred models. The tool can be used without an OpenAI key and can be exclusively used with Copilot for those not needing LLM integration.

README:

Talon-AI-Tools

Control large language models and AI tools through voice commands using the Talon Voice dictation engine.

This functionality is especially helpful for users who:

  • want to quickly edit text and fix dictation errors
  • code by voice using tools like Cursorless
  • have health issues affecting their hands and want to reduce keyboard use
  • want to speed up their workflow and use AI commands across the entire desktop

Prompts and extends the following tools:

  • Github Copilot
  • OpenAI API (with any GPT model) or simonw/llm CLI for text generation and processing
    • Any OpenAI compatible model endpoint can be used (Azure, local llamafiles, etc)
  • OpenAI API for image recognition

Setup:

  1. Download or git clone this repo into your Talon user directory.
  2. Choose one of the following three options to configure LLM access (unless you want to exclusively use this with Copilot):

Option 1: Direct OpenAI API Access

  1. Obtain an OpenAI API key.
  2. Create a Python file anywhere in your Talon user directory.
  3. Set the key environment variable within the Python file:

[!CAUTION] Make sure you do not push the key to a public repo!

# Example of setting the environment variable
import os

os.environ["OPENAI_API_KEY"] = "YOUR-KEY-HERE"

Option 2 (recommended): simonw/llm CLI

  1. Install simonw/llm and set up one or more models to use.

[!NOTE] Run llm keys set with the name of the provider you wish to use to set the API key for your requests. i.e. llm keys set openai.

  1. Add the following lines to your settings:
user.model_endpoint = "llm"
# If the llm binary is not found on Talon's PATH, uncomment and set:
# user.model_llm_path = "/path/to/llm"
  1. Choose a model in settings:
user.model_default = "claude-3.7-sonnet"  # or whichever model you installed
  1. By default, all model interactions will be logged locally and viewable on your machine via llm logs. If you prefer, you can disable this with llm logs off.

Option 3: Custom Endpoint URL

  1. Add the following line to settings to use your preferred endpoint:
user.model_endpoint = "https://your-custom-endpoint.com/v1/chat/completions"

This works with any API that follows the OpenAI schema, including:

  • Azure OpenAI Service[]
  • Local LLM servers (e.g., llamafiles, ollama)
  • Self-hosted models with OpenAI-compatible wrappers

Usage

See the GPT or Copilot folders for usage examples.

Quickstart Video

Talon-AI-Tools Quickstart

For Tasks:

Click tags to check more tools for each tasks

For Jobs:

Alternative AI tools for talon-ai-tools

Similar Open Source Tools

For similar tasks

For similar jobs