ai-gradio
None
Stars: 88
ai-gradio is a Python package that simplifies the creation of machine learning apps using various models like OpenAI, Google's Gemini, Anthropic's Claude, LumaAI, CrewAI, XAI's Grok, and Hyperbolic. It provides easy installation with support for different providers and offers features like text chat, voice chat, video chat, code generation interfaces, and AI agent teams. Users can set API keys for different providers and customize interfaces for specific tasks.
README:
A Python package that makes it easy for developers to create machine learning apps powered by OpenAI, Google's Gemini models, Anthropic's Claude, LumaAI, CrewAI, XAI's Grok, and Hyperbolic and more.
You can install ai-gradio
with different providers:
# Install with OpenAI support
pip install 'ai-gradio[openai]'
# Install with Gemini support
pip install 'ai-gradio[gemini]'
# Install with CrewAI support
pip install 'ai-gradio[crewai]'
# Install with Anthropic support
pip install 'ai-gradio[anthropic]'
# Install with LumaAI support
pip install 'ai-gradio[lumaai]'
# Install with XAI support
pip install 'ai-gradio[xai]'
# Install with Cohere support
pip install 'ai-gradio[cohere]'
# Install with SambaNova support
pip install 'ai-gradio[sambanova]'
# Install with Hyperbolic support
pip install 'ai-gradio[hyperbolic]'
# Install with all providers
pip install 'ai-gradio[all]'
# Installation additions:
pip install 'ai-gradio[fireworks]'
pip install 'ai-gradio[together]'
pip install 'ai-gradio[qwen]'
# Install with DeepSeek support
pip install 'ai-gradio[deepseek]'
# Install with SmolagentsAI support
pip install 'ai-gradio[smolagents]'
First, set your API key in the environment:
For OpenAI:
export OPENAI_API_KEY=<your token>
For Gemini:
export GEMINI_API_KEY=<your token>
For Anthropic:
export ANTHROPIC_API_KEY=<your token>
For LumaAI:
export LUMAAI_API_KEY=<your token>
For XAI:
export XAI_API_KEY=<your token>
For Cohere:
export COHERE_API_KEY=<your token>
For SambaNova:
export SAMBANOVA_API_KEY=<your token>
For Hyperbolic:
export HYPERBOLIC_API_KEY=<your token>
For DeepSeek:
export DEEPSEEK_API_KEY=<your token>
Then in a Python file:
import gradio as gr
import ai_gradio
# Create a Gradio interface
gr.load(
name='openai:gpt-4-turbo', # or 'gemini:gemini-1.5-flash' for Gemini, or 'xai:grok-beta' for Grok
src=ai_gradio.registry,
title='AI Chat',
description='Chat with an AI model'
).launch()
Basic text chat is supported for all text models. The interface provides a chat-like experience where you can have conversations with the AI model.
Voice chat is supported for OpenAI realtime models. You can enable it in two ways:
# Using a realtime model
gr.load(
name='openai:gpt-4o-realtime-preview-2024-10-01',
src=ai_gradio.registry
).launch()
# Or explicitly enabling voice chat with any realtime model
gr.load(
name='openai:gpt-4o-mini-realtime-preview-2024-12-17',
src=ai_gradio.registry,
enable_voice=True
).launch()
For voice chat functionality, you'll need:
- OpenAI API key (required):
export OPENAI_API_KEY=<your OpenAI token>
- Twilio credentials (recommended for better WebRTC performance):
export TWILIO_ACCOUNT_SID=<your Twilio account SID>
export TWILIO_AUTH_TOKEN=<your Twilio auth token>
You can get Twilio credentials by:
- Creating a free account at Twilio
- Finding your Account SID and Auth Token in the Twilio Console
Without Twilio credentials, voice chat will still work but might have connectivity issues in some network environments.
import gradio as gr
import ai_gradio
# Create a code generation interface with Gemini
gr.load(
name='gemini:gemini-pro',
src=ai_gradio.registry,
coder=True, # Enable code generation interface
title='Gemini Code Generator',
description='Generate web applications with Gemini'
).launch()
This creates an interactive code generation interface with:
- Input area for describing the desired web application
- Live preview of generated code
- Example templates
- System prompt configuration
- Code history tracking
- Real-time code preview
Example prompts:
examples = [
"Create a button that changes color when clicked",
"Create a simple todo list with add/remove functionality",
"Create a countdown timer with start/pause/reset controls"
]
Video chat is supported for Gemini models. You can enable it by setting enable_video=True
:
gr.load(
name='gemini:gemini-1.5-flash',
src=ai_gradio.registry,
enable_video=True
).launch()
DeepSeek models support text generation and coding assistance:
gr.load(
name='deepseek:deepseek-chat',
src=ai_gradio.registry,
title='DeepSeek Chat',
description='Chat with DeepSeek'
).launch()
# For code assistance
gr.load(
name='deepseek:deepseek-coder',
src=ai_gradio.registry,
title='DeepSeek Coder',
description='Get coding help from DeepSeek'
).launch()
# For vision tasks
gr.load(
name='deepseek:deepseek-vision',
src=ai_gradio.registry,
title='DeepSeek Vision',
description='Visual understanding with DeepSeek'
).launch()
Anthropic's Claude models are supported for text generation:
gr.load(
name='anthropic:claude-3-opus-20240229',
src=ai_gradio.registry,
title='Claude Chat',
description='Chat with Claude'
).launch()
LumaAI support allows you to generate videos and images from text prompts:
# For video generation
gr.load(
name='lumaai:dream-machine',
src=ai_gradio.registry,
title='LumaAI Video Generation'
).launch()
# For image generation
gr.load(
name='lumaai:photon-1',
src=ai_gradio.registry,
title='LumaAI Image Generation'
).launch()
Hyperbolic models support various LLMs including DeepSeek, LLaMA, and Qwen:
# Using DeepSeek V3
gr.load(
name='hyperbolic:deepseek-ai/DeepSeek-V3',
src=ai_gradio.registry,
title='DeepSeek Chat',
description='Chat with DeepSeek V3'
).launch()
# Using LLaMA 3.3
gr.load(
name='hyperbolic:meta-llama/llama-3.3-70b',
src=ai_gradio.registry,
title='LLaMA Chat',
description='Chat with LLaMA 3.3'
).launch()
# Using Qwen Coder
gr.load(
name='hyperbolic:Qwen/qwen2.5-coder-32b',
src=ai_gradio.registry,
title='Qwen Coder',
description='Get coding help from Qwen'
).launch()
smolagents support provides an intelligent assistant with web search capabilities:
gr.load(
name='smolagents:meta-llama/Llama-3.1-8B-Instruct',
src=ai_gradio.registry,
title='smolagents Assistant',
description='Agent Assistant'
).launch()
The SmolagentsAI integration includes:
- Interactive code generation
- Web search integration via DuckDuckGo
- Step-by-step thought process visibility
- Error handling and debugging assistance
- Real-time streaming responses
- meta-llama/Llama-3.1-8B-Instruct
- meta-llama/Llama-3.1-70B-Instruct
- meta-llama/Llama-3.2-70B-Instruct
CrewAI support allows you to create teams of AI agents that work together to solve complex tasks. Enable it by using the CrewAI provider:
gr.load(
name='crewai:gpt-4-turbo',
src=ai_gradio.registry,
title='AI Team Chat',
description='Chat with a team of specialized AI agents'
).launch()
The CrewAI integration supports different specialized agent teams:
-
support
: A team of support agents that help answer questions, including:- Senior Support Representative
- Support Quality Assurance Specialist
-
article
: A team of content creation agents, including:- Content Planner
- Content Writer
- Editor
You can specify the crew type when creating the interface:
gr.load(
name='crewai:gpt-4-turbo',
src=ai_gradio.registry,
crew_type='article', # or 'support'
title='AI Writing Team',
description='Create articles with a team of AI agents'
).launch()
When using the support
crew type, you can provide a documentation URL that the agents will reference when answering questions. The interface will automatically show a URL input field.
When loading a model, you can specify the provider explicitly using the format provider:model_name
.
# Explicit provider
gr.load(
name='gemini:gemini-pro',
src=ai_gradio.registry
).launch()
You can customize the interface by adding examples, changing the title, or adding a description:
gr.load(
name='gpt-4-turbo',
src=ai_gradio.registry,
title='Custom AI Chat',
description='Chat with an AI assistant',
examples=[
"Explain quantum computing to a 5-year old",
"What's the difference between machine learning and AI?"
]
).launch()
You can combine multiple models in a single interface using Gradio's Blocks:
import gradio as gr
import ai_gradio
with gr.Blocks() as demo:
with gr.Tab("GPT-4"):
gr.load('gpt-4-turbo', src=ai_gradio.registry)
with gr.Tab("Gemini"):
gr.load('gemini-pro', src=ai_gradio.registry)
with gr.Tab("Claude"):
gr.load('anthropic:claude-3-opus-20240229', src=ai_gradio.registry)
with gr.Tab("LumaAI"):
gr.load('lumaai:dream-machine', src=ai_gradio.registry)
with gr.Tab("CrewAI"):
gr.load('crewai:gpt-4-turbo', src=ai_gradio.registry)
with gr.Tab("Grok"):
gr.load('xai:grok-beta', src=ai_gradio.registry)
demo.launch()
- gpt-4-turbo
- gpt-4
- gpt-3.5-turbo
- gemini-pro
- gemini-pro-vision
- gemini-2.0-flash-exp
- claude-3-opus-20240229
- claude-3-sonnet-20240229
- claude-3-haiku-20240307
- claude-2.1
- claude-2.0
- claude-instant-1.2
- dream-machine (video generation)
- photon-1 (image generation)
- photon-flash-1 (fast image generation)
- crewai:gpt-4-turbo
- crewai:gpt-4
- crewai:gpt-3.5-turbo
- grok-beta
- grok-vision-beta
- command
- command-light
- command-nightly
- command-r
- llama2-70b-chat
- llama2-13b-chat
- llama2-7b-chat
- mixtral-8x7b-chat
- mistral-7b-chat
- whisper-v3
- whisper-v3-turbo
- f1-preview
- f1-mini
- meta-llama/Llama-Vision-Free
- meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo
- meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo
- meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo
- meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo
- meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo
- meta-llama/Meta-Llama-3-8B-Instruct-Turbo
- meta-llama/Meta-Llama-3-70B-Instruct-Turbo
- meta-llama/Llama-3.2-3B-Instruct-Turbo
- meta-llama/Meta-Llama-3-8B-Instruct-Lite
- meta-llama/Meta-Llama-3-70B-Instruct-Lite
- qwen-turbo-latest
- qwen-turbo
- qwen-plus
- qwen-max
- qwen1.5-110b-chat
- qwen1.5-72b-chat
- qwen1.5-32b-chat
- qwen1.5-14b-chat
- qwen1.5-7b-chat
- qwq-32b-preview
- qvq-72b-preview
- meta-llama/llama-3.3-70b
- Qwen/QwQ-32B-Preview
- Qwen/qwen2.5-coder-32b
- meta-llama/llama-3.2-3b
- Qwen/qwen2.5-72b
- deepseek/deepseek-v2.5
- meta-llama/llama-3-70b
- hermes/hermes-3-70b
- meta-llama/llama-3.1-405b
- meta-llama/llama-3.1-70b
- meta-llama/llama-3.1-8b
- deepseek-chat
- deepseek-coder
- deepseek-vision
- Python 3.10 or higher
- gradio >= 5.9.1
Additional dependencies are installed based on your chosen provider:
- OpenAI:
openai>=1.58.1
- Gemini:
google-generativeai
- CrewAI:
crewai>=0.1.0
,langchain>=0.1.0
,langchain-openai>=0.0.2
,crewai-tools>=0.0.1
- Anthropic:
anthropic>=1.0.0
- LumaAI:
lumaai>=0.0.3
- XAI:
xai>=0.1.0
- Cohere:
cohere>=5.0.0
- DeepSeek:
openai>=1.58.1
- Hyperbolic:
openai>=1.58.1
If you get a 401 authentication error, make sure your API key is properly set. You can set it manually in your Python session:
import os
# For OpenAI
os.environ["OPENAI_API_KEY"] = "your-api-key"
# For Gemini
os.environ["GEMINI_API_KEY"] = "your-api-key"
# For Anthropic
os.environ["ANTHROPIC_API_KEY"] = "your-api-key"
# For LumaAI
os.environ["LUMAAI_API_KEY"] = "your-api-key"
# For XAI
os.environ["XAI_API_KEY"] = "your-api-key"
# For Cohere
os.environ["COHERE_API_KEY"] = "your-api-key"
# For SambaNova
os.environ["SAMBANOVA_API_KEY"] = "your-api-key"
# Environment variables additions:
export FIREWORKS_API_KEY=<your token>
export TOGETHER_API_KEY=<your token>
export QWEN_API_KEY=<your token>
export HYPERBOLIC_API_KEY=<your token>
# Additional troubleshooting environment variables:
os.environ["FIREWORKS_API_KEY"] = "your-api-key"
os.environ["TOGETHER_API_KEY"] = "your-api-key"
os.environ["QWEN_API_KEY"] = "your-api-key"
os.environ["HYPERBOLIC_API_KEY"] = "your-api-key"
os.environ["DEEPSEEK_API_KEY"] = "your-api-key"
### No Providers Error
If you see an error about no providers being installed, make sure you've installed the package with the desired provider:
```bash
# Install with OpenAI support
pip install 'ai-gradio[openai]'
# Install with Gemini support
pip install 'ai-gradio[gemini]'
# Install with CrewAI support
pip install 'ai-gradio[crewai]'
# Install with Anthropic support
pip install 'ai-gradio[anthropic]'
# Install with LumaAI support
pip install 'ai-gradio[lumaai]'
# Install with XAI support
pip install 'ai-gradio[xai]'
# Install with Cohere support
pip install 'ai-gradio[cohere]'
# Install all providers
pip install 'ai-gradio[all]'
For voice chat functionality:
- gradio-webrtc
- numba==0.60.0
- pydub
- librosa
- websockets
- twilio
- gradio_webrtc[vad]
- numpy
For video chat functionality:
- opencv-python
- Pillow
import gradio as gr
import ai_gradio
# Simple chat with GPT-4
gr.load(
name='openai:gpt-4-turbo',
src=ai_gradio.registry,
title='GPT-4 Chat',
description='Chat with GPT-4'
).launch()
import gradio as gr
import ai_gradio
with gr.Blocks() as demo:
gr.Markdown("# AI Model Hub")
with gr.Tab("Text Models"):
with gr.Tab("GPT-4"):
gr.load('openai:gpt-4-turbo', src=ai_gradio.registry)
with gr.Tab("Claude"):
gr.load('anthropic:claude-3-opus-20240229', src=ai_gradio.registry)
with gr.Tab("DeepSeek"):
gr.load('deepseek:deepseek-chat', src=ai_gradio.registry)
with gr.Tab("Vision Models"):
with gr.Tab("Gemini Vision"):
gr.load('gemini:gemini-pro-vision', src=ai_gradio.registry, enable_video=True)
with gr.Tab("LumaAI"):
gr.load('lumaai:dream-machine', src=ai_gradio.registry)
with gr.Tab("Specialized"):
with gr.Tab("Code Assistant"):
gr.load('deepseek:deepseek-coder', src=ai_gradio.registry)
with gr.Tab("AI Team"):
gr.load('crewai:gpt-4-turbo', src=ai_gradio.registry, crew_type='article')
demo.launch()
import gradio as gr
import ai_gradio
# Enable voice chat with GPT-4
gr.load(
name='openai:gpt-4-turbo',
src=ai_gradio.registry,
enable_voice=True,
title='Voice Chat',
description='Talk with GPT-4'
).launch()
import gradio as gr
import ai_gradio
# Chat interface with custom examples and CSS
gr.load(
name='gemini:gemini-pro',
src=ai_gradio.registry,
title='Gemini Pro Assistant',
description='Your AI research companion',
examples=[
"Explain quantum entanglement",
"What are the main differences between RNA and DNA?",
"How does a neural network learn?"
],
css=".gradio-container {background-color: #f0f8ff}"
).launch()
import gradio as gr
import ai_gradio
# CrewAI setup for article writing
gr.load(
name='crewai:gpt-4-turbo',
src=ai_gradio.registry,
crew_type='article',
title='AI Writing Team',
description='Collaborate with AI agents to create articles',
examples=[
"Write a blog post about sustainable energy",
"Create a technical tutorial about Docker containers"
]
).launch()
import gradio as gr
import ai_gradio
# CrewAI support team with documentation reference
gr.load(
name='crewai:gpt-4-turbo',
src=ai_gradio.registry,
crew_type='support',
title='AI Support Team',
description='Get help from AI support agents',
documentation_url='https://docs.example.com'
).launch()
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for ai-gradio
Similar Open Source Tools
ai-gradio
ai-gradio is a Python package that simplifies the creation of machine learning apps using various models like OpenAI, Google's Gemini, Anthropic's Claude, LumaAI, CrewAI, XAI's Grok, and Hyperbolic. It provides easy installation with support for different providers and offers features like text chat, voice chat, video chat, code generation interfaces, and AI agent teams. Users can set API keys for different providers and customize interfaces for specific tasks.
aio-scrapy
Aio-scrapy is an asyncio-based web crawling and web scraping framework inspired by Scrapy. It supports distributed crawling/scraping, implements compatibility with scrapyd, and provides options for using redis queue and rabbitmq queue. The framework is designed for fast extraction of structured data from websites. Aio-scrapy requires Python 3.9+ and is compatible with Linux, Windows, macOS, and BSD systems.
agentipy
Agentipy is a powerful toolkit for interacting with the Solana blockchain, providing easy-to-use functions for token operations, trading, yield farming, LangChain integration, performance tracking, token data retrieval, pump & fun token launching, Meteora DLMM pool creation, and more. It offers features like token transfers, balance checks, staking, deploying new tokens, requesting faucet funds, trading with customizable slippage, yield farming with Lulo, and accessing LangChain tools for enhanced blockchain interactions. Users can also track current transactions per second (TPS), fetch token data by ticker or address, launch pump & fun tokens, create Meteora DLMM pools, buy/sell tokens with Raydium liquidity, and burn/close token accounts individually or in batches.
python-genai
The Google Gen AI SDK is a Python library that provides access to Google AI and Vertex AI services. It allows users to create clients for different services, work with parameter types, models, generate content, call functions, handle JSON response schemas, stream text and image content, perform async operations, count and compute tokens, embed content, generate and upscale images, edit images, work with files, create and get cached content, tune models, distill models, perform batch predictions, and more. The SDK supports various features like automatic function support, manual function declaration, JSON response schema support, streaming for text and image content, async methods, tuning job APIs, distillation, batch prediction, and more.
deepseek-free-api
DeepSeek Free API is a high-speed streaming output tool that supports multi-turn conversations and zero-configuration deployment. It is compatible with the ChatGPT interface and offers multiple token support. The tool provides eight free APIs for various AI interfaces. Users can access the tool online, prepare for integration, deploy using Docker, Docker-compose, Render, Vercel, or native deployment methods. It also offers client recommendations for faster integration and supports dialogue completion and userToken live checks. The tool comes with important considerations for Nginx reverse proxy optimization and token statistics.
Groq2API
Groq2API is a REST API wrapper around the Groq2 model, a large language model trained by Google. The API allows you to send text prompts to the model and receive generated text responses. The API is easy to use and can be integrated into a variety of applications.
eval-scope
Eval-Scope is a framework for evaluating and improving large language models (LLMs). It provides a set of commonly used test datasets, metrics, and a unified model interface for generating and evaluating LLM responses. Eval-Scope also includes an automatic evaluator that can score objective questions and use expert models to evaluate complex tasks. Additionally, it offers a visual report generator, an arena mode for comparing multiple models, and a variety of other features to support LLM evaluation and development.
metaso-free-api
Metaso AI Free service supports high-speed streaming output, secret tower AI super network search (full network or academic as well as concise, in-depth, research three modes), zero-configuration deployment, multi-token support. Fully compatible with ChatGPT interface. It also has seven other free APIs available for use. The tool provides various deployment options such as Docker, Docker-compose, Render, Vercel, and native deployment. Users can access the tool for chat completions and token live checks. Note: Reverse API is unstable, it is recommended to use the official Metaso AI website to avoid the risk of banning. This project is for research and learning purposes only, not for commercial use.
candle-vllm
Candle-vllm is an efficient and easy-to-use platform designed for inference and serving local LLMs, featuring an OpenAI compatible API server. It offers a highly extensible trait-based system for rapid implementation of new module pipelines, streaming support in generation, efficient management of key-value cache with PagedAttention, and continuous batching. The tool supports chat serving for various models and provides a seamless experience for users to interact with LLMs through different interfaces.
AnglE
AnglE is a library for training state-of-the-art BERT/LLM-based sentence embeddings with just a few lines of code. It also serves as a general sentence embedding inference framework, allowing for inferring a variety of transformer-based sentence embeddings. The library supports various loss functions such as AnglE loss, Contrastive loss, CoSENT loss, and Espresso loss. It provides backbones like BERT-based models, LLM-based models, and Bi-directional LLM-based models for training on single or multi-GPU setups. AnglE has achieved significant performance on various benchmarks and offers official pretrained models for both BERT-based and LLM-based models.
LLM-Tuning
LLM-Tuning is a collection of tools and resources for fine-tuning large language models (LLMs). It includes a library of pre-trained LoRA models, a set of tutorials and examples, and a community forum for discussion and support. LLM-Tuning makes it easy to fine-tune LLMs for a variety of tasks, including text classification, question answering, and dialogue generation. With LLM-Tuning, you can quickly and easily improve the performance of your LLMs on downstream tasks.
e2m
E2M is a Python library that can parse and convert various file types into Markdown format. It supports the conversion of multiple file formats, including doc, docx, epub, html, htm, url, pdf, ppt, pptx, mp3, and m4a. The ultimate goal of the E2M project is to provide high-quality data for Retrieval-Augmented Generation (RAG) and model training or fine-tuning. The core architecture consists of a Parser responsible for parsing various file types into text or image data, and a Converter responsible for converting text or image data into Markdown format.
cool-admin-midway
Cool-admin (midway version) is a cool open-source backend permission management system that supports modular, plugin-based, rapid CRUD development. It facilitates the quick construction and iteration of backend management systems, deployable in various ways such as serverless, docker, and traditional servers. It features AI coding for generating APIs and frontend pages, flow orchestration for drag-and-drop functionality, modular and plugin-based design for clear and maintainable code. The tech stack includes Node.js, Midway.js, Koa.js, TypeScript for backend, and Vue.js, Element-Plus, JSX, Pinia, Vue Router for frontend. It offers friendly technology choices for both frontend and backend developers, with TypeScript syntax similar to Java and PHP for backend developers. The tool is suitable for those looking for a modern, efficient, and fast development experience.
hezar
Hezar is an all-in-one AI library designed specifically for the Persian community. It brings together various AI models and tools, making it easy to use AI with just a few lines of code. The library seamlessly integrates with Hugging Face Hub, offering a developer-friendly interface and task-based model interface. In addition to models, Hezar provides tools like word embeddings, tokenizers, feature extractors, and more. It also includes supplementary ML tools for deployment, benchmarking, and optimization.
LongLLaVA
LongLLaVA is a tool for scaling multi-modal LLMs to 1000 images efficiently via hybrid architecture. It includes stages for single-image alignment, instruction-tuning, and multi-image instruction-tuning, with evaluation through a command line interface and model inference. The tool aims to achieve GPT-4V level capabilities and beyond, providing reproducibility of results and benchmarks for efficiency and performance.
ruby-nano-bots
Ruby Nano Bots is an implementation of the Nano Bots specification supporting various AI providers like Cohere Command, Google Gemini, Maritaca AI MariTalk, Mistral AI, Ollama, OpenAI ChatGPT, and others. It allows calling tools (functions) and provides a helpful assistant for interacting with AI language models. The tool can be used both from the command line and as a library in Ruby projects, offering features like REPL, debugging, and encryption for data privacy.
For similar tasks
ai-gradio
ai-gradio is a Python package that simplifies the creation of machine learning apps using various models like OpenAI, Google's Gemini, Anthropic's Claude, LumaAI, CrewAI, XAI's Grok, and Hyperbolic. It provides easy installation with support for different providers and offers features like text chat, voice chat, video chat, code generation interfaces, and AI agent teams. Users can set API keys for different providers and customize interfaces for specific tasks.
For similar jobs
promptflow
**Prompt flow** is a suite of development tools designed to streamline the end-to-end development cycle of LLM-based AI applications, from ideation, prototyping, testing, evaluation to production deployment and monitoring. It makes prompt engineering much easier and enables you to build LLM apps with production quality.
deepeval
DeepEval is a simple-to-use, open-source LLM evaluation framework specialized for unit testing LLM outputs. It incorporates various metrics such as G-Eval, hallucination, answer relevancy, RAGAS, etc., and runs locally on your machine for evaluation. It provides a wide range of ready-to-use evaluation metrics, allows for creating custom metrics, integrates with any CI/CD environment, and enables benchmarking LLMs on popular benchmarks. DeepEval is designed for evaluating RAG and fine-tuning applications, helping users optimize hyperparameters, prevent prompt drifting, and transition from OpenAI to hosting their own Llama2 with confidence.
MegaDetector
MegaDetector is an AI model that identifies animals, people, and vehicles in camera trap images (which also makes it useful for eliminating blank images). This model is trained on several million images from a variety of ecosystems. MegaDetector is just one of many tools that aims to make conservation biologists more efficient with AI. If you want to learn about other ways to use AI to accelerate camera trap workflows, check out our of the field, affectionately titled "Everything I know about machine learning and camera traps".
leapfrogai
LeapfrogAI is a self-hosted AI platform designed to be deployed in air-gapped resource-constrained environments. It brings sophisticated AI solutions to these environments by hosting all the necessary components of an AI stack, including vector databases, model backends, API, and UI. LeapfrogAI's API closely matches that of OpenAI, allowing tools built for OpenAI/ChatGPT to function seamlessly with a LeapfrogAI backend. It provides several backends for various use cases, including llama-cpp-python, whisper, text-embeddings, and vllm. LeapfrogAI leverages Chainguard's apko to harden base python images, ensuring the latest supported Python versions are used by the other components of the stack. The LeapfrogAI SDK provides a standard set of protobuffs and python utilities for implementing backends and gRPC. LeapfrogAI offers UI options for common use-cases like chat, summarization, and transcription. It can be deployed and run locally via UDS and Kubernetes, built out using Zarf packages. LeapfrogAI is supported by a community of users and contributors, including Defense Unicorns, Beast Code, Chainguard, Exovera, Hypergiant, Pulze, SOSi, United States Navy, United States Air Force, and United States Space Force.
llava-docker
This Docker image for LLaVA (Large Language and Vision Assistant) provides a convenient way to run LLaVA locally or on RunPod. LLaVA is a powerful AI tool that combines natural language processing and computer vision capabilities. With this Docker image, you can easily access LLaVA's functionalities for various tasks, including image captioning, visual question answering, text summarization, and more. The image comes pre-installed with LLaVA v1.2.0, Torch 2.1.2, xformers 0.0.23.post1, and other necessary dependencies. You can customize the model used by setting the MODEL environment variable. The image also includes a Jupyter Lab environment for interactive development and exploration. Overall, this Docker image offers a comprehensive and user-friendly platform for leveraging LLaVA's capabilities.
carrot
The 'carrot' repository on GitHub provides a list of free and user-friendly ChatGPT mirror sites for easy access. The repository includes sponsored sites offering various GPT models and services. Users can find and share sites, report errors, and access stable and recommended sites for ChatGPT usage. The repository also includes a detailed list of ChatGPT sites, their features, and accessibility options, making it a valuable resource for ChatGPT users seeking free and unlimited GPT services.
TrustLLM
TrustLLM is a comprehensive study of trustworthiness in LLMs, including principles for different dimensions of trustworthiness, established benchmark, evaluation, and analysis of trustworthiness for mainstream LLMs, and discussion of open challenges and future directions. Specifically, we first propose a set of principles for trustworthy LLMs that span eight different dimensions. Based on these principles, we further establish a benchmark across six dimensions including truthfulness, safety, fairness, robustness, privacy, and machine ethics. We then present a study evaluating 16 mainstream LLMs in TrustLLM, consisting of over 30 datasets. The document explains how to use the trustllm python package to help you assess the performance of your LLM in trustworthiness more quickly. For more details about TrustLLM, please refer to project website.
AI-YinMei
AI-YinMei is an AI virtual anchor Vtuber development tool (N card version). It supports fastgpt knowledge base chat dialogue, a complete set of solutions for LLM large language models: [fastgpt] + [one-api] + [Xinference], supports docking bilibili live broadcast barrage reply and entering live broadcast welcome speech, supports Microsoft edge-tts speech synthesis, supports Bert-VITS2 speech synthesis, supports GPT-SoVITS speech synthesis, supports expression control Vtuber Studio, supports painting stable-diffusion-webui output OBS live broadcast room, supports painting picture pornography public-NSFW-y-distinguish, supports search and image search service duckduckgo (requires magic Internet access), supports image search service Baidu image search (no magic Internet access), supports AI reply chat box [html plug-in], supports AI singing Auto-Convert-Music, supports playlist [html plug-in], supports dancing function, supports expression video playback, supports head touching action, supports gift smashing action, supports singing automatic start dancing function, chat and singing automatic cycle swing action, supports multi scene switching, background music switching, day and night automatic switching scene, supports open singing and painting, let AI automatically judge the content.