pollinations
Your Friendly Open-Source Gen-AI Platform
Stars: 3945
pollinations.ai is an open-source generative AI platform based in Berlin, empowering community projects with accessible text, image, video, and audio generation APIs. It offers a unified API endpoint for various AI generation needs, including text, images, audio, and video. The platform provides features like image generation using models such as Flux, GPT Image, Seedream, and Kontext, video generation with Seedance and Veo, and audio generation with text-to-speech and speech-to-text capabilities. Users can access the platform through a web interface or API, and authentication is managed through API keys. The platform is community-driven, transparent, and ethical, aiming to make AI technology open, accessible, and interconnected while fostering innovation and responsible development.
README:
| Name | Description | Author |
|---|---|---|
| Moltbot With Pollinations.ai provider | Multi-channel AI assistant for Telegram and WhatsApp using Pollinations AI. | @its3li |
| Pollinations Provider for Vercel AI SDK | Vercel AI SDK provider for Pollinations: text, image, and speech generation. | @artsiombarouski |
| Fable Friend | Interactive AI story adventure with selectable role, genre, and image style. | @Shreemahor |
| PolliDev | Developer-focused tool for managing generative image assets and prompts. | @FabioArieiraBaia |
| Lapse Game 2075 | AI-powered game where decisions shape a country's future over four pillars | @makeypocket |
| PixArt AI | iOS app that turns text prompts into AI-generated artwork via Pollinations. | @bhyahmed86-maker |
| Polly IDE | IDE to generate Angular and React sites using Pollinations AI models | @nulls-brawl-site |
| SteamKit | AI toolkit for Steam developers to generate marketing and store page assets. | @zouspants123321 |
| CHATTY AI | Enhance and update CHATTY AI web app to work seamlessly on Free Tier. | @tefa2007h-dev |
| Presintation web site | AI-powered web app that creates professional presentations in seconds. | @cyberuz001 |
We've launched https://gen.pollinations.ai โ a single endpoint for all your AI generation needs: text, images, audio, video โ all in one place.
-
Unified endpoint โ single API at
gen.pollinations.aifor all generation - Pollen credits โ simple pay-as-you-go system ($1 โ 1 Pollen)
- All models, one place โ Flux, GPT-5, Claude, Gemini, Seedream, and more
- API keys โ publishable keys for frontend, secret keys for backend
Get started at enter.pollinations.ai and check out the API docs
-
2026-02-09 โ ๐ Massive Model Drop Added
imagen-4,flux-2-dev,grok-video, andLTX-2for next-level image and video generation. -
2026-02-09 โ ๐ต Audio Studio Generate music with
elevenmusic, transcribe withwhisper, and use emotive TTS witheleven_v3. -
2026-02-09 โ ๐ง Smarter Claude Upgraded
claude-largeto the latest Claude 3 Opus 4.6 for enhanced reasoning. -
2026-02-06 โ ๐ New Provider: api.airforce Added
imagen(Google Imagen 3) andgrok-video(xAI Grok Imagine Video) models via api.airforce. -
2026-02-05 โ ๐ Paid Models Update
claude,grok,kontext,seedream, andseedance-proare moving to paid-only. - 2026-02-02 โ ๐ Kimi K2.5 Upgraded model with vision support and improved reasoning capabilities.
-
2026-02-02 โ ๐ง Web Research Claude can now use
perplexityandgemini-searchtools for real-time answers. -
2026-02-02 โ ๐ Premium Access High-end models
veo,claude-large, andseedream-proare now available for credit holders. - 2026-02-02 โ ๐ฑ Moltbot Integration Use Pollinations on Telegram & WhatsApp via the new Moltbot provider.
-
2026-01-30 โ ๐ Paid-Only Models
claude-large,gemini-large,veo,seedream-pro, andnanobanana-prowill require purchased pollen. Free credits work for all other models.
Hello folks! This year, pollinations.ai is participating in Google Summer of Code 2026!
We're looking for talented developers to contribute to our open-source AI platform.
Website: Pollinations GSOC 2026
Mentors: Thomas Haferlach, Ayushman Bhattacharya, Nihal Gazi.
Projects: Check out our Project Ideas.
Timeline: Applications open March 16th 2026 Coding starts May 25th 2026
Please reach out to us about any queries on Discord or via email at
[email protected]
pollinations.ai is an open-source generative AI platform based in Berlin, powering 500+ community projects with accessible text, image, video, and audio generation APIs. We build in the open and keep AI accessible to everyoneโthanks to our amazing supporters.
- ๐ 100% Open Source โ code, decisions, roadmap all public
- ๐ค Community-Built โ 500+ projects already using our APIs
- ๐ฑ Pollen Tiers โ earn daily credits by contributing (tiers in beta)
- ๐ผ๏ธ Image Generation โ Flux, GPT Image, Seedream, Kontext
- ๐ฌ Video Generation โ Seedance, Veo (alpha)
- ๐ต Audio โ Text-to-speech and speech-to-text
- ๐ฃ Easy-to-use Packages (Packages)
curl 'https://gen.pollinations.ai/image/a%20beautiful%20sunset' -o image.jpgOr visit pollinations.ai for an interactive experience.
curl 'https://gen.pollinations.ai/text/Hello%20world'Simple GET endpoint:
curl 'https://gen.pollinations.ai/audio/Hello%20from%20Pollinations?voice=nova&key=YOUR_API_KEY' -o speech.mp3OpenAI TTS compatible:
curl 'https://gen.pollinations.ai/v1/audio/speech' \
-H 'Content-Type: application/json' \
-H 'Authorization: Bearer YOUR_API_KEY' \
-d '{"model": "tts-1", "input": "Hello from Pollinations!", "voice": "nova"}' \
-o speech.mp3Available voices: alloy, echo, fable, onyx, nova, shimmer, plus 30+ ElevenLabs voices.
Our MCP (Model Context Protocol) server enables AI assistants like Claude to generate images and audio directly. Learn more
Add this to your MCP client configuration:
{
"mcpServers": {
"pollinations": {
"command": "npx",
"args": ["@pollinations/model-context-protocol"]
}
}
}npx @pollinations/model-context-protocolCommunity alternatives like MCPollinations and Sequa MCP Server are also available.
AI assistants can:
- Generate images from text descriptions
- Create text-to-speech audio with various voice options
- Play audio responses through the system speakers
- Access all pollinations.ai models and services
- List available models, voices, and capabilities
For more advanced usage, check out our API documentation.
Get your API key at enter.pollinations.ai
| Key | Prefix | Use Case | Rate Limits | Status |
|---|---|---|---|---|
| Publishable | pk_ |
Client-side, demos, prototypes | 1 pollen per IP per hour | |
| Secret | sk_ |
Server-side only | No rate limits | Stable |
โ ๏ธ Publishable keys: Turnstile protection coming soon. Not recommended for production yet.
โ ๏ธ Never exposesk_keys in client-side code, git repos, or public URLs
๐ก Building an app? Use Bring Your Own Pollen โ users pay for their own usage, you pay $0
Each API key can be scoped to specific models. When creating a key at enter.pollinations.ai, you can:
- Allow all models โ key works with any available model
-
Restrict to specific models โ select exactly which models the key can access (e.g., only
fluxandopenai, or justgptimage-large)
curl 'https://gen.pollinations.ai/image/a%20cat?key=YOUR_KEY'Environment variable (best practice):
export POLLINATIONS_API_KEY=sk_...See full API docs for detailed authentication information.
Our web interface is user-friendly and doesn't require any technical knowledge. Simply visit https://pollinations.ai and start creating!
Use our API directly in your browser or applications:
https://pollinations.ai/p/conceptual_isometric_world_of_pollinations_ai_surreal_hyperrealistic_digital_garden
Replace the description with your own, and you'll get a unique image based on your words!
Here's an example of a generated image:
Python code to download the generated image:
import requests
def download_image(prompt):
url = f"https://pollinations.ai/p/{prompt}"
response = requests.get(url)
with open('generated_image.jpg', 'wb') as file:
file.write(response.content)
print('Image downloaded!')
download_image("conceptual_isometric_world_of_pollinations_ai_surreal_hyperrealistic_digital_garden")
To generate text:
https://gen.pollinations.ai/text/What%20is%20artificial%20intelligence?
Generate speech from text:
https://gen.pollinations.ai/audio/Hello%20from%20Pollinations?voice=alloy&key=YOUR_API_KEY
Or use the OpenAI TTS-compatible endpoint:
curl 'https://gen.pollinations.ai/v1/audio/speech' \
-H 'Content-Type: application/json' \
-H 'Authorization: Bearer YOUR_API_KEY' \
-d '{"model": "tts-1", "input": "Hello from Pollinations!", "voice": "alloy"}' \
-o speech.mp3Check out our Pollinations SDK for Node.js, browser, and React integration.
graph LR
Q[Bots - Discord, Telegram, WhatsApp] --> GEN
N[30+ Mobile and Web Apps] --> GEN
A[pollinations.ai Web Frontend] --> GEN
R[AI Agents - Qwen, Sillytavern, ...] --> GEN
AI[AI Assistants - Claude] --> MCP[MCP Server]
MCP --> GEN
GEN[gen.pollinations.ai] --> ENTER[enter.pollinations.ai Gateway]
ENTER --> IMG[Image Service]
ENTER --> TXT[Text Service]
ENTER --> AUD[Audio Service]
IMG --> CF[Cloudflare Worker with R2 Cache]
CF --> B[image-origin.pollinations.ai]
B --> D[FLUX / GPT Image / Seedream - GPU VMs]
AUD --> EL[ElevenLabs TTS API]
TXT --> C[text.pollinations.ai]
C --> SC[Scaleway API]
C --> DS[Deepseek API]
C --> G[Azure-hosted LLMs]
C --> CFM[Cloudflare AI]We're constantly exploring new ways to push the boundaries of AI-driven content creation. Some areas we're excited about include:
- Digital Twins: Creating interactive AI-driven avatars
- Music Video Generation: Combining AI-generated visuals with music for unique video experiences
- Real-time AI-driven Visual Experiences: Projects like our Dreamachine, which create immersive, personalized visual journeys
pollinations.ai envisions a future where AI technology is:
-
Open & Accessible: We believe AI should be available to everyone โ earn daily Pollen by contributing, no credit card required
-
Transparent & Ethical: Our open-source approach ensures transparency in how our models work and behave
-
Community-Driven: We're building a platform where developers, creators, and AI enthusiasts can collaborate and innovate
-
Interconnected: We're creating an ecosystem where AI services can seamlessly work together, fostering innovation through composability
-
Evolving: We embrace the rapid evolution of AI technology while maintaining our commitment to openness and accessibility
We're committed to developing AI technology that serves humanity while respecting ethical boundaries and promoting responsible innovation. Join us in shaping the future of AI.
We believe in community-driven development. You can contribute to pollinations.ai in several ways:
-
Coding Assistant: The easiest way to contribute! Just create a GitHub issue describing the feature youโd like to see implemented. The MentatBot AI assistant will analyze and implement it directly! No coding required - just describe what you want.
-
Project Submissions: Have you built something with pollinations.ai? Use our project submission template (labeled as APPS) to share it with the community and get it featured in our README.
-
Feature Requests & Bug Reports: Have an idea or found a bug? Open an issue and let us know. Our team and the MentatBot assistant will review it.
-
Community Engagement: Join our vibrant Discord community to:
- Share your creations
- Get support and help others
- Collaborate with fellow AI enthusiasts
- Discuss feature ideas before creating issues
For any questions or support, please visit our Discord channel or create an issue on our GitHub repository.
Our codebase is organized into several key folders, each serving a specific purpose in the pollinations.ai ecosystem:
-
pollinations.ai/: The main React application for the Pollinations.ai website. -
image.pollinations.ai/: Backend service for image generation and caching with Cloudflare Workers and R2 storage. -
packages/sdk/: SDK and React component library for Pollinations.ai. -
text.pollinations.ai/: Backend service for text generation. -
packages/mcp/: Model Context Protocol (MCP) server for AI assistants like Claude to generate images directly.
This structure encompasses the frontend website, backend services for image and text generation, and integrations like the Discord bot and MCP server, providing a comprehensive framework for the pollinations.ai platform.
For development setup and environment management, see DEVELOP.md.
pollinations.ai is proud to be supported by:
- Perplexity AI: AI-powered search and conversational answer engine
- AWS Activate: GPU Cloud Credits
- io.net: Decentralized GPU network for AI compute
- BytePlus: Official ByteDance cloud services and AI solutions
- Google Cloud for Startups: GPU Cloud Credits
- NVIDIA Inception: AI startup support
- Azure (MS for Startups): OpenAI credits
- Cloudflare: Put the connectivity cloud to work for you.
- Scaleway: Europe's empowering cloud provider
- Modal: High-performance AI infrastructure
- NavyAI: AI API provider for OpenAI o3 and Gemini models
- Nebius: AI-optimized cloud infrastructure with NVIDIA GPU clusters
- api.airforce: AI API gateway powering Imagen 3 and Grok Imagine Video models
The best way to support pollinations.ai is by using our product! Get your API key and start building at enter.pollinations.ai.
๐ Twitter ยท Instagram ยท LinkedIn ยท Facebook ยท Reddit ยท YouTube
pollinations.ai is open-source software licensed under the MIT license.
Made with โค๏ธ by the pollinations.ai team
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for pollinations
Similar Open Source Tools
pollinations
pollinations.ai is an open-source generative AI platform based in Berlin, empowering community projects with accessible text, image, video, and audio generation APIs. It offers a unified API endpoint for various AI generation needs, including text, images, audio, and video. The platform provides features like image generation using models such as Flux, GPT Image, Seedream, and Kontext, video generation with Seedance and Veo, and audio generation with text-to-speech and speech-to-text capabilities. Users can access the platform through a web interface or API, and authentication is managed through API keys. The platform is community-driven, transparent, and ethical, aiming to make AI technology open, accessible, and interconnected while fostering innovation and responsible development.
PromptX
PromptX is a leading AI agent context platform that revolutionizes interaction design, enabling AI agents to become industry experts. It offers core capabilities such as an AI role creation platform, intelligent tool development platform, and cognitive memory system. PromptX allows users to easily discover experts, summon them for assistance, and engage in professional dialogues through natural conversations. The platform's core philosophy emphasizes treating AI as a person, enabling users to communicate naturally without the need for complex commands. With Nuwa Creation Workshop, users can design custom AI roles using meta-prompt technology, transforming abstract needs into concrete executable AI expert roles in just minutes.
BrowserAI
BrowserAI is a tool that allows users to run large language models (LLMs) directly in the browser, providing a simple, fast, and open-source solution. It prioritizes privacy by processing data locally, is cost-effective with no server costs, works offline after initial download, and offers WebGPU acceleration for high performance. It is developer-friendly with a simple API, supports multiple engines, and comes with pre-configured models for easy use. Ideal for web developers, companies needing privacy-conscious AI solutions, researchers experimenting with browser-based AI, and hobbyists exploring AI without infrastructure overhead.
RepoMaster
RepoMaster is an AI agent that leverages GitHub repositories to solve complex real-world tasks. It transforms how coding tasks are solved by automatically finding the right GitHub tools and making them work together seamlessly. Users can describe their tasks, and RepoMaster's AI analysis leads to auto discovery and smart execution, resulting in perfect outcomes. The tool provides a web interface for beginners and a command-line interface for advanced users, along with specialized agents for deep search, general assistance, and repository tasks.
crawl4ai
Crawl4AI is a powerful and free web crawling service that extracts valuable data from websites and provides LLM-friendly output formats. It supports crawling multiple URLs simultaneously, replaces media tags with ALT, and is completely free to use and open-source. Users can integrate Crawl4AI into Python projects as a library or run it as a standalone local server. The tool allows users to crawl and extract data from specified URLs using different providers and models, with options to include raw HTML content, force fresh crawls, and extract meaningful text blocks. Configuration settings can be adjusted in the `crawler/config.py` file to customize providers, API keys, chunk processing, and word thresholds. Contributions to Crawl4AI are welcome from the open-source community to enhance its value for AI enthusiasts and developers.
ai
Jetify's AI SDK for Go is a unified interface for interacting with multiple AI providers including OpenAI, Anthropic, and more. It addresses the challenges of fragmented ecosystems, vendor lock-in, poor Go developer experience, and complex multi-modal handling by providing a unified interface, Go-first design, production-ready features, multi-modal support, and extensible architecture. The SDK supports language models, embeddings, image generation, multi-provider support, multi-modal inputs, tool calling, and structured outputs.
natively-cluely-ai-assistant
Natively is a free, open-source, privacy-first AI assistant designed to help users in real time during meetings, interviews, presentations, and conversations. Unlike traditional AI tools that work after the conversation, Natively operates while the conversation is happening. It runs as an invisible, always-on-top desktop overlay, listens when prompted, observes the screen content, and provides instant, context-aware assistance. The tool is fully transparent, customizable, and grants users complete control over local vs cloud AI, data, and credentials.
UI-TARS-desktop
UI-TARS-desktop is a desktop application that provides a native GUI Agent based on the UI-TARS model. It offers features such as natural language control powered by Vision-Language Model, screenshot and visual recognition support, precise mouse and keyboard control, cross-platform support (Windows/MacOS/Browser), real-time feedback and status display, and private and secure fully local processing. The application aims to enhance the user's computer experience, introduce new browser operation features, and support the advanced UI-TARS-1.5 model for improved performance and precise control.
whatsapp-ai-bot
The WhatsApp AI Bot is a chatbot that utilizes various AI models APIs to generate responses to user input. Users can interact with the bot using commands to access different AI models such as Gemini, Gemini-Vision, CHAT-GPT, DALL-E, and Stability AI. Additionally, users have the flexibility to create their own custom models to personalize the bot's behavior. The bot operates on WhatsApp Web through Puppeteer and requires API keys for Gemini, OpenAI, and StabilityAI. It provides a range of functionalities and customization options for users interested in AI-powered chatbots.
RWKV_APP
RWKV App is an experimental application that enables users to run Large Language Models (LLMs) offline on their edge devices. It offers a privacy-first, on-device LLM experience for everyday devices. Users can engage in multi-turn conversations, text-to-speech, visual understanding, and more, all without requiring an internet connection. The app supports switching between different models, running locally without internet, and exploring various AI tasks such as chat, speech generation, and visual understanding. It is built using Flutter and Dart FFI for cross-platform compatibility and efficient communication with the C++ inference engine. The roadmap includes integrating features into the RWKV Chat app, supporting more model weights, hardware, operating systems, and devices.
GPTSwarm
GPTSwarm is a graph-based framework for LLM-based agents that enables the creation of LLM-based agents from graphs and facilitates the customized and automatic self-organization of agent swarms with self-improvement capabilities. The library includes components for domain-specific operations, graph-related functions, LLM backend selection, memory management, and optimization algorithms to enhance agent performance and swarm efficiency. Users can quickly run predefined swarms or utilize tools like the file analyzer. GPTSwarm supports local LM inference via LM Studio, allowing users to run with a local LLM model. The framework has been accepted by ICML2024 and offers advanced features for experimentation and customization.
project-blog
Welcome to the Blog Script Project, a collaborative platform for developers and writers to create, manage, and share content. With features like Markdown support, submodule integration, customizable templates, project contribution workflow, global visibility, community discussions, SEO optimization, and role-based dashboard, Blog Script enhances collaboration and visibility for your work. You can contribute by adding new projects, improving existing projects, updating documentation, fixing bugs, optimizing, and ensuring code readability. Follow the contribution guidelines to star the repository, find tasks, fork the repository, make changes, add screenshots, submit a pull request, and contribute to the open-source community. Additionally, you can add your project as a submodule by following the provided guidelines. Join us, contribute, and grow together!
GMTalker
GMTalker is an interactive digital human rendered by Unreal Engine, developed by the Media Intelligence Team at Bright Laboratory. The system integrates speech recognition, speech synthesis, natural language understanding, and lip-sync animation driving. It supports rapid deployment on Windows with only 2GB of VRAM required. The project showcases two 3D cartoon digital human avatars suitable for presentations, expansions, and commercial integration.
bytebot
Bytebot is an open-source AI desktop agent that provides a virtual employee with its own computer to complete tasks for users. It can use various applications, download and organize files, log into websites, process documents, and perform complex multi-step workflows. By giving AI access to a complete desktop environment, Bytebot unlocks capabilities not possible with browser-only agents or API integrations, enabling complete task autonomy, document processing, and usage of real applications.
WebMasterLog
WebMasterLog is a comprehensive repository showcasing various web development projects built with front-end and back-end technologies. It highlights interactive user interfaces, dynamic web applications, and a spectrum of web development solutions. The repository encourages contributions in areas such as adding new projects, improving existing projects, updating documentation, fixing bugs, implementing responsive design, enhancing code readability, and optimizing project functionalities. Contributors are guided to follow specific guidelines for project submissions, including directory naming conventions, README file inclusion, project screenshots, and commit practices. Pull requests are reviewed based on criteria such as proper PR template completion, originality of work, code comments for clarity, and sharing screenshots for frontend updates. The repository also participates in various open-source programs like JWOC, GSSoC, Hacktoberfest, KWOC, 24 Pull Requests, IWOC, SWOC, and DWOC, welcoming valuable contributors.
AI-Infra-Guard
A.I.G (AI-Infra-Guard) is an AI red teaming platform by Tencent Zhuque Lab that integrates capabilities such as AI infra vulnerability scan, MCP Server risk scan, and Jailbreak Evaluation. It aims to provide users with a comprehensive, intelligent, and user-friendly solution for AI security risk self-examination. The platform offers features like AI Infra Scan, AI Tool Protocol Scan, and Jailbreak Evaluation, along with a modern web interface, complete API, multi-language support, cross-platform deployment, and being free and open-source under the MIT license.
For similar tasks
pollinations
pollinations.ai is an open-source generative AI platform based in Berlin, empowering community projects with accessible text, image, video, and audio generation APIs. It offers a unified API endpoint for various AI generation needs, including text, images, audio, and video. The platform provides features like image generation using models such as Flux, GPT Image, Seedream, and Kontext, video generation with Seedance and Veo, and audio generation with text-to-speech and speech-to-text capabilities. Users can access the platform through a web interface or API, and authentication is managed through API keys. The platform is community-driven, transparent, and ethical, aiming to make AI technology open, accessible, and interconnected while fostering innovation and responsible development.
obsidian-arcana
Arcana is a plugin for Obsidian that offers a collection of AI-powered tools inspired by famous historical figures to enhance creativity and productivity. It includes tools for conversation, text-to-speech transcription, speech-to-text replies, metadata markup, text generation, file moving, flashcard generation, auto tagging, and note naming. Users can interact with these tools using the command palette and sidebar views, with an OpenAI API key required for usage. The plugin aims to assist users in various note-taking and knowledge management tasks within the Obsidian vault environment.
TigerBot
TigerBot is a cutting-edge foundation for your very own LLM, providing a world-class large model for innovative Chinese-style contributions. It offers various upgrades and features, such as search mode enhancements, support for large context lengths, and the ability to play text-based games. TigerBot is suitable for prompt-based game engine development, interactive game design, and real-time feedback for playable games.
For similar jobs
promptflow
**Prompt flow** is a suite of development tools designed to streamline the end-to-end development cycle of LLM-based AI applications, from ideation, prototyping, testing, evaluation to production deployment and monitoring. It makes prompt engineering much easier and enables you to build LLM apps with production quality.
deepeval
DeepEval is a simple-to-use, open-source LLM evaluation framework specialized for unit testing LLM outputs. It incorporates various metrics such as G-Eval, hallucination, answer relevancy, RAGAS, etc., and runs locally on your machine for evaluation. It provides a wide range of ready-to-use evaluation metrics, allows for creating custom metrics, integrates with any CI/CD environment, and enables benchmarking LLMs on popular benchmarks. DeepEval is designed for evaluating RAG and fine-tuning applications, helping users optimize hyperparameters, prevent prompt drifting, and transition from OpenAI to hosting their own Llama2 with confidence.
MegaDetector
MegaDetector is an AI model that identifies animals, people, and vehicles in camera trap images (which also makes it useful for eliminating blank images). This model is trained on several million images from a variety of ecosystems. MegaDetector is just one of many tools that aims to make conservation biologists more efficient with AI. If you want to learn about other ways to use AI to accelerate camera trap workflows, check out our of the field, affectionately titled "Everything I know about machine learning and camera traps".
leapfrogai
LeapfrogAI is a self-hosted AI platform designed to be deployed in air-gapped resource-constrained environments. It brings sophisticated AI solutions to these environments by hosting all the necessary components of an AI stack, including vector databases, model backends, API, and UI. LeapfrogAI's API closely matches that of OpenAI, allowing tools built for OpenAI/ChatGPT to function seamlessly with a LeapfrogAI backend. It provides several backends for various use cases, including llama-cpp-python, whisper, text-embeddings, and vllm. LeapfrogAI leverages Chainguard's apko to harden base python images, ensuring the latest supported Python versions are used by the other components of the stack. The LeapfrogAI SDK provides a standard set of protobuffs and python utilities for implementing backends and gRPC. LeapfrogAI offers UI options for common use-cases like chat, summarization, and transcription. It can be deployed and run locally via UDS and Kubernetes, built out using Zarf packages. LeapfrogAI is supported by a community of users and contributors, including Defense Unicorns, Beast Code, Chainguard, Exovera, Hypergiant, Pulze, SOSi, United States Navy, United States Air Force, and United States Space Force.
llava-docker
This Docker image for LLaVA (Large Language and Vision Assistant) provides a convenient way to run LLaVA locally or on RunPod. LLaVA is a powerful AI tool that combines natural language processing and computer vision capabilities. With this Docker image, you can easily access LLaVA's functionalities for various tasks, including image captioning, visual question answering, text summarization, and more. The image comes pre-installed with LLaVA v1.2.0, Torch 2.1.2, xformers 0.0.23.post1, and other necessary dependencies. You can customize the model used by setting the MODEL environment variable. The image also includes a Jupyter Lab environment for interactive development and exploration. Overall, this Docker image offers a comprehensive and user-friendly platform for leveraging LLaVA's capabilities.
carrot
The 'carrot' repository on GitHub provides a list of free and user-friendly ChatGPT mirror sites for easy access. The repository includes sponsored sites offering various GPT models and services. Users can find and share sites, report errors, and access stable and recommended sites for ChatGPT usage. The repository also includes a detailed list of ChatGPT sites, their features, and accessibility options, making it a valuable resource for ChatGPT users seeking free and unlimited GPT services.
TrustLLM
TrustLLM is a comprehensive study of trustworthiness in LLMs, including principles for different dimensions of trustworthiness, established benchmark, evaluation, and analysis of trustworthiness for mainstream LLMs, and discussion of open challenges and future directions. Specifically, we first propose a set of principles for trustworthy LLMs that span eight different dimensions. Based on these principles, we further establish a benchmark across six dimensions including truthfulness, safety, fairness, robustness, privacy, and machine ethics. We then present a study evaluating 16 mainstream LLMs in TrustLLM, consisting of over 30 datasets. The document explains how to use the trustllm python package to help you assess the performance of your LLM in trustworthiness more quickly. For more details about TrustLLM, please refer to project website.
AI-YinMei
AI-YinMei is an AI virtual anchor Vtuber development tool (N card version). It supports fastgpt knowledge base chat dialogue, a complete set of solutions for LLM large language models: [fastgpt] + [one-api] + [Xinference], supports docking bilibili live broadcast barrage reply and entering live broadcast welcome speech, supports Microsoft edge-tts speech synthesis, supports Bert-VITS2 speech synthesis, supports GPT-SoVITS speech synthesis, supports expression control Vtuber Studio, supports painting stable-diffusion-webui output OBS live broadcast room, supports painting picture pornography public-NSFW-y-distinguish, supports search and image search service duckduckgo (requires magic Internet access), supports image search service Baidu image search (no magic Internet access), supports AI reply chat box [html plug-in], supports AI singing Auto-Convert-Music, supports playlist [html plug-in], supports dancing function, supports expression video playback, supports head touching action, supports gift smashing action, supports singing automatic start dancing function, chat and singing automatic cycle swing action, supports multi scene switching, background music switching, day and night automatic switching scene, supports open singing and painting, let AI automatically judge the content.