
AI-Blueprints
📁 This repository hosts a growing collection of AI blueprint projects that run end-to-end using Jupyter notebooks, MLflow deployments, and Streamlit web apps.🛠️ All projects are built using HP AI Studio with ❤️ If you find this useful, please don’t forget to star the repository ⭐ and support our work 🚀
Stars: 179

This repository hosts a collection of AI blueprint projects for HP AI Studio, providing end-to-end solutions across key AI domains like data science, machine learning, deep learning, and generative AI. The projects are designed to be plug-and-play, utilizing open-source and hosted models to offer ready-to-use solutions. The repository structure includes projects related to classical machine learning, deep learning applications, generative AI, NGC integration, and troubleshooting guidelines for common issues. Each project is accompanied by detailed descriptions and use cases, showcasing the versatility and applicability of AI technologies in various domains.
README:
- Overview
- Repository Structure
- Data Science
- Deep Learning
- Generative AI
- NVIDIA GPU Cloud (NGC) Integration
- Troubleshooting
- Contact and Support
This repository hosts a growing collection of AI blueprint projects that run end-to-end using Jupyter notebooks, MLflow deployments, and Streamlit web apps. Each blueprint is designed to be plug-and-play, giving you ready-to-use solutions across key AI domains like data science, machine learning, deep learning, and generative AI. All projects are built to work seamlessly with HP AI Studio.
These blueprints make use of local, open-source models including:
- LLaMA (Meta)
- Qwen (Alibaba)
- BERT (Google)
- LLaMA Nemotron (NVIDIA)
In addition, some examples use hosted models from Hugging Face for added flexibility.
- Data visualization
- Data analysis
- Image classification
- Super resolution
- LLM fine-tuning
- Multi-modal RAG
- Audio translation
- Agentic RAG
- And much more
We’re actively expanding this collection. If you have ideas or want to suggest a blueprint for HP AI Studio, feel free to open an issue. We’d love to hear from you.
To run any of these projects quickly and effortlessly, you can download HP AI Studio for free.
# Root Directory
├── data-science/ # Projects related to classical machine learning and statistical analysis
│ ├── classification-with-svm/ # SVM-based classification implementation
│ └── data-analysis-with-var/ # Vector AutoRegression analysis workflow
│
├── deep-learning/ # Deep learning applications using popular frameworks
│ ├── classification-with-keras/ # Image classification using Keras
│ ├── question-answering-with-bert/ # QA system built on top of BERT model
│ ├── recommendation-system-with-tensorflow/ # TensorFlow-based recommendation engine
│ ├── spam-detection-with-nlp/ # NLP-driven spam classifier
│ ├── super-resolution-with-fsrcnn/ # Image enhancement using FSRCNN
│ └── text-generation-with-rnn/ # RNN-based generative model for text
│
├── generative-ai/ # Generative AI applications across text, code, and image
│ ├── automated-evaluation-with-structured-outputs/ # Eval pipeline for structured generation
│ ├── code-generation-with-langchain/ # Code synthesis using LangChain
│ ├── fine-tuning-with-orpo/ # ORPO-based fine-tuning procedure
│ ├── image-generation-with-stablediffusion/ # StableDiffusion-powered image generation
│ ├── text-generation-with-langchain/ # Text generation leveraging LangChain stack
│ ├── text-summarization-with-langchain/ # Summarization pipeline using LangChain
│ └── vanilla-rag-with-langchain/ # Basic Retrieval-Augmented Generation with LangChain
│
├── ngc-integration/ # Projects leveraging NVIDIA GPU Cloud and libraries
│ ├── agentic-rag-with-tensorrtllm/ # RAG system using TensorRT-LLM and agentic planning
│ ├── audio-translation-with-nemo/ # Speech translation with NVIDIA NeMo
│ ├── data-analysis-with-cudf/ # RAPIDS cuDF-based data manipulation
│ ├── data-visualization-with-cudf/ # Visualizations using GPU-accelerated tools
│ └── vacation-recommendation-with-bert/ # Recommendation app using BERT embeddings
The blueprint projects in this folder demonstrate how to build data science applications with HP AI Studio.
We provide 2 blueprint projects, each designed for quick and easy use to help you get started efficiently.
This project is a simple classification experiment focused on predicting species of iris flowers.
It runs on the Data Science Workspace, demonstrating basic supervised learning techniques for multi-class classification tasks.
This project explores a regression experiment using mobility data collected during the COVID-19 pandemic.
It highlights how city-level movement patterns changed during the crisis. The experiment runs on the Data Science Workspace.
The blueprint projects in this folder demonstrate how to build deep learning applications with HP AI Studio.
We provide 6 blueprint projects, each designed for quick and easy use to help you get started efficiently.
This project performs basic image classification using the TensorFlow framework.
It trains a model to classify handwritten digits from the MNIST dataset and runs on the Deep Learning Workspace.
This project demonstrates a simple BERT Question Answering (QA) experiment. It provides code to train a BERT-based model, as well as instructions to load a pretrained model from Hugging Face.
The model is deployed using MLflow to expose an inference service capable of answering questions based on input text.
This project builds a simple recommender system for movies using TensorFlow.
It trains on user-item interaction data to predict movie preferences and runs on the Deep Learning Workspace.
This project implements a text classification system to detect spam messages.
It uses deep learning techniques and requires the Deep Learning Workspace for training and inference.
This project showcases a Computer Vision experiment that applies convolutional neural networks for image super-resolution — enhancing the quality and resolution of input images.
This project illustrates how to build a simple character-by-character text generation model.
It trains on a dataset containing Shakespeare's texts, demonstrating the fundamentals of text generation by predicting one character at a time.
The blueprint projects in this folder demonstrate how to build generative AI applications with HP AI Studio.
We provide 7 blueprint projects, each designed for quick and easy use to help you get started efficiently.
Automated Evaluation with Structured Outputs turns a local Meta-Llama-3 model into an MLflow-served scorer that rates any batch of texts (e.g., project abstracts) against arbitrary rubric criteria.
- Generates scores locally via
llama.cpp
(no data leaves your machine) - Registers the evaluator as a pyfunc model in MLflow
- Exposes a REST
/invocations
endpoint - Ships two front-ends — a Streamlit dashboard and a pure HTML/JS UI — for instant human-friendly interaction and CSV download.
This notebook performs automatic code explanation by extracting code snippets from Jupyter notebooks and generating natural language descriptions using LLMs. It supports contextual enrichment based on adjacent markdown cells, enables configurable prompt templating, and integrates with PromptQuality and Galileo for evaluation and tracking. The pipeline is modular, supports local or hosted model inference, and is compatible with LLaMA, Mistral, and Hugging Face-based models. It also includes GitHub notebook crawling, metadata structuring, and vector store integration for downstream tasks like RAG and semantic search.
This project demonstrates a full-stack LLM fine-tuning experiment using ORPO (Open-Source Reinforcement Pretraining Objective) to align a base language model with human preference data. It leverages the Z by HP AI Studio Local GenAI environment, and uses models such as LLaMA 3, Gemma 1B, and Mistral 7B as foundations.
We incorporate:
Galileo PromptQuality for evaluating model responses with human-like scorers (e.g., context adherence) TensorBoard for human feedback visualization before fine-tuning A flexible model selector and inference runner architecture A comparative setup to benchmark base vs fine-tuned models on the same prompts
This notebook performs image generation inference using the Stable Diffusion architecture, with support for both standard and DreamBooth fine-tuned models. It loads configuration and secrets from YAML files, enables local or deployed inference execution, and calculates custom image quality metrics, such as entropy and complexity. The pipeline is modular, supports Hugging Face model loading, and integrates with PromptQuality for evaluation and tracking.
This notebook implements a full Retrieval-Augmented Generation (RAG) pipeline for automatically generating a scientific presentation script. It integrates paper retrieval from arXiv, text extraction and chunking, embedding generation with HuggingFace, vector storage with ChromaDB, and context-aware generation using LLMs. It also integrates Galileo Prompt Quality for evaluation and logging, and supports multi-source model loading, including local Llama.cpp, HuggingFace-hosted, and HuggingFace-cloud models like Mistral or DeepSeek.
This project demonstrates how to build a semantic chunking and summarization pipeline for texts using LangChain, Sentence Transformers, and Galileo for model evaluation, protection, and observability. It leverages the Z by HP AI Studio Local GenAI image and the Meta Llama 3.1 model with 8B parameters to generate concise and contextually accurate summaries from text data.
This project is an AI-powered vanilla RAG (Retrieval-Augmented Generation) chatbot built using LangChain and Galileo for model evaluation, protection, and observability. It leverages the Z by HP AI Studio Local GenAI image and the Meta Llama 3.1 model with 8B parameters to generate contextual and document-grounded answers to user queries about Z by HP AI Studio.
The blueprint projects in this folder demonstrate how to integrate NVIDIA NGC (NVIDIA GPU Cloud) resources with HP AI Studio.
We provide 5 blueprint projects, each designed for quick and easy use to help you get started efficiently.
This project contains a single integrated pipeline—Agentic RAG for AI Studio with TRT-LLM and LangGraph—that implements a Retrieval-Augmented Generation (RAG) workflow using:
TensorRT-backed Llama-3.1-Nano (TRT-LLM): for fast, GPU-accelerated inference. LangGraph: to orchestrate an agentic, multi-step decision flow (relevance check, memory lookup, query rewriting, retrieval, answer generation, and memory update). ChromaDB: as a local vector store over Markdown context files (about AI Studio). SimpleKVMemory: a lightweight on-disk key-value store to cache query-answer pairs.
This project demonstrates an end-to-end audio translation pipeline using NVIDIA NeMo models. It takes an English audio sample and performs:
- Speech-to-Text (STT) conversion using Citrinet
- Text Translation (TT) from English to Spanish using NMT
- Text-to-Speech (TTS) synthesis in Spanish using FastPitch and HiFiGAN
All steps are GPU-accelerated, and the full workflow is integrated with MLflow for experiment tracking and model registration.
In this project, we provide notebooks to compare the execution time of dataset operations using traditional Pandas (CPU) versus NVIDIA’s cuDF, a GPU-accelerated drop-in replacement for Pandas. This example is presented in two different formats:
-
Original Example Notebook: This version, created by NVIDIA, runs the entire evaluation within a single notebook. It includes downloading the data and restarting the kernel to activate the cuDF extension.
-
Data Analysis Notebooks: These notebooks use preprocessed datasets of varying sizes from the datafabric folder in AI Studio. The evaluation is split across two notebooks—one using Pandas (CPU) and the other using cuDF (GPU)—with performance metrics logged to MLflow.
This project is a GPU-accelerated, interactive exploratory data analysis (EDA) dashboard for the OpenCellID dataset. It uses Panel and cuDF to deliver lightning-fast geospatial analysis and visualization.
You can explore cell tower distributions by radio type, operator, country, and time window — rendered live on an interactive map with full GPU acceleration.
This project implements an AI-powered recommendation agent that delivers personalized travel suggestions based on user queries.
It leverages the NVIDIA NeMo Framework and BERT embeddings to understand user intent and generate highly relevant, tailored vacation recommendations.
This section provides solutions for common issues users may encounter when working with AI Blueprint projects in HP AI Studio:
-
Check Hardware Compatibility Each project’s README includes recommended minimum hardware specifications (e.g., RAM, VRAM). Make sure your system meets these requirements—especially when working with large models or during deployment, as insufficient resources can cause failures.
-
Models or Datasets Not Visible After Download If you download models or datasets while your workspace is running, they might not appear in the workspace. In such cases, restart your workspace to ensure they are properly recognized.
-
Connection or SSL Errors in Notebooks If you encounter SSL certificate or connection errors while accessing websites from notebooks (especially on restricted networks), verify your network settings. Consider using a proxy to bypass restrictive network constraints.
-
File or Path Not Found Errors Ensure that all required files and directories are correctly placed as specified in the project’s README. If any paths or files are missing, create or move them to the correct locations.
-
GPU Not Available For projects requiring NVIDIA GPUs, verify GPU availability by running
nvidia-smi
in the terminal. Ensure that a compatible GPU is accessible and has sufficient free memory to run the project. -
Deployment Errors Despite Meeting Requirements Even if your hardware meets the specs, limited available RAM or VRAM can cause deployment issues. Close other running workspaces or programs to free up memory.
-
API Timeout Issues API requests triggered for the local MLflow deployment in the AI Studio have a response timeout limit (approximately 4 minutes). For long-running tasks or large inputs, use the provided notebooks instead of the API to avoid timeout errors.
-
Large Datasets If your dataset is too large to fit into RAM, avoid using libraries like pandas, which rely on eager loading. Instead, use libraries that support lazy loading (processing data in chunks on demand), such as Polars or DuckDB. These tools are designed to handle large-scale data more efficiently without exhausting system memory.
-
Troubleshooting: Refer to the Troubleshooting section of the main README in our public AI-Blueprints GitHub repo for solutions to common issues.
-
Issues & Bugs: Open a new issue in our AI-Blueprints GitHub repo.
-
Docs: AI Studio Documentation.
-
Community: Join the HP AI Creator Community for questions and help.
Built with ❤️ using HP AI Studio.
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for AI-Blueprints
Similar Open Source Tools

AI-Blueprints
This repository hosts a collection of AI blueprint projects for HP AI Studio, providing end-to-end solutions across key AI domains like data science, machine learning, deep learning, and generative AI. The projects are designed to be plug-and-play, utilizing open-source and hosted models to offer ready-to-use solutions. The repository structure includes projects related to classical machine learning, deep learning applications, generative AI, NGC integration, and troubleshooting guidelines for common issues. Each project is accompanied by detailed descriptions and use cases, showcasing the versatility and applicability of AI technologies in various domains.

kodit
Kodit is a Code Indexing MCP Server that connects AI coding assistants to external codebases, providing accurate and up-to-date code snippets. It improves AI-assisted coding by offering canonical examples, indexing local and public codebases, integrating with AI coding assistants, enabling keyword and semantic search, and supporting OpenAI-compatible or custom APIs/models. Kodit helps engineers working with AI-powered coding assistants by providing relevant examples to reduce errors and hallucinations.

Simplifine
Simplifine is an open-source library designed for easy LLM finetuning, enabling users to perform tasks such as supervised fine tuning, question-answer finetuning, contrastive loss for embedding tasks, multi-label classification finetuning, and more. It provides features like WandB logging, in-built evaluation tools, automated finetuning parameters, and state-of-the-art optimization techniques. The library offers bug fixes, new features, and documentation updates in its latest version. Users can install Simplifine via pip or directly from GitHub. The project welcomes contributors and provides comprehensive documentation and support for users.

echosharp
EchoSharp is an open-source library designed for near-real-time audio processing, orchestrating different AI models seamlessly for various audio analysis scopes. It focuses on flexibility and performance, allowing near-real-time Transcription and Translation by integrating components for Speech-to-Text and Voice Activity Detection. With interchangeable components, easy orchestration, and first-party components like Whisper.net, SileroVad, OpenAI Whisper, AzureAI SpeechServices, WebRtcVadSharp, Onnx.Whisper, and Onnx.Sherpa, EchoSharp provides efficient audio analysis solutions for developers.

policy-synth
Policy Synth is a TypeScript class library that empowers better decision-making for governments and companies by integrating collective and artificial intelligence. It streamlines processes through multi-scale AI agent logic flows, robust APIs, and cutting-edge real-time AI-driven web applications. The tool supports organizations in generating, refining, and implementing smarter, data-informed strategies, fostering collaboration with AI to tackle complex challenges effectively.

AgentForge
AgentForge is a low-code framework tailored for the rapid development, testing, and iteration of AI-powered autonomous agents and Cognitive Architectures. It is compatible with a range of LLM models and offers flexibility to run different models for different agents based on specific needs. The framework is designed for seamless extensibility and database-flexibility, making it an ideal playground for various AI projects. AgentForge is a beta-testing ground and future-proof hub for crafting intelligent, model-agnostic autonomous agents.

ChatFAQ
ChatFAQ is an open-source comprehensive platform for creating a wide variety of chatbots: generic ones, business-trained, or even capable of redirecting requests to human operators. It includes a specialized NLP/NLG engine based on a RAG architecture and customized chat widgets, ensuring a tailored experience for users and avoiding vendor lock-in.

Awesome-AI-Agents-HUB-for-CrewAI
A comprehensive repository featuring a curated collection of AI-powered projects and Multi Agent Systems (MAS) built with the Crew AI framework. It provides innovative AI solutions for various domains, including marketing automation, health planning, legal advice, and more. Users can explore and deploy AI agents, Multi Agent Systems, and advanced machine learning techniques through a diverse selection of projects that leverage state-of-the-art AI technologies like RAG (Retrieval-Augmented Generation). The projects offer practical applications and customizable solutions for integrating AI into existing workflows or new projects.

cline-based-code-generator
HAI Code Generator is a cutting-edge tool designed to simplify and automate task execution while enhancing code generation workflows. Leveraging Specif AI, it streamlines processes like task execution, file identification, and code documentation through intelligent automation and AI-driven capabilities. Built on Cline's powerful foundation for AI-assisted development, HAI Code Generator boosts productivity and precision by automating task execution and integrating file management capabilities. It combines intelligent file indexing, context generation, and LLM-driven automation to minimize manual effort and ensure task accuracy. Perfect for developers and teams aiming to enhance their workflows.

refact-vscode
Refact.ai is an open-source AI coding assistant that boosts developer's productivity. It supports 25+ programming languages and offers features like code completion, AI Toolbox for code explanation and refactoring, integrated in-IDE chat, and self-hosting or cloud version. The Enterprise plan provides enhanced customization, security, fine-tuning, user statistics, efficient inference, priority support, and access to 20+ LLMs for up to 50 engineers per GPU.

postgresml
PostgresML is a powerful Postgres extension that seamlessly combines data storage and machine learning inference within your database. It enables running machine learning and AI operations directly within PostgreSQL, leveraging GPU acceleration for faster computations, integrating state-of-the-art large language models, providing built-in functions for text processing, enabling efficient similarity search, offering diverse ML algorithms, ensuring high performance, scalability, and security, supporting a wide range of NLP tasks, and seamlessly integrating with existing PostgreSQL tools and client libraries.

learnhouse
LearnHouse is an open-source platform that allows anyone to easily provide world-class educational content. It supports various content types, including dynamic pages, videos, and documents. The platform is still in early development and should not be used in production environments. However, it offers several features, such as dynamic Notion-like pages, ease of use, multi-organization support, support for uploading videos and documents, course collections, user management, quizzes, course progress tracking, and an AI-powered assistant for teachers and students. LearnHouse is built using various open-source projects, including Next.js, TailwindCSS, Radix UI, Tiptap, FastAPI, YJS, PostgreSQL, LangChain, and React.

codegate
CodeGate is a local gateway that enhances the safety of AI coding assistants by ensuring AI-generated recommendations adhere to best practices, safeguarding code integrity, and protecting individual privacy. Developed by Stacklok, CodeGate allows users to confidently leverage AI in their development workflow without compromising security or productivity. It works seamlessly with coding assistants, providing real-time security analysis of AI suggestions. CodeGate is designed with privacy at its core, keeping all data on the user's machine and offering complete control over data.

ProjectAirSim
Project AirSim is a simulation platform for drones, robots, and autonomous systems. Leveraging Unreal Engine 5, it provides photo-realistic visuals and a simulation framework for custom physics, controllers, actuators, and sensors. It consists of three main layers: Sim Libs, Plugin, and Client Library. It supports Windows 11 and Ubuntu 22, inviting collaboration and enterprise support. Users can join the community, contribute to the roadmap, and get started with pre-built binaries or building from source. It offers headless running options and references for configuration settings, API, controllers, sensors, scene, physics, and FAQ.

heurist-agent-framework
Heurist Agent Framework is a flexible multi-interface AI agent framework that allows processing text and voice messages, generating images and videos, interacting across multiple platforms, fetching and storing information in a knowledge base, accessing external APIs and tools, and composing complex workflows using Mesh Agents. It supports various platforms like Telegram, Discord, Twitter, Farcaster, REST API, and MCP. The framework is built on a modular architecture and provides core components, tools, workflows, and tool integration with MCP support.

comfyui_LLM_Polymath
LLM Polymath Chat Node is an advanced Chat Node for ComfyUI that integrates large language models to build text-driven applications and automate data processes, enhancing prompt responses by incorporating real-time web search, linked content extraction, and custom agent instructions. It supports both OpenAI’s GPT-like models and alternative models served via a local Ollama API. The core functionalities include Comfy Node Finder and Smart Assistant, along with additional agents like Flux Prompter, Custom Instructors, Python debugger, and scripter. The tool offers features for prompt processing, web search integration, model & API integration, custom instructions, image handling, logging & debugging, output compression, and more.
For similar tasks

Azure-Analytics-and-AI-Engagement
The Azure-Analytics-and-AI-Engagement repository provides packaged Industry Scenario DREAM Demos with ARM templates (Containing a demo web application, Power BI reports, Synapse resources, AML Notebooks etc.) that can be deployed in a customer’s subscription using the CAPE tool within a matter of few hours. Partners can also deploy DREAM Demos in their own subscriptions using DPoC.

sorrentum
Sorrentum is an open-source project that aims to combine open-source development, startups, and brilliant students to build machine learning, AI, and Web3 / DeFi protocols geared towards finance and economics. The project provides opportunities for internships, research assistantships, and development grants, as well as the chance to work on cutting-edge problems, learn about startups, write academic papers, and get internships and full-time positions at companies working on Sorrentum applications.

tidb
TiDB is an open-source distributed SQL database that supports Hybrid Transactional and Analytical Processing (HTAP) workloads. It is MySQL compatible and features horizontal scalability, strong consistency, and high availability.

zep-python
Zep is an open-source platform for building and deploying large language model (LLM) applications. It provides a suite of tools and services that make it easy to integrate LLMs into your applications, including chat history memory, embedding, vector search, and data enrichment. Zep is designed to be scalable, reliable, and easy to use, making it a great choice for developers who want to build LLM-powered applications quickly and easily.

telemetry-airflow
This repository codifies the Airflow cluster that is deployed at workflow.telemetry.mozilla.org (behind SSO) and commonly referred to as "WTMO" or simply "Airflow". Some links relevant to users and developers of WTMO: * The `dags` directory in this repository contains some custom DAG definitions * Many of the DAGs registered with WTMO don't live in this repository, but are instead generated from ETL task definitions in bigquery-etl * The Data SRE team maintains a WTMO Developer Guide (behind SSO)

mojo
Mojo is a new programming language that bridges the gap between research and production by combining Python syntax and ecosystem with systems programming and metaprogramming features. Mojo is still young, but it is designed to become a superset of Python over time.

pandas-ai
PandasAI is a Python library that makes it easy to ask questions to your data in natural language. It helps you to explore, clean, and analyze your data using generative AI.

databend
Databend is an open-source cloud data warehouse that serves as a cost-effective alternative to Snowflake. With its focus on fast query execution and data ingestion, it's designed for complex analysis of the world's largest datasets.
For similar jobs

promptflow
**Prompt flow** is a suite of development tools designed to streamline the end-to-end development cycle of LLM-based AI applications, from ideation, prototyping, testing, evaluation to production deployment and monitoring. It makes prompt engineering much easier and enables you to build LLM apps with production quality.

deepeval
DeepEval is a simple-to-use, open-source LLM evaluation framework specialized for unit testing LLM outputs. It incorporates various metrics such as G-Eval, hallucination, answer relevancy, RAGAS, etc., and runs locally on your machine for evaluation. It provides a wide range of ready-to-use evaluation metrics, allows for creating custom metrics, integrates with any CI/CD environment, and enables benchmarking LLMs on popular benchmarks. DeepEval is designed for evaluating RAG and fine-tuning applications, helping users optimize hyperparameters, prevent prompt drifting, and transition from OpenAI to hosting their own Llama2 with confidence.

MegaDetector
MegaDetector is an AI model that identifies animals, people, and vehicles in camera trap images (which also makes it useful for eliminating blank images). This model is trained on several million images from a variety of ecosystems. MegaDetector is just one of many tools that aims to make conservation biologists more efficient with AI. If you want to learn about other ways to use AI to accelerate camera trap workflows, check out our of the field, affectionately titled "Everything I know about machine learning and camera traps".

leapfrogai
LeapfrogAI is a self-hosted AI platform designed to be deployed in air-gapped resource-constrained environments. It brings sophisticated AI solutions to these environments by hosting all the necessary components of an AI stack, including vector databases, model backends, API, and UI. LeapfrogAI's API closely matches that of OpenAI, allowing tools built for OpenAI/ChatGPT to function seamlessly with a LeapfrogAI backend. It provides several backends for various use cases, including llama-cpp-python, whisper, text-embeddings, and vllm. LeapfrogAI leverages Chainguard's apko to harden base python images, ensuring the latest supported Python versions are used by the other components of the stack. The LeapfrogAI SDK provides a standard set of protobuffs and python utilities for implementing backends and gRPC. LeapfrogAI offers UI options for common use-cases like chat, summarization, and transcription. It can be deployed and run locally via UDS and Kubernetes, built out using Zarf packages. LeapfrogAI is supported by a community of users and contributors, including Defense Unicorns, Beast Code, Chainguard, Exovera, Hypergiant, Pulze, SOSi, United States Navy, United States Air Force, and United States Space Force.

llava-docker
This Docker image for LLaVA (Large Language and Vision Assistant) provides a convenient way to run LLaVA locally or on RunPod. LLaVA is a powerful AI tool that combines natural language processing and computer vision capabilities. With this Docker image, you can easily access LLaVA's functionalities for various tasks, including image captioning, visual question answering, text summarization, and more. The image comes pre-installed with LLaVA v1.2.0, Torch 2.1.2, xformers 0.0.23.post1, and other necessary dependencies. You can customize the model used by setting the MODEL environment variable. The image also includes a Jupyter Lab environment for interactive development and exploration. Overall, this Docker image offers a comprehensive and user-friendly platform for leveraging LLaVA's capabilities.

carrot
The 'carrot' repository on GitHub provides a list of free and user-friendly ChatGPT mirror sites for easy access. The repository includes sponsored sites offering various GPT models and services. Users can find and share sites, report errors, and access stable and recommended sites for ChatGPT usage. The repository also includes a detailed list of ChatGPT sites, their features, and accessibility options, making it a valuable resource for ChatGPT users seeking free and unlimited GPT services.

TrustLLM
TrustLLM is a comprehensive study of trustworthiness in LLMs, including principles for different dimensions of trustworthiness, established benchmark, evaluation, and analysis of trustworthiness for mainstream LLMs, and discussion of open challenges and future directions. Specifically, we first propose a set of principles for trustworthy LLMs that span eight different dimensions. Based on these principles, we further establish a benchmark across six dimensions including truthfulness, safety, fairness, robustness, privacy, and machine ethics. We then present a study evaluating 16 mainstream LLMs in TrustLLM, consisting of over 30 datasets. The document explains how to use the trustllm python package to help you assess the performance of your LLM in trustworthiness more quickly. For more details about TrustLLM, please refer to project website.

AI-YinMei
AI-YinMei is an AI virtual anchor Vtuber development tool (N card version). It supports fastgpt knowledge base chat dialogue, a complete set of solutions for LLM large language models: [fastgpt] + [one-api] + [Xinference], supports docking bilibili live broadcast barrage reply and entering live broadcast welcome speech, supports Microsoft edge-tts speech synthesis, supports Bert-VITS2 speech synthesis, supports GPT-SoVITS speech synthesis, supports expression control Vtuber Studio, supports painting stable-diffusion-webui output OBS live broadcast room, supports painting picture pornography public-NSFW-y-distinguish, supports search and image search service duckduckgo (requires magic Internet access), supports image search service Baidu image search (no magic Internet access), supports AI reply chat box [html plug-in], supports AI singing Auto-Convert-Music, supports playlist [html plug-in], supports dancing function, supports expression video playback, supports head touching action, supports gift smashing action, supports singing automatic start dancing function, chat and singing automatic cycle swing action, supports multi scene switching, background music switching, day and night automatic switching scene, supports open singing and painting, let AI automatically judge the content.