Muice-Chatbot
沐雪,一个会自动找你聊天的AI女孩子
Stars: 247
Muice-Chatbot is an AI chatbot designed to proactively engage in conversations with users. It is based on the ChatGLM2-6B and Qwen-7B models, with a training dataset of 1.8K+ dialogues. The chatbot has a speaking style similar to a 2D girl, being somewhat tsundere but willing to share daily life details and greet users differently every day. It provides various functionalities, including initiating chats and offering 5 available commands. The project supports model loading through different methods and provides onebot service support for QQ users. Users can interact with the chatbot by running the main.py file in the project directory.
README:
沐雪,一只会主动找你聊天的AI女孩子,其对话模型基于ChatGLM2-6B与Qwen微调而成,训练集长度3K+ ,具有二次元女孩子的说话风格,比较傲娇,但乐于和你分享生活的琐碎,每天会给你不一样的问候。
✅ 支持近乎全自动安装环境
✅ 提供本人由3k+对话数据微调的ChatGLM2-6B P-Tuning V2模型与Qwen Qlora微调模型
✅ 主动发起聊天(随机和每天早中晚固定时间)
✅ 提供5条命令以便在聊天中进行刷新回复等操作
✅ OFA图像识别:识别表情包、理解表情包、发送表情包
✅ 支持通过fishaudio/fish-speech进行语言合成(沐雪tts模型尚未发布)
✅ 在群聊中聊天(支持被 @ 回复或不被 @ 随机回复)
✅ 在控制台中实时对话(暂不支持打QQ电话)
⬜ 对记忆模块进行优化,实现长期记忆与短期记忆
建议环境:
- Python 3.10
- 一张拥有6GB+ 显存的显卡(int4量化最低要求: 4G ; CPU推理内存要求:16G+)
目前已做到自动安装所有软件、依赖,通过code->download zip下载解压最新源码。
双击install_env.bat
安装(不能启用旧版控制台),或在命令行中运行以下命令:
.\install_env.bat
自动安装可能需要较长时间,请耐心等待,安装完成后,你仍需手动下载模型。
请注意,自动安装脚本使用的是python的虚拟环境,不需要conda,留意安装脚本的提示。
git clone https://github.com/Moemu/Muice-Chatbot
cd Muice-Chatbot
conda create --name Muice python=3.10.10
conda activate Muice
pip install -r requirements.txt
目前支持的基底模型如下表:
基底模型 | 对应微调模型版本号 | 额外依赖库 |
---|---|---|
ChatGLM2-6B-Int4 | 2.2-2.4 | cpm_kernels |
ChatGLM2-6B | 2.0-2.3 | |
Qwen-7B-Chat-Int4 | 2.3、2.6.2 | llmtuner |
Qwen2-1.5B-Instruct-GPTQ-Int4 | 2.5.3 | llmtuner |
RWKV(Seikaijyu微调) | 参见HF | (需要下载配置RWKV-Runner) |
微调模型下载:Releases
请将基底模型与微调模型放放入model
文件夹中(确保微调模型目录下存在.model文件而不是文件夹,部分微调模型由于疏忽还套了一层checkpoint文件夹)
本仓库目前支持如下模型加载方式:
- 通过API加载
- 通过transformers的
AutoTokenizer
,AutoModel
函数加载 - 通过
llmtuner.chat
(LLaMA-Factory
)的ChatModel
类加载 - 通过
RWKV-Runner
提供的API服务加载
在已测试的模型中,我们建议以下模型通过对应的方式加载,其他模型亦可以通过类似的方式加载:
基底模型 | 微调方式 | 加载方法 |
---|---|---|
ChatGLM | P-tuning V2 | transformers |
Qwen | sft | llmtuner |
RWKV(Seikaijyu微调) | pissa | rwkv-api |
在配置文件中可调整模型的加载方式:
"model_loader": "api/transformers/llmtuner/rwkv-api",
"model_name_or_path": "基底模型位置",
"adapter_name_or_path": "沐雪微调模型位置"
(若是API/rwkv-api加载,model_name_or_path
填api地址)
如果你没有合适的显卡,需要通过CPU加载模型,请安装配置gcc环境并勾选openmp.
本项目支持OFA图像识别,即对聊天图片进行特征提取,并通过OFA模型进行描述生成。发送信息时,会检索本地图片库,若存在匹配的图片,则会发送图片。
若您希望使用OFA图像识别,请在配置文件中设置enable_ofa_image
为true
,并设置ofa_image_model_name_or_path
为OFA图像识别模型的路径。
OFA使用到的依赖需要额外安装:
conda activate Muice
pip install -r ofa_requirements.txt
目前支持的OFA模型如下表:
OFA模型 |
---|
OFA-Image-Caption-Meme-Large-ZH (建议) |
ofa_image-caption_muge_base_zh |
本地图片数据库位于image_data
文件夹中,程序会每五分钟备份一次数据库,默认保留5个备份槽位。
若您想要回退数据库到某个时间点,请将image_data
文件夹中的image_data.db
删除,并将备份槽位文件image_data.db.backup_xxx_xxx
修改为image_data.db
。
若您希望使用语音回复,请在配置文件中设置Voice_Reply_Rate
为大于0的整数,机器人将会以一定概率回复语音消息。
语音回复使用到的项目:fishaudio/fish-speech
在fish-speech的webui启动(使用--infer参数)后,更改fish_speech_api.py
中的Client
、reference_audio
、reference_text
即可。
-
Client
为fish-speech的webui地址 -
reference_audio
为参考音频文件路径,此音频用于变声效果。 -
reference_text
为参考音频文件的参考文本。
1.安装依赖:
conda activate Muice
pip install -r audio_requirements.txt
2.安装配置语音回复(详见上文)
3.获取语音识别模型
目前支持的模型如下表:
语音识别模型 |
---|
SenseVoice多语言语音理解模型Small |
你可以通过以下命令下载并解压模型:
modelscope download --model iic/SenseVoiceSmall --local_path ./SenseVoice
下载完成后,在配置文件中设置audio_name_or_path
为模型文件夹的路径。
4.配置信息和设备
你可以通过以下命令查看输入输出设备信息:
python realtime_refence.py --get_device
在realtime_refence.py
中配置输入输出设备信息:
CHUNK = 1024 # 每次读取的音频块大小
FORMAT = pyaudio.paFloat32 # 音频格式
CHANNELS = 1 # 输入设备声道
RATE = 22050 # 采样率(16000/22050/44100)
THRESHOLD = 75 # 声音响度阈值(60-150左右,请根据实际情况调节)
SILENCE_THRESHOLD_MS = 1500 # 静音持续时间阈值(毫秒)
SILENCE_COUNT = int(SILENCE_THRESHOLD_MS / (1000 * CHUNK / RATE)) # 静音计数器阈值
use_virtual_device = False # 是否使用虚拟设备(当你需要通过语音通话时,请设置为True)
if use_virtual_device:
speaker_device_index = 3 # 输入设备索引
mic_device_index = 10 # 输出设备索引
device_index = speaker_device_index
else:
device_index = 1 # 录音设备索引
4.启动实时语音聊天
python realtime_refence.py
现以提供onebot服务支持, 无需担心gocq的风控(喜)
本项目使用onebotV11协议, 若您希望于QQ使用, 推荐参考LLOneBot使用onebot服务
注:请在安装好LLOneBot后, 于设置中开启反向WebSocket服务, 填写ws://127.0.0.1:21050/ws/api
, 以正常运行
您也可以使用Lagrange.Core以及OpenShamrock等来链接QQ, 或其他适配器链接其他软件,详见onebotV11适配器
能使用请勿随意更新ntQQ, 若无法使用请尝试降级ntQQ
在完成这些操作后,你应该得到类似如下所示的文件结构:
Muice-Chatbot <- 主路径
├─llm
├─model
│ ├─ chatglm2-6b <- 原始模型 (三者三选一)
│ ├─ chatglm2-6b-int4 <- int4原始模型
│ ├─ Qwen-7B-Chat-Int4 <- Qwen-7B-int4原始模型
│ └─ Muice
├─configs.json <- 配置文件
├─ws.py <- ws服务
├─main.py <- 主函数
├─requirements.txt
└─...
本项目使用configs.json
作为配置文件,目前支持配置的项如下:
{
"model_loader": "transformers",
"model_name_or_path": "./model/chatglm2-6b",
"adapter_name_or_path": "./model/Muice",
"enable_ofa_image": false,
"ofa_image_model_name_or_path": "",
"Trust_QQ_list": [],
"AutoCreateTopic": false,
"read_memory_from_file": true,
"known_topic_probability": 0.003,
"time_topic_probability": 0.75,
"port":21050,
"Reply_Wait": true,
"bot_qq_id":123456789,
"Is_CQ_Code": false,
"Group_Message_Reply_Only_To_Trusted": true,
"Reply_Rate": 50,
"At_Reply": false,
"NonReply_Prefix": [],
"Voice_Reply_Rate": 0
}
model_loader
: 指定模型加载器的类型,当前支持api/transformers/llmtuner/rwkv-api
。
model_name_or_path
: 指定基底模型的名称或路径,例如./model/chatglm2-6b
。
adapter_name_or_path
: 指定预训练模型的名称或路径, 例如./model/Muice
。
enable_ofa_image
: 是否使用OFA图像识别。
ofa_image_model_name_or_path
: OFA图像识别模型的名称或路径。
Trust_QQ_list
: 信任QQ号列表,只有在此列表中的QQ号发送的消息,机器人才会回复。
AutoCreateTopic
: 是否自动发起新对话。如果启用,将默认以Trust_QQ_list中的第一个QQ号作为对话发起对象。
read_memory_from_file
: 是否从文件中读取记忆。这对于项目重启后恢复之前的对话状态非常有用。
known_topic_probability
: 随机发起已知话题的概率。
time_topic_probability
: 根据时间(早、中、傍、晚)触发日常问候的概率。
port
: 反向WebSocket服务的端口号,默认21050
。
Reply_Wait
: (私聊)是否回复时等待一段时间。
bot_qq_id
: 机器人的QQ号。
Is_CQ_Code
: 是否启用CQ码处理信息。
Group_Message_Reply_Only_To_Trusted
: (群聊)是否仅对信任的qq回复。
Reply_Rate
: (群聊)机器人回复的概率,取值范围为0-100。
At_Reply
: (群聊)是否只回复@机器人的消息。
NonReply_Prefix
: 消息前缀,机器人不会回复以这些前缀开头的消息。
Voice_Reply_Rate
: 语音回复的概率,取值范围为0-100。
在本项目根目录下运行 main.py
conda activate Muice
python main.py
命令 | 释义 |
---|---|
/clean | 清空本轮对话历史 |
/refresh | 刷新本次对话 |
/help | 显示所有可用的命令列表 |
/reset | 重置所有对话数据(将存档对话数据) |
/undo | 撤销上一次对话 |
参见公开的训练集Moemu/Muice-Dataset
与其他聊天机器人项目不同,本项目提供由本人通过自家对话数据集微调后的模型,在Release中提供下载,关于微调后的模型人设,目前公开的信息如下:
训练集开源地址: Moemu/Muice-Dataset
原始模型:THUDM/ChatGLM2-6B & QwenLM/Qwen)
本项目源码使用MIT license,对于微调后的模型文件,不建议作为商业用途
代码实现:Moemu、MoeSnowyFox、NaivG
训练集编写与模型微调:Moemu (RWKV微调:Seikaijyu)
总代码贡献:
Star History:
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for Muice-Chatbot
Similar Open Source Tools
Muice-Chatbot
Muice-Chatbot is an AI chatbot designed to proactively engage in conversations with users. It is based on the ChatGLM2-6B and Qwen-7B models, with a training dataset of 1.8K+ dialogues. The chatbot has a speaking style similar to a 2D girl, being somewhat tsundere but willing to share daily life details and greet users differently every day. It provides various functionalities, including initiating chats and offering 5 available commands. The project supports model loading through different methods and provides onebot service support for QQ users. Users can interact with the chatbot by running the main.py file in the project directory.
api-for-open-llm
This project provides a unified backend interface for open large language models (LLMs), offering a consistent experience with OpenAI's ChatGPT API. It supports various open-source LLMs, enabling developers to seamlessly integrate them into their applications. The interface features streaming responses, text embedding capabilities, and support for LangChain, a tool for developing LLM-based applications. By modifying environment variables, developers can easily use open-source models as alternatives to ChatGPT, providing a cost-effective and customizable solution for various use cases.
moonpalace
MoonPalace is a debugging tool for API provided by Moonshot AI. It supports all platforms (Mac, Windows, Linux) and is simple to use by replacing 'base_url' with 'http://localhost:9988'. It captures complete requests, including 'accident scenes' during network errors, and allows quick retrieval and viewing of request information using 'request_id' and 'chatcmpl_id'. It also enables one-click export of BadCase structured reporting data to help improve Kimi model capabilities. MoonPalace is recommended for use as an API 'supplier' during code writing and debugging stages to quickly identify and locate various issues related to API calls and code writing processes, and to export request details for submission to Moonshot AI to improve Kimi model.
Langchain-Chatchat
LangChain-Chatchat is an open-source, offline-deployable retrieval-enhanced generation (RAG) large model knowledge base project based on large language models such as ChatGLM and application frameworks such as Langchain. It aims to establish a knowledge base Q&A solution that is friendly to Chinese scenarios, supports open-source models, and can run offline.
Chinese-Mixtral-8x7B
Chinese-Mixtral-8x7B is an open-source project based on Mistral's Mixtral-8x7B model for incremental pre-training of Chinese vocabulary, aiming to advance research on MoE models in the Chinese natural language processing community. The expanded vocabulary significantly improves the model's encoding and decoding efficiency for Chinese, and the model is pre-trained incrementally on a large-scale open-source corpus, enabling it with powerful Chinese generation and comprehension capabilities. The project includes a large model with expanded Chinese vocabulary and incremental pre-training code.
litellm
LiteLLM is a tool that allows you to call all LLM APIs using the OpenAI format. This includes Bedrock, Huggingface, VertexAI, TogetherAI, Azure, OpenAI, and more. LiteLLM manages translating inputs to provider's `completion`, `embedding`, and `image_generation` endpoints, providing consistent output, and retry/fallback logic across multiple deployments. It also supports setting budgets and rate limits per project, api key, and model.
tiny-llm-zh
Tiny LLM zh is a project aimed at building a small-parameter Chinese language large model for quick entry into learning large model-related knowledge. The project implements a two-stage training process for large models and subsequent human alignment, including tokenization, pre-training, instruction fine-tuning, human alignment, evaluation, and deployment. It is deployed on ModeScope Tiny LLM website and features open access to all data and code, including pre-training data and tokenizer. The project trains a tokenizer using 10GB of Chinese encyclopedia text to build a Tiny LLM vocabulary. It supports training with Transformers deepspeed, multiple machine and card support, and Zero optimization techniques. The project has three main branches: llama2_torch, main tiny_llm, and tiny_llm_moe, each with specific modifications and features.
ddddocr
ddddocr is a Rust version of a simple OCR API server that provides easy deployment for captcha recognition without relying on the OpenCV library. It offers a user-friendly general-purpose captcha recognition Rust library. The tool supports recognizing various types of captchas, including single-line text, transparent black PNG images, target detection, and slider matching algorithms. Users can also import custom OCR training models and utilize the OCR API server for flexible OCR result control and range limitation. The tool is cross-platform and can be easily deployed.
gollama
Gollama is a tool designed for managing Ollama models through a Text User Interface (TUI). Users can list, inspect, delete, copy, and push Ollama models, as well as link them to LM Studio. The application offers interactive model selection, sorting by various criteria, and actions using hotkeys. It provides features like sorting and filtering capabilities, displaying model metadata, model linking, copying, pushing, and more. Gollama aims to be user-friendly and useful for managing models, especially for cleaning up old models.
app-builder
AppBuilder SDK is a one-stop development tool for AI native applications, providing basic cloud resources, AI capability engine, Qianfan large model, and related capability components to improve the development efficiency of AI native applications.
chatglm.cpp
ChatGLM.cpp is a C++ implementation of ChatGLM-6B, ChatGLM2-6B, ChatGLM3-6B and more LLMs for real-time chatting on your MacBook. It is based on ggml, working in the same way as llama.cpp. ChatGLM.cpp features accelerated memory-efficient CPU inference with int4/int8 quantization, optimized KV cache and parallel computing. It also supports P-Tuning v2 and LoRA finetuned models, streaming generation with typewriter effect, Python binding, web demo, api servers and more possibilities.
nexa-sdk
Nexa SDK is a comprehensive toolkit supporting ONNX and GGML models for text generation, image generation, vision-language models (VLM), and text-to-speech (TTS) capabilities. It offers an OpenAI-compatible API server with JSON schema mode and streaming support, along with a user-friendly Streamlit UI. Users can run Nexa SDK on any device with Python environment, with GPU acceleration supported. The toolkit provides model support, conversion engine, inference engine for various tasks, and differentiating features from other tools.
Streamer-Sales
Streamer-Sales is a large model for live streamers that can explain products based on their characteristics and inspire users to make purchases. It is designed to enhance sales efficiency and user experience, whether for online live sales or offline store promotions. The model can deeply understand product features and create tailored explanations in vivid and precise language, sparking user's desire to purchase. It aims to revolutionize the shopping experience by providing detailed and unique product descriptions to engage users effectively.
ChatTTS-Forge
ChatTTS-Forge is a powerful text-to-speech generation tool that supports generating rich audio long texts using a SSML-like syntax and provides comprehensive API services, suitable for various scenarios. It offers features such as batch generation, support for generating super long texts, style prompt injection, full API services, user-friendly debugging GUI, OpenAI-style API, Google-style API, support for SSML-like syntax, speaker management, style management, independent refine API, text normalization optimized for ChatTTS, and automatic detection and processing of markdown format text. The tool can be experienced and deployed online through HuggingFace Spaces, launched with one click on Colab, deployed using containers, or locally deployed after cloning the project, preparing models, and installing necessary dependencies.
Llama-Chinese
Llama中文社区是一个专注于Llama模型在中文方面的优化和上层建设的高级技术社区。 **已经基于大规模中文数据,从预训练开始对Llama2模型进行中文能力的持续迭代升级【Done】**。**正在对Llama3模型进行中文能力的持续迭代升级【Doing】** 我们热忱欢迎对大模型LLM充满热情的开发者和研究者加入我们的行列。
Speech-AI-Forge
Speech-AI-Forge is a project developed around TTS generation models, implementing an API Server and a WebUI based on Gradio. The project offers various ways to experience and deploy Speech-AI-Forge, including online experience on HuggingFace Spaces, one-click launch on Colab, container deployment with Docker, and local deployment. The WebUI features include TTS model functionality, speaker switch for changing voices, style control, long text support with automatic text segmentation, refiner for ChatTTS native text refinement, various tools for voice control and enhancement, support for multiple TTS models, SSML synthesis control, podcast creation tools, voice creation, voice testing, ASR tools, and post-processing tools. The API Server can be launched separately for higher API throughput. The project roadmap includes support for various TTS models, ASR models, voice clone models, and enhancer models. Model downloads can be manually initiated using provided scripts. The project aims to provide inference services and may include training-related functionalities in the future.
For similar tasks
LLMStack
LLMStack is a no-code platform for building generative AI agents, workflows, and chatbots. It allows users to connect their own data, internal tools, and GPT-powered models without any coding experience. LLMStack can be deployed to the cloud or on-premise and can be accessed via HTTP API or triggered from Slack or Discord.
ai-guide
This guide is dedicated to Large Language Models (LLMs) that you can run on your home computer. It assumes your PC is a lower-end, non-gaming setup.
onnxruntime-genai
ONNX Runtime Generative AI is a library that provides the generative AI loop for ONNX models, including inference with ONNX Runtime, logits processing, search and sampling, and KV cache management. Users can call a high level `generate()` method, or run each iteration of the model in a loop. It supports greedy/beam search and TopP, TopK sampling to generate token sequences, has built in logits processing like repetition penalties, and allows for easy custom scoring.
jupyter-ai
Jupyter AI connects generative AI with Jupyter notebooks. It provides a user-friendly and powerful way to explore generative AI models in notebooks and improve your productivity in JupyterLab and the Jupyter Notebook. Specifically, Jupyter AI offers: * An `%%ai` magic that turns the Jupyter notebook into a reproducible generative AI playground. This works anywhere the IPython kernel runs (JupyterLab, Jupyter Notebook, Google Colab, Kaggle, VSCode, etc.). * A native chat UI in JupyterLab that enables you to work with generative AI as a conversational assistant. * Support for a wide range of generative model providers, including AI21, Anthropic, AWS, Cohere, Gemini, Hugging Face, NVIDIA, and OpenAI. * Local model support through GPT4All, enabling use of generative AI models on consumer grade machines with ease and privacy.
khoj
Khoj is an open-source, personal AI assistant that extends your capabilities by creating always-available AI agents. You can share your notes and documents to extend your digital brain, and your AI agents have access to the internet, allowing you to incorporate real-time information. Khoj is accessible on Desktop, Emacs, Obsidian, Web, and Whatsapp, and you can share PDF, markdown, org-mode, notion files, and GitHub repositories. You'll get fast, accurate semantic search on top of your docs, and your agents can create deeply personal images and understand your speech. Khoj is self-hostable and always will be.
langchain_dart
LangChain.dart is a Dart port of the popular LangChain Python framework created by Harrison Chase. LangChain provides a set of ready-to-use components for working with language models and a standard interface for chaining them together to formulate more advanced use cases (e.g. chatbots, Q&A with RAG, agents, summarization, extraction, etc.). The components can be grouped into a few core modules: * **Model I/O:** LangChain offers a unified API for interacting with various LLM providers (e.g. OpenAI, Google, Mistral, Ollama, etc.), allowing developers to switch between them with ease. Additionally, it provides tools for managing model inputs (prompt templates and example selectors) and parsing the resulting model outputs (output parsers). * **Retrieval:** assists in loading user data (via document loaders), transforming it (with text splitters), extracting its meaning (using embedding models), storing (in vector stores) and retrieving it (through retrievers) so that it can be used to ground the model's responses (i.e. Retrieval-Augmented Generation or RAG). * **Agents:** "bots" that leverage LLMs to make informed decisions about which available tools (such as web search, calculators, database lookup, etc.) to use to accomplish the designated task. The different components can be composed together using the LangChain Expression Language (LCEL).
danswer
Danswer is an open-source Gen-AI Chat and Unified Search tool that connects to your company's docs, apps, and people. It provides a Chat interface and plugs into any LLM of your choice. Danswer can be deployed anywhere and for any scale - on a laptop, on-premise, or to cloud. Since you own the deployment, your user data and chats are fully in your own control. Danswer is MIT licensed and designed to be modular and easily extensible. The system also comes fully ready for production usage with user authentication, role management (admin/basic users), chat persistence, and a UI for configuring Personas (AI Assistants) and their Prompts. Danswer also serves as a Unified Search across all common workplace tools such as Slack, Google Drive, Confluence, etc. By combining LLMs and team specific knowledge, Danswer becomes a subject matter expert for the team. Imagine ChatGPT if it had access to your team's unique knowledge! It enables questions such as "A customer wants feature X, is this already supported?" or "Where's the pull request for feature Y?"
infinity
Infinity is an AI-native database designed for LLM applications, providing incredibly fast full-text and vector search capabilities. It supports a wide range of data types, including vectors, full-text, and structured data, and offers a fused search feature that combines multiple embeddings and full text. Infinity is easy to use, with an intuitive Python API and a single-binary architecture that simplifies deployment. It achieves high performance, with 0.1 milliseconds query latency on million-scale vector datasets and up to 15K QPS.
For similar jobs
sweep
Sweep is an AI junior developer that turns bugs and feature requests into code changes. It automatically handles developer experience improvements like adding type hints and improving test coverage.
teams-ai
The Teams AI Library is a software development kit (SDK) that helps developers create bots that can interact with Teams and Microsoft 365 applications. It is built on top of the Bot Framework SDK and simplifies the process of developing bots that interact with Teams' artificial intelligence capabilities. The SDK is available for JavaScript/TypeScript, .NET, and Python.
ai-guide
This guide is dedicated to Large Language Models (LLMs) that you can run on your home computer. It assumes your PC is a lower-end, non-gaming setup.
classifai
Supercharge WordPress Content Workflows and Engagement with Artificial Intelligence. Tap into leading cloud-based services like OpenAI, Microsoft Azure AI, Google Gemini and IBM Watson to augment your WordPress-powered websites. Publish content faster while improving SEO performance and increasing audience engagement. ClassifAI integrates Artificial Intelligence and Machine Learning technologies to lighten your workload and eliminate tedious tasks, giving you more time to create original content that matters.
chatbot-ui
Chatbot UI is an open-source AI chat app that allows users to create and deploy their own AI chatbots. It is easy to use and can be customized to fit any need. Chatbot UI is perfect for businesses, developers, and anyone who wants to create a chatbot.
BricksLLM
BricksLLM is a cloud native AI gateway written in Go. Currently, it provides native support for OpenAI, Anthropic, Azure OpenAI and vLLM. BricksLLM aims to provide enterprise level infrastructure that can power any LLM production use cases. Here are some use cases for BricksLLM: * Set LLM usage limits for users on different pricing tiers * Track LLM usage on a per user and per organization basis * Block or redact requests containing PIIs * Improve LLM reliability with failovers, retries and caching * Distribute API keys with rate limits and cost limits for internal development/production use cases * Distribute API keys with rate limits and cost limits for students
uAgents
uAgents is a Python library developed by Fetch.ai that allows for the creation of autonomous AI agents. These agents can perform various tasks on a schedule or take action on various events. uAgents are easy to create and manage, and they are connected to a fast-growing network of other uAgents. They are also secure, with cryptographically secured messages and wallets.
griptape
Griptape is a modular Python framework for building AI-powered applications that securely connect to your enterprise data and APIs. It offers developers the ability to maintain control and flexibility at every step. Griptape's core components include Structures (Agents, Pipelines, and Workflows), Tasks, Tools, Memory (Conversation Memory, Task Memory, and Meta Memory), Drivers (Prompt and Embedding Drivers, Vector Store Drivers, Image Generation Drivers, Image Query Drivers, SQL Drivers, Web Scraper Drivers, and Conversation Memory Drivers), Engines (Query Engines, Extraction Engines, Summary Engines, Image Generation Engines, and Image Query Engines), and additional components (Rulesets, Loaders, Artifacts, Chunkers, and Tokenizers). Griptape enables developers to create AI-powered applications with ease and efficiency.