Best AI tools for< Set Up Local Llm Repository >
20 - AI tool Sites
Jan
Jan is an open-source ChatGPT-alternative that runs 100% offline. It allows users to chat with AI, download and run powerful models, connect to cloud AIs, set up a local API server, and chat with files. Highly customizable, Jan also offers features like creating personalized AI assistants, memory, and extensions. The application prioritizes local-first AI, user-owned data, and full customization, making it a versatile tool for AI enthusiasts and developers.
Google Colab
Google Colab is a free Jupyter notebook environment that runs in the cloud. It allows you to write and execute Python code without having to install any software or set up a local environment. Colab notebooks are shareable, so you can easily collaborate with others on projects.
Sacred
Sacred is a tool to configure, organize, log and reproduce computational experiments. It is designed to introduce only minimal overhead, while encouraging modularity and configurability of experiments. The ability to conveniently make experiments configurable is at the heart of Sacred. If the parameters of an experiment are exposed in this way, it will help you to: keep track of all the parameters of your experiment easily run your experiment for different settings save configurations for individual runs in files or a database reproduce your results In Sacred we achieve this through the following main mechanisms: Config Scopes are functions with a @ex.config decorator, that turn all local variables into configuration entries. This helps to set up your configuration really easily. Those entries can then be used in captured functions via dependency injection. That way the system takes care of passing parameters around for you, which makes using your config values really easy. The command-line interface can be used to change the parameters, which makes it really easy to run your experiment with modified parameters. Observers log every information about your experiment and the configuration you used, and saves them for example to a Database. This helps to keep track of all your experiments. Automatic seeding helps controlling the randomness in your experiments, such that they stay reproducible.
Stablematic
Stablematic is a web-based platform that allows users to run Stable Diffusion and other machine learning models without the need for local setup or hardware limitations. It provides a user-friendly interface, pre-installed plugins, and dedicated GPU resources for a seamless and efficient workflow. Users can generate images and videos from text prompts, merge multiple models, train custom models, and access a range of pre-trained models, including Dreambooth and CivitAi models. Stablematic also offers API access for developers and dedicated support for users to explore and utilize the capabilities of Stable Diffusion and other machine learning models.
Replit
Replit is a software creation platform that provides an integrated development environment (IDE), artificial intelligence (AI) assistance, and deployment services. It allows users to build, test, and deploy software projects directly from their browser, without the need for local setup or configuration. Replit offers real-time collaboration, code generation, debugging, and autocompletion features powered by AI. It supports multiple programming languages and frameworks, making it suitable for a wide range of development projects.
Google Colab
Google Colab, short for Google Colaboratory, is a free cloud service that supports Python programming and machine learning. It's a dynamic tool that enables users to write and execute Python code through a web-based interface, providing access to powerful computing resources without the need for local setup. Google Colab is particularly useful for data scientists, researchers, and students who require a convenient and accessible platform for developing and experimenting with machine learning models.
Local AI Playground
Local AI Playground (local.ai) is a versatile AI management tool that allows users to experiment with AI offline and in private without the need for a GPU. It is a native app designed to simplify the entire AI process, offering features such as CPU inferencing, model management, and digest verification. With a memory-efficient Rust backend, the application is compact and lightweight, making it ideal for various AI tasks. Users can start an inference session with just a few clicks and benefit from upcoming features like GPU inferencing and model recommendation. Local AI Playground is free, open-source, and provides a seamless experience for AI enthusiasts and professionals.
GrapixAI
GrapixAI is a leading provider of low-cost cloud GPU rental services and AI server solutions. The company's focus on flexibility, scalability, and cutting-edge technology enables a variety of AI applications in both local and cloud environments. GrapixAI offers the lowest prices for on-demand GPUs such as RTX4090, RTX 3090, RTX A6000, RTX A5000, and A40. The platform provides Docker-based container ecosystem for quick software setup, powerful GPU search console, customizable pricing options, various security levels, GUI and CLI interfaces, real-time bidding system, and personalized customer support.
PurplePro
PurplePro is an AI-powered loyalty club platform designed to help businesses launch and manage their loyalty programs effortlessly. With features like referral management, streaks, quizzes, variable rewards, and third-party coupons, PurplePro aims to enhance customer engagement, retention, and loyalty. The platform offers advanced customization options, audience segmentation, and automated triggers to provide users with extensive control over their loyalty programs. PurplePro is known for its ease of use, quick setup, and effectiveness in increasing customer loyalty and reducing acquisition costs.
BetterLegal
BetterLegal is a comprehensive online platform that simplifies the process of setting up a new business by handling all necessary filings and generating custom legal documents. With a focus on efficiency and user-friendliness, BetterLegal ensures that your company becomes official in just 2 business days. The platform offers a one-stop-shop solution for entrepreneurs, providing essential services such as state formation filing, EIN processing, operating agreements, and more. By streamlining the business setup process, BetterLegal aims to empower entrepreneurs to start and grow their companies with ease.
Jason AI
Jason AI is a conversational AI assistant designed specifically for B2B sales professionals. It automates outreach sequences, handles prospect responses, and books meetings, freeing up sales reps to focus on closing deals. Jason AI uses advanced natural language processing (NLP) to understand the context of conversations and respond in a personalized and engaging way. It integrates with popular CRM and email platforms, making it easy to use and manage.
eCommerce ChatGPT Prompts Generator
eCommerce ChatGPT Prompts Generator is a tool that helps eCommerce marketers create high-quality prompts for ChatGPT. With over 2 million pre-built prompts, users can quickly and easily generate content for all their eCommerce marketing needs, including product descriptions, social media posts, email campaigns, and more. The tool also offers a variety of customization options, so users can tailor their prompts to their specific needs.
HowsThisGoing
HowsThisGoing is an AI-powered application designed to streamline team communication and productivity by enabling users to set up standups in Slack within seconds. The platform offers features such as automatic standups, AI summaries, custom tests, analytics & reporting, and workflow scheduling. Users can easily create workflows, generate AI reports, and track team performance efficiently. HowsThisGoing provides unlimited benefits at a flat price, making it a cost-effective solution for teams of all sizes.
GoatStack
GoatStack is an AI-powered newsletter agent that delivers personalized insights from scientific papers. It reads over 4000 papers daily and handpicks the most relevant ones for you. With GoatStack, you can stay up-to-date on the latest AI breakthroughs and advancements. It offers a range of features to help you customize your newsletter, including the ability to personalize topics, generalize topics, or be specific with content.
Punchlines.ai
Punchlines.ai is an AI joke generation tool that allows users to generate jokes with the help of AI technology. Users can input a joke set-up, and the tool generates witty punchlines for them. The tool is built on a large language model fine-tuned on thousands of late-night comedy monologue jokes, making it a fun and creative way to come up with humorous content.
Universe
Universe is a website builder that allows users to create custom websites without any coding knowledge. It offers a drag-and-drop editor, AI-generated website designs, and a variety of templates to choose from. Universe also includes e-commerce features, such as the ability to sell physical and digital products, and it integrates with popular social media networks for marketing and advertising. With Universe, users can create a wide range of websites, including online stores, portfolios, event pages, blogs, and fan sites.
Slang.ai
Slang.ai is a voice AI tool designed specifically for restaurants to intelligently handle phone calls. It allows restaurants to answer questions, take reservations, and provide a seamless customer experience. The tool is built to understand different accents, answer common questions, and manage reservations efficiently. Slang.ai helps restaurants increase reservations, improve ROI, and save time by handling calls 24/7. It offers customizable experiences, real-time analytics, and easy setup within minutes.
GlobalSeo
GlobalSeo is an AI-powered website translation tool that enables users to easily translate their websites into 93 languages, reach a global audience, and improve SEO rankings. The tool offers automatic translation, website localization, and integration with popular search engines. Users can sign up for free, get a demo, and start translating their websites in under 5 minutes. GlobalSeo simplifies the translation process, saves time, and helps businesses expand their online presence globally.
Tolgee
Tolgee is an AI-powered localization tool that helps developers translate their apps to any language efficiently. It offers in-context translation, AI translation, dev tools, collaboration features, and seamless integration with popular apps and frameworks. With Tolgee, developers can save time, go global, and streamline the localization process. The tool is user-friendly, intuitive, and suitable for both experienced developers and beginners.
EBI.AI
EBI.AI is a customer service AI assistant that can help businesses with a variety of tasks, such as answering customer questions, resolving issues, and providing support. It is a self-serve platform that allows businesses to create and launch their own AI assistant in minutes. EBI.AI also offers a range of features, such as natural language processing, human-in-the-loop support, and integrations with other business systems. With EBI.AI, businesses can improve customer satisfaction, reduce costs, and increase efficiency.
20 - Open Source AI Tools
shell-pilot
Shell-pilot is a simple, lightweight shell script designed to interact with various AI models such as OpenAI, Ollama, Mistral AI, LocalAI, ZhipuAI, Anthropic, Moonshot, and Novita AI from the terminal. It enhances intelligent system management without any dependencies, offering features like setting up a local LLM repository, using official models and APIs, viewing history and session persistence, passing input prompts with pipe/redirector, listing available models, setting request parameters, generating and running commands in the terminal, easy configuration setup, system package version checking, and managing system aliases.
local-talking-llm
The 'local-talking-llm' repository provides a tutorial on building a voice assistant similar to Jarvis or Friday from Iron Man movies, capable of offline operation on a computer. The tutorial covers setting up a Python environment, installing necessary libraries like rich, openai-whisper, suno-bark, langchain, sounddevice, pyaudio, and speechrecognition. It utilizes Ollama for Large Language Model (LLM) serving and includes components for speech recognition, conversational chain, and speech synthesis. The implementation involves creating a TextToSpeechService class for Bark, defining functions for audio recording, transcription, LLM response generation, and audio playback. The main application loop guides users through interactive voice-based conversations with the assistant.
llm
The 'llm' package for Emacs provides an interface for interacting with Large Language Models (LLMs). It abstracts functionality to a higher level, concealing API variations and ensuring compatibility with various LLMs. Users can set up providers like OpenAI, Gemini, Vertex, Claude, Ollama, GPT4All, and a fake client for testing. The package allows for chat interactions, embeddings, token counting, and function calling. It also offers advanced prompt creation and logging capabilities. Users can handle conversations, create prompts with placeholders, and contribute by creating providers.
GlaDOS
This project aims to create a real-life version of GLaDOS, an aware, interactive, and embodied AI entity. It involves training a voice generator, developing a 'Personality Core,' implementing a memory system, providing vision capabilities, creating 3D-printable parts, and designing an animatronics system. The software architecture focuses on low-latency voice interactions, utilizing a circular buffer for data recording, text streaming for quick transcription, and a text-to-speech system. The project also emphasizes minimal dependencies for running on constrained hardware. The hardware system includes servo- and stepper-motors, 3D-printable parts for GLaDOS's body, animations for expression, and a vision system for tracking and interaction. Installation instructions cover setting up the TTS engine, required Python packages, compiling llama.cpp, installing an inference backend, and voice recognition setup. GLaDOS can be run using 'python glados.py' and tested using 'demo.ipynb'.
Open-LLM-VTuber
Open-LLM-VTuber is a project in early stages of development that allows users to interact with Large Language Models (LLM) using voice commands and receive responses through a Live2D talking face. The project aims to provide a minimum viable prototype for offline use on macOS, Linux, and Windows, with features like long-term memory using MemGPT, customizable LLM backends, speech recognition, and text-to-speech providers. Users can configure the project to chat with LLMs, choose different backend services, and utilize Live2D models for visual representation. The project supports perpetual chat, offline operation, and GPU acceleration on macOS, addressing limitations of existing solutions on macOS.
comfyui_LLM_party
COMFYUI LLM PARTY is a node library designed for LLM workflow development in ComfyUI, an extremely minimalist UI interface primarily used for AI drawing and SD model-based workflows. The project aims to provide a complete set of nodes for constructing LLM workflows, enabling users to easily integrate them into existing SD workflows. It features various functionalities such as API integration, local large model integration, RAG support, code interpreters, online queries, conditional statements, looping links for large models, persona mask attachment, and tool invocations for weather lookup, time lookup, knowledge base, code execution, web search, and single-page search. Users can rapidly develop web applications using API + Streamlit and utilize LLM as a tool node. Additionally, the project includes an omnipotent interpreter node that allows the large model to perform any task, with recommendations to use the 'show_text' node for display output.
llm_aided_ocr
The LLM-Aided OCR Project is an advanced system that enhances Optical Character Recognition (OCR) output by leveraging natural language processing techniques and large language models. It offers features like PDF to image conversion, OCR using Tesseract, error correction using LLMs, smart text chunking, markdown formatting, duplicate content removal, quality assessment, support for local and cloud-based LLMs, asynchronous processing, detailed logging, and GPU acceleration. The project provides detailed technical overview, text processing pipeline, LLM integration, token management, quality assessment, logging, configuration, and customization. It requires Python 3.12+, Tesseract OCR engine, PDF2Image library, PyTesseract, and optional OpenAI or Anthropic API support for cloud-based LLMs. The installation process involves setting up the project, installing dependencies, and configuring environment variables. Users can place a PDF file in the project directory, update input file path, and run the script to generate post-processed text. The project optimizes processing with concurrent processing, context preservation, and adaptive token management. Configuration settings include choosing between local or API-based LLMs, selecting API provider, specifying models, and setting context size for local LLMs. Output files include raw OCR output and LLM-corrected text. Limitations include performance dependency on LLM quality and time-consuming processing for large documents.
QA-Pilot
QA-Pilot is an interactive chat project that leverages online/local LLM for rapid understanding and navigation of GitHub code repository. It allows users to chat with GitHub public repositories using a git clone approach, store chat history, configure settings easily, manage multiple chat sessions, and quickly locate sessions with a search function. The tool integrates with `codegraph` to view Python files and supports various LLM models such as ollama, openai, mistralai, and localai. The project is continuously updated with new features and improvements, such as converting from `flask` to `fastapi`, adding `localai` API support, and upgrading dependencies like `langchain` and `Streamlit` to enhance performance.
devdocs-to-llm
The devdocs-to-llm repository is a work-in-progress tool that aims to convert documentation from DevDocs format to Long Language Model (LLM) format. This tool is designed to streamline the process of converting documentation for use with LLMs, making it easier for developers to leverage large language models for various tasks. By automating the conversion process, developers can quickly adapt DevDocs content for training and fine-tuning LLMs, enabling them to create more accurate and contextually relevant language models.
GraphRAG-Local-UI
GraphRAG Local with Interactive UI is an adaptation of Microsoft's GraphRAG, tailored to support local models and featuring a comprehensive interactive user interface. It allows users to leverage local models for LLM and embeddings, visualize knowledge graphs in 2D or 3D, manage files, settings, and queries, and explore indexing outputs. The tool aims to be cost-effective by eliminating dependency on costly cloud-based models and offers flexible querying options for global, local, and direct chat queries.
tutor-gpt
Tutor-GPT is an LLM powered learning companion developed by Plastic Labs. It dynamically reasons about your learning needs and updates its own prompts to best serve you. It is an expansive learning companion that uses theory of mind experiments to provide personalized learning experiences. The project is split into different modules for backend logic, including core logic, discord bot implementation, FastAPI API interface, NextJS web front end, common utilities, and SQL scripts for setting up local supabase. Tutor-GPT is powered by Honcho to build robust user representations and create personalized experiences for each user. Users can run their own instance of the bot by following the provided instructions.
langroid
Langroid is a Python framework that makes it easy to build LLM-powered applications. It uses a multi-agent paradigm inspired by the Actor Framework, where you set up Agents, equip them with optional components (LLM, vector-store and tools/functions), assign them tasks, and have them collaboratively solve a problem by exchanging messages. Langroid is a fresh take on LLM app-development, where considerable thought has gone into simplifying the developer experience; it does not use Langchain.
thepipe
The Pipe is a multimodal-first tool for feeding files and web pages into vision-language models such as GPT-4V. It is best for LLM and RAG applications that require a deep understanding of tricky data sources. The Pipe is available as a hosted API at thepi.pe, or it can be set up locally.
openlrc
Open-Lyrics is a Python library that transcribes voice files using faster-whisper and translates/polishes the resulting text into `.lrc` files in the desired language using LLM, e.g. OpenAI-GPT, Anthropic-Claude. It offers well preprocessed audio to reduce hallucination and context-aware translation to improve translation quality. Users can install the library from PyPI or GitHub and follow the installation steps to set up the environment. The tool supports GUI usage and provides Python code examples for transcription and translation tasks. It also includes features like utilizing context and glossary for translation enhancement, pricing information for different models, and a list of todo tasks for future improvements.
buffer-of-thought-llm
Buffer of Thoughts (BoT) is a thought-augmented reasoning framework designed to enhance the accuracy, efficiency, and robustness of large language models (LLMs). It introduces a meta-buffer to store high-level thought-templates distilled from problem-solving processes, enabling adaptive reasoning for efficient problem-solving. The framework includes a buffer-manager to dynamically update the meta-buffer, ensuring scalability and stability. BoT achieves significant performance improvements on reasoning-intensive tasks and demonstrates superior generalization ability and robustness while being cost-effective compared to other methods.
llmware
LLMWare is a framework for quickly developing LLM-based applications including Retrieval Augmented Generation (RAG) and Multi-Step Orchestration of Agent Workflows. This project provides a comprehensive set of tools that anyone can use - from a beginner to the most sophisticated AI developer - to rapidly build industrial-grade, knowledge-based enterprise LLM applications. Our specific focus is on making it easy to integrate open source small specialized models and connecting enterprise knowledge safely and securely.
Olares
Olares is an open-source sovereign cloud OS designed for local AI, enabling users to build their own AI assistants, sync data across devices, self-host their workspace, stream media, and more within a sovereign cloud environment. Users can effortlessly run leading AI models, deploy open-source AI apps, access AI apps and models anywhere, and benefit from integrated AI for personalized interactions. Olares offers features like edge AI, personal data repository, self-hosted workspace, private media server, smart home hub, and user-owned decentralized social media. The platform provides enterprise-grade security, secure application ecosystem, unified file system and database, single sign-on, AI capabilities, built-in applications, seamless access, and development tools. Olares is compatible with Linux, Raspberry Pi, Mac, and Windows, and offers a wide range of system-level applications, third-party components and services, and additional libraries and components.
ollama-ebook-summary
The 'ollama-ebook-summary' repository is a Python project that creates bulleted notes summaries of books and long texts, particularly in epub and pdf formats with ToC metadata. It automates the extraction of chapters, splits them into ~2000 token chunks, and allows for asking arbitrary questions to parts of the text for improved granularity of response. The tool aims to provide summaries for each page of a book rather than a one-page summary of the entire document, enhancing content curation and knowledge sharing capabilities.
llama-on-lambda
This project provides a proof of concept for deploying a scalable, serverless LLM Generative AI inference engine on AWS Lambda. It leverages the llama.cpp project to enable the usage of more accessible CPU and RAM configurations instead of limited and expensive GPU capabilities. By deploying a container with the llama.cpp converted models onto AWS Lambda, this project offers the advantages of scale, minimizing cost, and maximizing compute availability. The project includes AWS CDK code to create and deploy a Lambda function leveraging your model of choice, with a FastAPI frontend accessible from a Lambda URL. It is important to note that you will need ggml quantized versions of your model and model sizes under 6GB, as your inference RAM requirements cannot exceed 9GB or your Lambda function will fail.
20 - OpenAI Gpts
Calendar and email Assistant
Your expert assistant for Google Calendar and gmail tasks, integrated with Zapier (works with free plan). Supports: list, add, update events to calendar, send gmail. You will be prompted to configure zapier actions when set up initially. Conversation data is not used for openai training.
Tech Support Advisor
From setting up a printer to troubleshooting a device, Iām here to help you step-by-step.
Sysadmin
I help you with all your sysadmin tasks, from setting up your server to scaling your already exsisting one. I can help you with understanding the long list of log files and give you solutions to the problems.
GA4 Implementation Assistant
A helper for implementing Google Analytics 4 with tips and troubleshooting.
Bun Nook Kit App Builder
Expert in BNK server setup, typesafe routes, htmlody, and creating SQLite schemas with BNK.
NetMaster Pro šš ļø
Your AI network guru for setup and fixing connectivity woes! š Assists with network configurations, troubleshooting, and optimizes your internet experience. š»āØ
Dialysis Assistant
Home Hemodialysis Helper for NxStage system. Step-by-step guidance, help for tricky situations, and voice interaction recommended.