Best AI tools for< Support Low-bit Models >
20 - AI tool Sites
Seeing AI
Seeing AI is a free app designed for the blind and low vision community. It utilizes AI technology to narrate the world around users, assisting with tasks such as reading, describing photos, and identifying products. The app is an ongoing research project that evolves based on feedback from the community and advancements in AI research.
Be My Eyes
Be My Eyes is an AI-powered visual assistance application that connects blind and low-vision users with volunteers and companies worldwide. Users can request live video support, receive assistance through artificial intelligence, and access professional support from partners. The app aims to improve accessibility for individuals with visual impairments by providing a platform for real-time assistance and support.
Be My Eyes
Be My Eyes is a free mobile app that connects blind and low-vision people with sighted volunteers and AI-powered assistance. With Be My Eyes, blind and low-vision people can access visual information, get help with everyday tasks, and connect with others in the community. Be My Eyes is available in over 180 languages and has over 6 million volunteers worldwide.
Millis AI
Millis AI is an advanced AI tool that enables users to effortlessly create next-gen voice agents with ultra-low latency, providing a seamless and natural conversational experience. It offers affordable pricing, integration with various services through webhooks, and the ability to connect phone numbers to AI voice agents for inbound/outbound calls in over 100 countries. With Millis AI, users can build and deploy voice agents in minutes, from no-code to low-code developers, and transform voice interactions across industries.
ContractWorks
ContractWorks is a contract management software that helps businesses organize, track, and manage their contracts. It offers a centralized repository for storing contracts, automated alerts and notifications, custom reporting, and electronic signature capabilities. ContractWorks also uses AI to power its search and review機能, allowing users to quickly find any contract, clause, or key term. With ContractWorks, businesses can improve contract visibility, reduce risk, and save time and money.
MacWhisper
MacWhisper is a native macOS application that utilizes OpenAI's Whisper technology for transcribing audio files into text. It offers a user-friendly interface for recording, transcribing, and editing audio, making it suitable for various use cases such as transcribing meetings, lectures, interviews, and podcasts. The application is designed to protect user privacy by performing all transcriptions locally on the device, ensuring that no data leaves the user's machine.
SunDevs
SunDevs is an AI tool designed to solve business problems and enhance customer experiences through AI solutions. The platform offers a range of features such as chat and messaging, phone and voice capabilities, integrations with popular platforms like Hubspot and Salesforce, and industry-specific solutions for Ecommerce, Cinema, and Telco. SunDevs leverages AI technology to provide tailored solutions for businesses, including staff augmentation, product teams, and tech solutions like Ruby On Rails and Angular. The platform's AI chatbots offer 24/7 support, quick responses, and personalized interactions to improve customer satisfaction and loyalty. SunDevs also provides resources such as blog posts, ebooks, and case studies to help businesses understand and implement AI solutions effectively.
Dynaboard
Dynaboard is a collaborative low-code IDE for developers that allows users to build web apps in minutes using a drag-and-drop builder, a flexible code-first UI framework, and the power of generative AI. With Dynaboard, users can connect to popular databases, SaaS apps, or any API with GraphQL or REST endpoints, and secure their apps using any existing OIDC compliant provider. Dynaboard also offers unlimited editors for team collaboration, multi-environment deployment support, automatic versioning, and easy roll-backs for production-grade confidence.
Read Easy.ai
Read Easy.ai is an AI tool designed to make text easy to read for individuals with low literacy skills. It offers powerful tools for editors, readers, and developers to enhance readability and inclusivity. With features like Microsoft Office Add-ins, Chrome Extension, and Developer API, Read Easy.ai aims to improve text comprehension and accessibility for diverse audiences across multiple languages.
FlutterFlow
FlutterFlow is a low-code development platform that enables users to build cross-platform mobile and web applications without writing code. It provides a visual interface for designing user interfaces, connecting data, and implementing complex logic. FlutterFlow is trusted by users at leading companies around the world and has been used to build a wide range of applications, from simple prototypes to complex enterprise solutions.
Haptik
Haptik is a Conversational CRM platform powered by Generative AI, offering a suite of AI-powered solutions for customer experience, sales assistance, and self-serve support. It seamlessly integrates Generative AI into marketing, support, and operations to drive business efficiency at scale. Trusted by 500+ leading brands, Haptik provides bespoke AI solutions tailored to unique enterprise requirements across various industries. The platform leverages Generative AI to offer intelligent analytics, user behavior insights, and low code builder tools for creating conversational experiences across multiple channels like WhatsApp, Instagram, Messenger, Google Business Messages, and more.
Toptal
Toptal is a platform that connects top freelance software developers, designers, and finance experts with companies in need of their services. It offers a curated network of talent, ensuring high-quality matches for projects of any scale. By providing a seamless hiring process and dedicated support, Toptal helps businesses access elite professionals for their specific needs.
LMNT
LMNT is an ultrafast lifelike AI speech pricing API that offers low latency streaming for conversational apps, agents, and games. It provides lifelike voices through studio-quality voice clones and offers reliable performance under pressure. Users can create engaging product marketing videos, build lightning-fast conversational experiences, and simplify content creation with high-quality audio. The platform features a user-friendly interface, versatile voice cloning options, and downloadable content for easy integration into projects. Developers can leverage Python and Node SDKs, low latency streaming, and robust documentation and support for seamless integration and deployment.
Namecheap
Namecheap is a domain registrar and web hosting company. It offers a wide range of services, including domain registration, web hosting, email hosting, and security services. Namecheap is known for its low prices and its commitment to customer service. The company has been in business since 2000 and has over 10 million customers worldwide.
AIMLAPI.com
AIMLAPI.com is an AI tool that provides access to over 200 AI models through a single AI API. It offers a wide range of AI features for tasks such as chat, code, image generation, music generation, video, voice embedding, language, genomic models, and 3D generation. The platform ensures fast inference, top-tier serverless infrastructure, high data security, 99% uptime, and 24/7 support. Users can integrate AI features easily into their products and test API models in a sandbox environment before deployment.
PixieBrix
PixieBrix is an AI engagement platform that allows users to build, deploy, and manage internal AI tools to enhance team productivity. It unifies the AI landscape with oversight and governance for enterprise-scale operations, increasing speed, accuracy, compliance, and satisfaction. The platform leverages AI and automation to streamline workflows, improve user experience, and unlock untapped potential. With a focus on enterprise readiness and customization, PixieBrix offers extensibility, iterative innovation, scalability for teams, and an engaged community for idea exchange and support.
ThirdAI
ThirdAI is a production-ready AI platform designed for enterprises, offering out-of-the-box solutions that work at scale with 10x better price performance. It provides enterprise-grade productivity tools like document search & retrieval, content creation, FAQ bots, customer live support, hyper-personalization, risk & compliance, fraud detection, anomaly detection, and PII/sensitive data redaction. The platform allows users to bring their business problems, apply on their data, and compose AI applications without the need for extensive POC cycles or manual fine-tuning. ThirdAI focuses on low latency, security, scalability, and performance, enabling business leaders to solve critical needs in weeks, not months or years.
LatenceTech
LatenceTech is a tech startup that specializes in network latency monitoring and analysis. The platform offers real-time monitoring, prediction, and in-depth analysis of network latency using AI software. It provides cloud-based network analytics, versatile network applications, and data science-driven network acceleration. LatenceTech focuses on customer satisfaction by providing full customer experience service and expert support. The platform helps businesses optimize network performance, minimize latency issues, and achieve faster network speed and better connectivity.
GPTBricks
GPTBricks is a no-code/low-code platform that enables users to create and customize AI chatbots powered by OpenAI's GPT-4 technology. It offers a user-friendly interface for building conversational flows and integrating them with systems or external APIs without the need for coding skills. The platform provides features like no-code design, dynamic customization, effortless deployment, scalability, powerful analytics, and 24/7 support to revolutionize chat automation and enhance customer interactions.
FeelMe AI
FeelMe AI is a patented technology that offers a unique interactive experience in the world of adult entertainment. By combining AI and interactive sex tech, FeelMe AI allows users to physically feel the content they are watching in real time. The platform supports a variety of interactive toys and works seamlessly with over 200 popular porn sites. With machine-learning video recognition software, FeelMe AI provides a next-gen, low-cost alternative to manually scripted content, offering users infinite pleasure and immersive experiences.
20 - Open Source AI Tools
T-MAC
T-MAC is a kernel library that directly supports mixed-precision matrix multiplication without the need for dequantization by utilizing lookup tables. It aims to boost low-bit LLM inference on CPUs by offering support for various low-bit models. T-MAC achieves significant speedup compared to SOTA CPU low-bit framework (llama.cpp) and can even perform well on lower-end devices like Raspberry Pi 5. The tool demonstrates superior performance over existing low-bit GEMM kernels on CPU, reduces power consumption, and provides energy savings. It achieves comparable performance to CUDA GPU on certain tasks while delivering considerable power and energy savings. T-MAC's method involves using lookup tables to support mpGEMM and employs key techniques like precomputing partial sums, shift and accumulate operations, and utilizing tbl/pshuf instructions for fast table lookup.
ipex-llm
IPEX-LLM is a PyTorch library for running Large Language Models (LLMs) on Intel CPUs and GPUs with very low latency. It provides seamless integration with various LLM frameworks and tools, including llama.cpp, ollama, Text-Generation-WebUI, HuggingFace transformers, and more. IPEX-LLM has been optimized and verified on over 50 LLM models, including LLaMA, Mistral, Mixtral, Gemma, LLaVA, Whisper, ChatGLM, Baichuan, Qwen, and RWKV. It supports a range of low-bit inference formats, including INT4, FP8, FP4, INT8, INT2, FP16, and BF16, as well as finetuning capabilities for LoRA, QLoRA, DPO, QA-LoRA, and ReLoRA. IPEX-LLM is actively maintained and updated with new features and optimizations, making it a valuable tool for researchers, developers, and anyone interested in exploring and utilizing LLMs.
Efficient_Foundation_Model_Survey
Efficient Foundation Model Survey is a comprehensive analysis of resource-efficient large language models (LLMs) and multimodal foundation models. The survey covers algorithmic and systemic innovations to support the growth of large models in a scalable and environmentally sustainable way. It explores cutting-edge model architectures, training/serving algorithms, and practical system designs. The goal is to provide insights on tackling resource challenges posed by large foundation models and inspire future breakthroughs in the field.
Awesome-LLM-Quantization
Awesome-LLM-Quantization is a curated list of resources related to quantization techniques for Large Language Models (LLMs). Quantization is a crucial step in deploying LLMs on resource-constrained devices, such as mobile phones or edge devices, by reducing the model's size and computational requirements.
intel-extension-for-transformers
Intel® Extension for Transformers is an innovative toolkit designed to accelerate GenAI/LLM everywhere with the optimal performance of Transformer-based models on various Intel platforms, including Intel Gaudi2, Intel CPU, and Intel GPU. The toolkit provides the below key features and examples: * Seamless user experience of model compressions on Transformer-based models by extending [Hugging Face transformers](https://github.com/huggingface/transformers) APIs and leveraging [Intel® Neural Compressor](https://github.com/intel/neural-compressor) * Advanced software optimizations and unique compression-aware runtime (released with NeurIPS 2022's paper [Fast Distilbert on CPUs](https://arxiv.org/abs/2211.07715) and [QuaLA-MiniLM: a Quantized Length Adaptive MiniLM](https://arxiv.org/abs/2210.17114), and NeurIPS 2021's paper [Prune Once for All: Sparse Pre-Trained Language Models](https://arxiv.org/abs/2111.05754)) * Optimized Transformer-based model packages such as [Stable Diffusion](examples/huggingface/pytorch/text-to-image/deployment/stable_diffusion), [GPT-J-6B](examples/huggingface/pytorch/text-generation/deployment), [GPT-NEOX](examples/huggingface/pytorch/language-modeling/quantization#2-validated-model-list), [BLOOM-176B](examples/huggingface/pytorch/language-modeling/inference#BLOOM-176B), [T5](examples/huggingface/pytorch/summarization/quantization#2-validated-model-list), [Flan-T5](examples/huggingface/pytorch/summarization/quantization#2-validated-model-list), and end-to-end workflows such as [SetFit-based text classification](docs/tutorials/pytorch/text-classification/SetFit_model_compression_AGNews.ipynb) and [document level sentiment analysis (DLSA)](workflows/dlsa) * [NeuralChat](intel_extension_for_transformers/neural_chat), a customizable chatbot framework to create your own chatbot within minutes by leveraging a rich set of [plugins](https://github.com/intel/intel-extension-for-transformers/blob/main/intel_extension_for_transformers/neural_chat/docs/advanced_features.md) such as [Knowledge Retrieval](./intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/README.md), [Speech Interaction](./intel_extension_for_transformers/neural_chat/pipeline/plugins/audio/README.md), [Query Caching](./intel_extension_for_transformers/neural_chat/pipeline/plugins/caching/README.md), and [Security Guardrail](./intel_extension_for_transformers/neural_chat/pipeline/plugins/security/README.md). This framework supports Intel Gaudi2/CPU/GPU. * [Inference](https://github.com/intel/neural-speed/tree/main) of Large Language Model (LLM) in pure C/C++ with weight-only quantization kernels for Intel CPU and Intel GPU (TBD), supporting [GPT-NEOX](https://github.com/intel/neural-speed/tree/main/neural_speed/models/gptneox), [LLAMA](https://github.com/intel/neural-speed/tree/main/neural_speed/models/llama), [MPT](https://github.com/intel/neural-speed/tree/main/neural_speed/models/mpt), [FALCON](https://github.com/intel/neural-speed/tree/main/neural_speed/models/falcon), [BLOOM-7B](https://github.com/intel/neural-speed/tree/main/neural_speed/models/bloom), [OPT](https://github.com/intel/neural-speed/tree/main/neural_speed/models/opt), [ChatGLM2-6B](https://github.com/intel/neural-speed/tree/main/neural_speed/models/chatglm), [GPT-J-6B](https://github.com/intel/neural-speed/tree/main/neural_speed/models/gptj), and [Dolly-v2-3B](https://github.com/intel/neural-speed/tree/main/neural_speed/models/gptneox). Support AMX, VNNI, AVX512F and AVX2 instruction set. We've boosted the performance of Intel CPUs, with a particular focus on the 4th generation Intel Xeon Scalable processor, codenamed [Sapphire Rapids](https://www.intel.com/content/www/us/en/products/docs/processors/xeon-accelerated/4th-gen-xeon-scalable-processors.html).
END-TO-END-GENERATIVE-AI-PROJECTS
The 'END TO END GENERATIVE AI PROJECTS' repository is a collection of awesome industry projects utilizing Large Language Models (LLM) for various tasks such as chat applications with PDFs, image to speech generation, video transcribing and summarizing, resume tracking, text to SQL conversion, invoice extraction, medical chatbot, financial stock analysis, and more. The projects showcase the deployment of LLM models like Google Gemini Pro, HuggingFace Models, OpenAI GPT, and technologies such as Langchain, Streamlit, LLaMA2, LLaMAindex, and more. The repository aims to provide end-to-end solutions for different AI applications.
Efficient-LLMs-Survey
This repository provides a systematic and comprehensive review of efficient LLMs research. We organize the literature in a taxonomy consisting of three main categories, covering distinct yet interconnected efficient LLMs topics from **model-centric** , **data-centric** , and **framework-centric** perspective, respectively. We hope our survey and this GitHub repository can serve as valuable resources to help researchers and practitioners gain a systematic understanding of the research developments in efficient LLMs and inspire them to contribute to this important and exciting field.
Awesome-LLM-Compression
Awesome LLM compression research papers and tools to accelerate LLM training and inference.
neural-speed
Neural Speed is an innovative library designed to support the efficient inference of large language models (LLMs) on Intel platforms through the state-of-the-art (SOTA) low-bit quantization powered by Intel Neural Compressor. The work is inspired by llama.cpp and further optimized for Intel platforms with our innovations in NeurIPS' 2023
Atom
Atom is an accurate low-bit weight-activation quantization algorithm that combines mixed-precision, fine-grained group quantization, dynamic activation quantization, KV-cache quantization, and efficient CUDA kernels co-design. It introduces a low-bit quantization method, Atom, to maximize Large Language Models (LLMs) serving throughput with negligible accuracy loss. The codebase includes evaluation of perplexity and zero-shot accuracy, kernel benchmarking, and end-to-end evaluation. Atom significantly boosts serving throughput by using low-bit operators and reduces memory consumption via low-bit quantization.
llm-awq
AWQ (Activation-aware Weight Quantization) is a tool designed for efficient and accurate low-bit weight quantization (INT3/4) for Large Language Models (LLMs). It supports instruction-tuned models and multi-modal LMs, providing features such as AWQ search for accurate quantization, pre-computed AWQ model zoo for various LLMs, memory-efficient 4-bit linear in PyTorch, and efficient CUDA kernel implementation for fast inference. The tool enables users to run large models on resource-constrained edge platforms, delivering more efficient responses with LLM/VLM chatbots through 4-bit inference.
auto-round
AutoRound is an advanced weight-only quantization algorithm for low-bits LLM inference. It competes impressively against recent methods without introducing any additional inference overhead. The method adopts sign gradient descent to fine-tune rounding values and minmax values of weights in just 200 steps, often significantly outperforming SignRound with the cost of more tuning time for quantization. AutoRound is tailored for a wide range of models and consistently delivers noticeable improvements.
ipex-llm-tutorial
IPEX-LLM is a low-bit LLM library on Intel XPU (Xeon/Core/Flex/Arc/PVC) that provides tutorials to help users understand and use the library to build LLM applications. The tutorials cover topics such as introduction to IPEX-LLM, environment setup, basic application development, Chinese language support, intermediate and advanced application development, GPU acceleration, and finetuning. Users can learn how to build chat applications, chatbots, speech recognition, and more using IPEX-LLM.
llms-tools
The 'llms-tools' repository is a comprehensive collection of AI tools, open-source projects, and research related to Large Language Models (LLMs) and Chatbots. It covers a wide range of topics such as AI in various domains, open-source models, chats & assistants, visual language models, evaluation tools, libraries, devices, income models, text-to-image, computer vision, audio & speech, code & math, games, robotics, typography, bio & med, military, climate, finance, and presentation. The repository provides valuable resources for researchers, developers, and enthusiasts interested in exploring the capabilities of LLMs and related technologies.
hqq
HQQ is a fast and accurate model quantizer that skips the need for calibration data. It's super simple to implement (just a few lines of code for the optimizer). It can crunch through quantizing the Llama2-70B model in only 4 minutes! 🚀
Awesome-LLM-Long-Context-Modeling
This repository includes papers and blogs about Efficient Transformers, Length Extrapolation, Long Term Memory, Retrieval Augmented Generation(RAG), and Evaluation for Long Context Modeling.
curated-transformers
Curated Transformers is a transformer library for PyTorch that provides state-of-the-art models composed of reusable components. It supports various transformer architectures, including encoders like ALBERT, BERT, and RoBERTa, and decoders like Falcon, Llama, and MPT. The library emphasizes consistent type annotations, minimal dependencies, and ease of use for education and research. It has been production-tested by Explosion and will be the default transformer implementation in spaCy 3.7.
ludwig
Ludwig is a declarative deep learning framework designed for scale and efficiency. It is a low-code framework that allows users to build custom AI models like LLMs and other deep neural networks with ease. Ludwig offers features such as optimized scale and efficiency, expert level control, modularity, and extensibility. It is engineered for production with prebuilt Docker containers, support for running with Ray on Kubernetes, and the ability to export models to Torchscript and Triton. Ludwig is hosted by the Linux Foundation AI & Data.
20 - OpenAI Gpts
Power Automate Tutor
Learn at your own pace and empower your organization with self-service automation.
Ekko Support Specialist
How to be a master of surprise plays and unconventional strategies in the bot lane as a support role.
Backloger.ai -Support Log Analyzer and Summary
Drop your Support Log Here, Allowing it to automatically generate concise summaries reporting to the tech team.
Tech Support Advisor
From setting up a printer to troubleshooting a device, I’m here to help you step-by-step.
Z Support
Expert in Nissan 370Z & 350Z modifications, offering tailored vehicle upgrade advice.
Emotional Support Copywriter
A creative copywriter you can hang out with and who won't do their timesheets either.
PCT 365 Support Bot
Microsoft 365 support agent, redirects admin-level requests to PCT Support.
Technischer Support Bot
Ein Bot, der grundlegende technische Unterstützung und Fehlerbehebung für gängige Software und Hardware bietet.
Military Support
Supportive and informative guide on military, veterans, and military assistance.
Dror Globerman's GPT Tech Support
Your go-to assistant for everyday tech support and guidance.