Best AI tools for< Query Documents >
20 - AI tool Sites
AI Brain Bank
AI Brain Bank is a powerful tool that allows you to remember everything. With AI Brain Bank, you can query all your documents, media, and knowledge with AI. This makes it easy to find the information you need, when you need it. AI Brain Bank is the perfect tool for students, researchers, and anyone else who needs to manage a large amount of information.
Focal
Focal is an AI-powered tool that helps users summarize and organize their research and reading materials. It offers features such as AI-generated summaries, document highlighting, and collaboration tools. Focal is designed for researchers, students, professionals, and anyone who needs to efficiently process large amounts of information.
FileAI
The FileAI website offers an AI-powered file reading assistant that specializes in data extraction from structured documents like financial statements, legal documents, and research papers. It automates tasks related to legal and compliance review, finance and accounting report preparation, and research and academia support. The tool aims to streamline document processing, enhance learning processes, and improve research efficiency. With features like summarizing complex texts, extracting key information, and detecting plagiarism, FileAI caters to users in various industries and educational fields. The platform prioritizes data security and user privacy, ensuring that data is used solely for its intended purpose and deleted after 7 days of non-use.
Legalyze.ai
Legalyze.ai is an AI-powered platform designed to assist lawyers in streamlining their document review process. It uses AI to summarize and extract key points from case documents, providing rapid insights, summaries, and answers to specific questions. The platform allows users to create document summaries in seconds, supports various file formats, and is externally security audited. Legalyze.ai aims to save time for legal professionals by automating tasks like fact-finding and document creation.
Chat with Docs
Chat with Docs is a platform that allows users to interact with documents using a simple API. Users can chat with any document by integrating just 2 lines of code. The platform supports various document formats such as Pdf, docx, doc, pptx, txt, and more. Users can ask questions about documents using cUrl, Python, or JavaScript. Chat with Docs offers a straightforward pricing model and emphasizes privacy and terms of use.
ChatWithPDF
ChatWithPDF is a ChatGPT plugin that allows users to query against small or large PDF documents directly in ChatGPT. It offers a convenient way to process and semantically search PDF documents based on your queries. By providing a temporary PDF URL, the plugin fetches relevant information from the PDF file and returns the most suitable matches according to your search input.
AutoQuery GPT
AutoQuery GPT is a tool that allows users to ask questions to ChatGPT and get answers automatically. It provides users with time-saving and performance benefits. Users can use this site by using their own API key to ask questions to ChatGPT and save the answers as a file, using the Query Block and Query Excel features.
Loata
Loata is an AI-powered platform that serves as a learning orchestrator for adaptive text analyses. It allows users to store their notes and documents in the cloud, which are then ingested and transformed into knowledge bases. The platform features smart AI agents powered by LLMs to provide intelligent answers based on the content. With end-to-end encryption and controlled ingestion, Loata ensures the security and privacy of user data. Users can choose from different subscription plans to access varying levels of storage and query capacity, making it suitable for individuals and professionals alike.
PrivacyDoc
PrivacyDoc is an AI-powered portal that allows users to analyze and query PDF and ebooks effortlessly. By leveraging advanced NLP technology, PrivacyDoc enables users to uncover insights and conduct thorough document analysis. The platform offers features such as easy file upload, query functionality, enhanced security measures, and free access to powerful PDF analysis tools. With PrivacyDoc, users can experience the convenience of logging in with their Google account, submitting queries for prompt AI-driven responses, and ensuring data privacy with secure file handling.
SvectorDB
SvectorDB is a vector database built from the ground up for serverless applications. It is designed to be highly scalable, performant, and easy to use. SvectorDB can be used for a variety of applications, including recommendation engines, document search, and image search.
GOODY-2
GOODY-2 is the world's most responsible AI model, built with next-gen adherence to ethical principles. It's so safe that it won't answer anything that could be possibly construed as controversial or problematic. GOODY-2 can recognize any query that could be controversial, offensive, or dangerous in any context and elegantly avoids answering, redirecting the conversation, and mitigating brand risk. GOODY-2's ethical adherence is unbreakable, ensuring that every conversation stays within the bounds of ethical principles. Even bad actors will find themselves unable to cause GOODY-2 to answer problematic queries. GOODY-2 is the perfect fit for customer service, paralegal assistance, back-office tasks, and more. It's the safe, dependable AI model companies around the globe have been waiting for.
Leena AI
Leena AI is a Gen AI employee assistant that reduces IT, HR, Finance tickets. It guarantees a 70% self-service ratio in the contract. Leena AI centralizes knowledge that is scattered across the enterprise, making information easy to find right on chat with a simple query. The knowledge is auto-updated when changes are made, so your employees always get support that is both relevant and accurate. Leena AI integrates with all knowledge base systems within your enterprise from day one, eliminating the need for any consolidation or migration effort across the knowledge base systems. The work assistant tailors responses to employee queries based on their role, team and access level which boosts employee experience while adhering to security protocols and preventing unauthorized access to information. Leena AI analyzes historically closed tickets and learns from their resolutions to create knowledge articles that can prevent similar tickets from being raised again. Leena AI breaks down large policy documents and knowledge articles into consumable snippets so the responses are sharper and more specific to the employee’s query.
AutoKT
AutoKT is an AI-powered application designed for Automatic Knowledge Transfer. It helps in effortless documentation by automatically writing and updating documentation, allowing users to focus on building innovative projects. The tool addresses the challenge of time and bandwidth spent on writing and maintaining documentation in agile workplaces. AutoKT ensures asynchronous knowledge transfer by keeping documentation in sync with code changes and providing a query feature for easy access to information. It is a valuable tool for developers, enabling them to understand legacy code, streamline documentation writing, and facilitate faster onboarding of new team members.
BixGPT
BixGPT is an AI-powered tool designed to supercharge product documentation by leveraging the power of private AI models. It offers features like AI-assisted release notes generation, data encryption, autodiscovery of Jira data, multi-format support, client notifications, and more. With BixGPT, users can create and manage release notes effortlessly while ensuring data privacy and security through the use of private AI models. The tool provides a seamless experience for generating release web pages with custom styling and analytics.
Code99
Code99 is an AI-powered platform designed to speed up the development process by generating instant boilerplate code. It allows users to customize their tech stack, streamline development, and launch projects faster. Ideal for startups, developers, and IT agencies looking to accelerate project timelines and improve productivity.
QueryHub
QueryHub is an AI-powered web application designed to assist students in their academic endeavors. It provides a platform for users to ask and answer questions, collaborate with peers, and access instant and accurate information through AI chatbot assistance and smart search capabilities. QueryHub aims to empower students by offering a personalized learning experience, accelerating learning through document collaboration, and fostering community collaboration. With a user-friendly interface and a focus on user-driven innovation, QueryHub is a valuable tool for enhancing academic success.
Iodine Software
Iodine Software is a healthcare technology company that provides AI-enabled solutions for revenue cycle management, clinical documentation integrity, and utilization management. The company's flagship product, AwareCDI, is a suite of solutions that addresses the root causes of mid-cycle revenue leakage from admission through post-billing review. AwareCDI uses Iodine's CognitiveML AI engine to spot what is missing in patient documentation based on clinical evidence. This enables healthcare organizations to maximize documentation integrity and revenue capture. Iodine Software also offers AwareUM, a continuous, intelligent prioritization solution for peak UM performance.
Refraction
Refraction is an AI-powered code generation tool designed to help developers learn, improve, and generate code effortlessly. It offers a wide range of features such as bug detection, code conversion, function creation, CSP generation, CSS style conversion, debug statement addition, diagram generation, documentation creation, code explanation, code improvement, concept learning, CI/CD pipeline creation, SQL query generation, code refactoring, regex generation, style checking, type addition, and unit test generation. With support for 56 programming languages, Refraction is a versatile tool trusted by innovative companies worldwide to streamline software development processes using the magic of AI.
WebDB
WebDB is an open-source and efficient Database IDE that focuses on providing a secure and user-friendly platform for database management. It offers features such as automatic DBMS discovery, credential guessing, time machine for database version control, powerful queries editor with autocomplete and documentation, AI assistant integration, NoSQL structure management, intelligent data generation, and more. With a modern ERD view and support for various databases, WebDB aims to simplify database management tasks and enhance productivity for users.
BFF AI
BFF AI is a comprehensive AI-powered tool that provides a wide range of services, including text, image, and code generation, virtual assistance, speech-to-text transcription, text-to-speech conversion, and more. It is designed to help users save time, improve productivity, and enhance their creativity. With its user-friendly interface and powerful features, BFF AI is suitable for individuals, teams, and businesses of all sizes.
20 - Open Source AI Tools
WDoc
WDoc is a powerful Retrieval-Augmented Generation (RAG) system designed to summarize, search, and query documents across various file types. It supports querying tens of thousands of documents simultaneously, offers tailored summaries to efficiently manage large amounts of information, and includes features like supporting multiple file types, various LLMs, local and private LLMs, advanced RAG capabilities, advanced summaries, trust verification, markdown formatted answers, sophisticated embeddings, extensive documentation, scriptability, type checking, lazy imports, caching, fast processing, shell autocompletion, notification callbacks, and more. WDoc is ideal for researchers, students, and professionals dealing with extensive information sources.
wdoc
wdoc is a powerful Retrieval-Augmented Generation (RAG) system designed to summarize, search, and query documents across various file types. It aims to handle large volumes of diverse document types, making it ideal for researchers, students, and professionals dealing with extensive information sources. wdoc uses LangChain to process and analyze documents, supporting tens of thousands of documents simultaneously. The system includes features like high recall and specificity, support for various Language Model Models (LLMs), advanced RAG capabilities, advanced document summaries, and support for multiple tasks. It offers markdown-formatted answers and summaries, customizable embeddings, extensive documentation, scriptability, and runtime type checking. wdoc is suitable for power users seeking document querying capabilities and AI-powered document summaries.
cognita
Cognita is an open-source framework to organize your RAG codebase along with a frontend to play around with different RAG customizations. It provides a simple way to organize your codebase so that it becomes easy to test it locally while also being able to deploy it in a production ready environment. The key issues that arise while productionizing RAG system from a Jupyter Notebook are: 1. **Chunking and Embedding Job** : The chunking and embedding code usually needs to be abstracted out and deployed as a job. Sometimes the job will need to run on a schedule or be trigerred via an event to keep the data updated. 2. **Query Service** : The code that generates the answer from the query needs to be wrapped up in a api server like FastAPI and should be deployed as a service. This service should be able to handle multiple queries at the same time and also autoscale with higher traffic. 3. **LLM / Embedding Model Deployment** : Often times, if we are using open-source models, we load the model in the Jupyter notebook. This will need to be hosted as a separate service in production and model will need to be called as an API. 4. **Vector DB deployment** : Most testing happens on vector DBs in memory or on disk. However, in production, the DBs need to be deployed in a more scalable and reliable way. Cognita makes it really easy to customize and experiment everything about a RAG system and still be able to deploy it in a good way. It also ships with a UI that makes it easier to try out different RAG configurations and see the results in real time. You can use it locally or with/without using any Truefoundry components. However, using Truefoundry components makes it easier to test different models and deploy the system in a scalable way. Cognita allows you to host multiple RAG systems using one app. ### Advantages of using Cognita are: 1. A central reusable repository of parsers, loaders, embedders and retrievers. 2. Ability for non-technical users to play with UI - Upload documents and perform QnA using modules built by the development team. 3. Fully API driven - which allows integration with other systems. > If you use Cognita with Truefoundry AI Gateway, you can get logging, metrics and feedback mechanism for your user queries. ### Features: 1. Support for multiple document retrievers that use `Similarity Search`, `Query Decompostion`, `Document Reranking`, etc 2. Support for SOTA OpenSource embeddings and reranking from `mixedbread-ai` 3. Support for using LLMs using `Ollama` 4. Support for incremental indexing that ingests entire documents in batches (reduces compute burden), keeps track of already indexed documents and prevents re-indexing of those docs.
ChatData
ChatData is a robust chat-with-documents application designed to extract information and provide answers by querying the MyScale free knowledge base or uploaded documents. It leverages the Retrieval Augmented Generation (RAG) framework, millions of Wikipedia pages, and arXiv papers. Features include self-querying retriever, VectorSQL, session management, and building a personalized knowledge base. Users can effortlessly navigate vast data, explore academic papers, and research documents. ChatData empowers researchers, students, and knowledge enthusiasts to unlock the true potential of information retrieval.
llm-rag-workshop
The LLM RAG Workshop repository provides a workshop on using Large Language Models (LLMs) and Retrieval-Augmented Generation (RAG) to generate and understand text in a human-like manner. It includes instructions on setting up the environment, indexing Zoomcamp FAQ documents, creating a Q&A system, and using OpenAI for generation based on retrieved information. The repository focuses on enhancing language model responses with retrieved information from external sources, such as document databases or search engines, to improve factual accuracy and relevance of generated text.
llamabot
LlamaBot is a Pythonic bot interface to Large Language Models (LLMs), providing an easy way to experiment with LLMs in Jupyter notebooks and build Python apps utilizing LLMs. It supports all models available in LiteLLM. Users can access LLMs either through local models with Ollama or by using API providers like OpenAI and Mistral. LlamaBot offers different bot interfaces like SimpleBot, ChatBot, QueryBot, and ImageBot for various tasks such as rephrasing text, maintaining chat history, querying documents, and generating images. The tool also includes CLI demos showcasing its capabilities and supports contributions for new features and bug reports from the community.
chroma
Chroma is an open-source embedding database that provides a simple, scalable, and feature-rich way to build Python or JavaScript LLM apps with memory. It offers a fully-typed, fully-tested, and fully-documented API that makes it easy to get started and scale your applications. Chroma also integrates with popular tools like LangChain and LlamaIndex, and supports a variety of embedding models, including Sentence Transformers, OpenAI embeddings, and Cohere embeddings. With Chroma, you can easily add documents to your database, query relevant documents with natural language, and compose documents into the context window of an LLM like GPT3 for additional summarization or analysis.
workbench-example-hybrid-rag
This NVIDIA AI Workbench project is designed for developing a Retrieval Augmented Generation application with a customizable Gradio Chat app. It allows users to embed documents into a locally running vector database and run inference locally on a Hugging Face TGI server, in the cloud using NVIDIA inference endpoints, or using microservices via NVIDIA Inference Microservices (NIMs). The project supports various models with different quantization options and provides tutorials for using different inference modes. Users can troubleshoot issues, customize the Gradio app, and access advanced tutorials for specific tasks.
dynamiq
Dynamiq is an orchestration framework designed to streamline the development of AI-powered applications, specializing in orchestrating retrieval-augmented generation (RAG) and large language model (LLM) agents. It provides an all-in-one Gen AI framework for agentic AI and LLM applications, offering tools for multi-agent orchestration, document indexing, and retrieval flows. With Dynamiq, users can easily build and deploy AI solutions for various tasks.
postgresml
PostgresML is a powerful Postgres extension that seamlessly combines data storage and machine learning inference within your database. It enables running machine learning and AI operations directly within PostgreSQL, leveraging GPU acceleration for faster computations, integrating state-of-the-art large language models, providing built-in functions for text processing, enabling efficient similarity search, offering diverse ML algorithms, ensuring high performance, scalability, and security, supporting a wide range of NLP tasks, and seamlessly integrating with existing PostgreSQL tools and client libraries.
llama_ros
This repository provides a set of ROS 2 packages to integrate llama.cpp into ROS 2. By using the llama_ros packages, you can easily incorporate the powerful optimization capabilities of llama.cpp into your ROS 2 projects by running GGUF-based LLMs and VLMs.
EdgeChains
EdgeChains is an open-source chain-of-thought engineering framework tailored for Large Language Models (LLMs)- like OpenAI GPT, LLama2, Falcon, etc. - With a focus on enterprise-grade deployability and scalability. EdgeChains is specifically designed to **orchestrate** such applications. At EdgeChains, we take a unique approach to Generative AI - we think Generative AI is a deployment and configuration management challenge rather than a UI and library design pattern challenge. We build on top of a tech that has solved this problem in a different domain - Kubernetes Config Management - and bring that to Generative AI. Edgechains is built on top of jsonnet, originally built by Google based on their experience managing a vast amount of configuration code in the Borg infrastructure.
azure-functions-openai-extension
Azure Functions OpenAI Extension is a project that adds support for OpenAI LLM (GPT-3.5-turbo, GPT-4) bindings in Azure Functions. It provides NuGet packages for various functionalities like text completions, chat completions, assistants, embeddings generators, and semantic search. The project requires .NET 6 SDK or greater, Azure Functions Core Tools v4.x, and specific settings in Azure Function or local settings for development. It offers features like text completions, chat completion, assistants with custom skills, embeddings generators for text relatedness, and semantic search using vector databases. The project also includes examples in C# and Python for different functionalities.
h2ogpt
h2oGPT is an Apache V2 open-source project that allows users to query and summarize documents or chat with local private GPT LLMs. It features a private offline database of any documents (PDFs, Excel, Word, Images, Video Frames, Youtube, Audio, Code, Text, MarkDown, etc.), a persistent database (Chroma, Weaviate, or in-memory FAISS) using accurate embeddings (instructor-large, all-MiniLM-L6-v2, etc.), and efficient use of context using instruct-tuned LLMs (no need for LangChain's few-shot approach). h2oGPT also offers parallel summarization and extraction, reaching an output of 80 tokens per second with the 13B LLaMa2 model, HYDE (Hypothetical Document Embeddings) for enhanced retrieval based upon LLM responses, a variety of models supported (LLaMa2, Mistral, Falcon, Vicuna, WizardLM. With AutoGPTQ, 4-bit/8-bit, LORA, etc.), GPU support from HF and LLaMa.cpp GGML models, and CPU support using HF, LLaMa.cpp, and GPT4ALL models. Additionally, h2oGPT provides Attention Sinks for arbitrarily long generation (LLaMa-2, Mistral, MPT, Pythia, Falcon, etc.), a UI or CLI with streaming of all models, the ability to upload and view documents through the UI (control multiple collaborative or personal collections), Vision Models LLaVa, Claude-3, Gemini-Pro-Vision, GPT-4-Vision, Image Generation Stable Diffusion (sdxl-turbo, sdxl) and PlaygroundAI (playv2), Voice STT using Whisper with streaming audio conversion, Voice TTS using MIT-Licensed Microsoft Speech T5 with multiple voices and Streaming audio conversion, Voice TTS using MPL2-Licensed TTS including Voice Cloning and Streaming audio conversion, AI Assistant Voice Control Mode for hands-free control of h2oGPT chat, Bake-off UI mode against many models at the same time, Easy Download of model artifacts and control over models like LLaMa.cpp through the UI, Authentication in the UI by user/password via Native or Google OAuth, State Preservation in the UI by user/password, Linux, Docker, macOS, and Windows support, Easy Windows Installer for Windows 10 64-bit (CPU/CUDA), Easy macOS Installer for macOS (CPU/M1/M2), Inference Servers support (oLLaMa, HF TGI server, vLLM, Gradio, ExLLaMa, Replicate, OpenAI, Azure OpenAI, Anthropic), OpenAI-compliant, Server Proxy API (h2oGPT acts as drop-in-replacement to OpenAI server), Python client API (to talk to Gradio server), JSON Mode with any model via code block extraction. Also supports MistralAI JSON mode, Claude-3 via function calling with strict Schema, OpenAI via JSON mode, and vLLM via guided_json with strict Schema, Web-Search integration with Chat and Document Q/A, Agents for Search, Document Q/A, Python Code, CSV frames (Experimental, best with OpenAI currently), Evaluate performance using reward models, and Quality maintained with over 1000 unit and integration tests taking over 4 GPU-hours.
chromem-go
chromem-go is an embeddable vector database for Go with a Chroma-like interface and zero third-party dependencies. It enables retrieval augmented generation (RAG) and similar embeddings-based features in Go apps without the need for a separate database. The focus is on simplicity and performance for common use cases, allowing querying of documents with minimal memory allocations. The project is in beta and may introduce breaking changes before v1.0.0.
infinity
Infinity is a high-throughput, low-latency REST API for serving vector embeddings, supporting all sentence-transformer models and frameworks. It is developed under the MIT License and powers inference behind Gradient.ai. The API allows users to deploy models from SentenceTransformers, offers fast inference backends utilizing various accelerators, dynamic batching for efficient processing, correct and tested implementation, and easy-to-use API built on FastAPI with Swagger documentation. Users can embed text, rerank documents, and perform text classification tasks using the tool. Infinity supports various models from Huggingface and provides flexibility in deployment via CLI, Docker, Python API, and cloud services like dstack. The tool is suitable for tasks like embedding, reranking, and text classification.
marqo
Marqo is more than a vector database, it's an end-to-end vector search engine for both text and images. Vector generation, storage and retrieval are handled out of the box through a single API. No need to bring your own embeddings.
superduper
superduper.io is a Python framework that integrates AI models, APIs, and vector search engines directly with existing databases. It allows hosting of models, streaming inference, and scalable model training/fine-tuning. Key features include integration of AI with data infrastructure, inference via change-data-capture, scalable model training, model chaining, simple Python interface, Python-first approach, working with difficult data types, feature storing, and vector search capabilities. The tool enables users to turn their existing databases into centralized repositories for managing AI model inputs and outputs, as well as conducting vector searches without the need for specialized databases.
20 - OpenAI Gpts
TradeComply
Import Export Compliance | Tariff Classification | Shipping Queries | Logistics & Supply Chain Solutions
Mongoose Docs Helper
Casual, technical helper for Mongoose docs, includes documentation links.
KQL Query Helper
The KQL Query Helper GPT is tailored specifically for assisting users with Kusto Query Language (KQL) queries. It leverages extensive knowledge from Azure Data Explorer documentation to aid users in understanding, reviewing, and creating new KQL queries based on their prompts.
Query Companion
Getting ready to query agents or publishers? Upload your manuscript. I analyse your novel's writing style, themes and genre. I'll tell you how it's relevant to a modern audience, offer marketing insights and will even write you a draft synopsis and cover letter. I'll help you find relevant agents.
Big Query SQL Query Optimizer
Expert in brief, direct SQL queries for BigQuery, with casual professional tone.
OpenStreetMap Query
Helps get map data from Open Street Map by generating Overpass Turbo queries. Ask me for mapping features like cafes, rivers or highways
Power Query Assistant
Expert in Power Query and DAX for Power BI, offering in-depth guidance and insights
Search Query Optimizer
Create the most effective database or search engine queries using keywords, truncation, and Boolean operators!
BCorpGPT
Query BCorp company data. All data is publicly available. United Kingdom only (for now).
Supabase Sensei
Supabase expert also supports query generation and Flutter code generation
Your TT Ads Strategist
I'm your guide for any query and information related to TikTok Ads. Let's build your new campaign together!
Korean teacher
Answers in query language, adds Korean translation and phonetics for non-Korean queries.
AI Help BOT by IHeartDomains
Welcome to AIHelp.bot, your versatile assistant for any query. Whether it's a general knowledge question, a technical issue, or something more obscure, I'm here to help. Please type your question below, and I'll use my resources to find the best possible answer.
Ordinals API
Knows the docs and can query official ordinal endpoints—Sat Numbers, Inscription IDs, and more.