Best AI tools for< Improve Data Literacy >
20 - AI tool Sites
Grapha.ai
Grapha.ai is a data exploration tool that helps users to quickly and easily explore and understand their data. It provides a variety of features to help users to visualize their data, identify trends and patterns, and make informed decisions. Grapha.ai is designed to be easy to use, even for users with no prior experience with data analysis.
SimplyPut
SimplyPut is an AI-powered data analytics platform that allows users to ask questions about their data in natural language and get instant answers. It is designed to be intuitive and easy to use, and it can be integrated with a variety of data sources. SimplyPut is used by businesses of all sizes to improve their data literacy and make better decisions.
AI Unlock
AI Unlock is a sales training platform that focuses on teaching sales teams how to leverage AI to improve efficiency and productivity. The platform offers comprehensive training on AI fundamentals, sales techniques, industry knowledge, AI products, consultative selling, data literacy, communication skills, competitive analysis, ethical considerations, and continuous learning in AI. AI Unlock aims to help companies stay ahead in the fast-evolving AI industry by providing effective, fast, and low-cost AI training programs. The platform also offers on-going training, updates, and a community of AI sales professionals to support continuous learning and development.
Storywizard.ai
Storywizard.ai is an AI-powered platform that helps students create and refine their own fully illustrated and editable stories in any of eight languages. It also provides teachers with tools to create and track customized reading and writing assignments. Storywizard.ai is designed to help students improve their literacy skills, engage with fun and educational exercises, and increase their motivation to learn.
DQLabs
DQLabs is a modern data quality platform that leverages observability to deliver reliable and accurate data for better business outcomes. It combines the power of Data Quality and Data Observability to enable data producers, consumers, and leaders to achieve decentralized data ownership and turn data into action faster, easier, and more collaboratively. The platform offers features such as data observability, remediation-centric data relevance, decentralized data ownership, enhanced data collaboration, and AI/ML-enabled semantic data discovery.
Datuum
Datuum is an AI-powered data onboarding solution that offers seamless integration for businesses. It simplifies the data onboarding process by automating manual tasks, generating code, and ensuring data accuracy with AI-driven validation. Datuum helps businesses achieve faster time to value, reduce costs, improve scalability, and enhance data quality and consistency.
Coginiti
Coginiti is a collaborative analytics platform and tools designed for SQL developers, data scientists, engineers, and analysts. It offers capabilities such as AI assistant, data mesh, database & object store support, powerful query & analysis, and share & reuse curated assets. Coginiti empowers teams and organizations to manage collaborative practices, data efficiency, and deliver trusted data products faster. The platform integrates modular analytic development, collaborative versioned teamwork, and a data quality framework to enhance productivity and ensure data reliability. Coginiti also provides an AI-enabled virtual analytics advisor to boost team efficiency and empower data heroes.
PaperEntry AI
Deep Cognition offers PaperEntry AI, an Intelligent Document Processing solution powered by generative AI. It automates data entry tasks with high accuracy, scalability, and configurability, handling complex documents of any type or format. The application is trusted by leading global organizations for customs clearance automation and government document processing, delivering significant time and cost savings. With industry-specific features and a proven track record, Deep Cognition provides a state-of-the-art solution for businesses seeking efficient data extraction and automation.
Eigen Technologies
Eigen Technologies is an AI-powered data extraction platform designed for business users to automate the extraction of data from various documents. The platform offers solutions for intelligent document processing and automation, enabling users to streamline business processes, make informed decisions, and achieve significant efficiency gains. Eigen's platform is purpose-built to deliver real ROI by reducing manual processes, improving data accuracy, and accelerating decision-making across industries such as corporates, banks, financial services, insurance, law, and manufacturing. With features like generative insights, table extraction, pre-processing hub, and model governance, Eigen empowers users to automate data extraction workflows efficiently. The platform is known for its unmatched accuracy, speed, and capability, providing customers with a flexible and scalable solution that integrates seamlessly with existing systems.
Seudo
Seudo is a data workflow automation platform that uses AI to help businesses automate their data processes. It provides a variety of features to help businesses with data integration, data cleansing, data transformation, and data analysis. Seudo is designed to be easy to use, even for businesses with no prior experience with AI. It offers a drag-and-drop interface that makes it easy to create and manage data workflows. Seudo also provides a variety of pre-built templates that can be used to get started quickly.
Sicara
Sicara is a data and AI expert platform that helps clients define and implement data strategies, build data platforms, develop data science products, and automate production processes with computer vision. They offer services to improve data performance, accelerate data use cases, integrate generative AI, and support ESG transformation. Sicara collaborates with technology partners to provide tailor-made solutions for data and AI challenges. The platform also features a blog, job offers, and a team of experts dedicated to enhancing productivity and quality in data projects.
nuvo
nuvo is an AI-powered data import solution that offers fast, secure, and scalable data import solutions for software companies. It provides tools like nuvo Data Importer SDK and nuvo Data Pipeline to streamline manual and recurring ETL data imports, enabling users to manage data imports independently. With AI-enhanced automation, nuvo helps prepare clean data for preferred systems quickly and efficiently, reducing manual effort and improving data quality. The platform allows users to upload unlimited data in various formats, match imported data to system schemas, clean and validate data, and import clean data into target systems with just a click.
Roundtable
Roundtable is an AI-assisted data cleaning tool designed for enterprise survey programming. It offers an easy-to-integrate API for cleaning open-ended survey responses, saving up to 70% of time. The tool uses real-time behavioral tracking to detect unnatural typing and programmatic entries, and it provides multilingual functionality for deploying studies to various markets. Roundtable also features GPT detection to identify bots and participants, dynamic clustering to group duplicate responses, and programmatic pre-screening to auto-reject low-quality participants. The tool is trusted by leaders and innovators for improving data quality efforts and providing reliable human-generated insights.
Swyft AI
Swyft AI is an AI-powered tool that automates CRM data capture and sales motions. It integrates with popular CRM and web conferencing tools, allowing sales teams to save time and improve data accuracy. Swyft AI's key features include automatic CRM data capture, workflow automation, and call summarization. It offers advantages such as improved data hygiene, increased sales productivity, and reduced manual work for revenue teams.
Prompt Security
Prompt Security is a platform that secures all uses of Generative AI in the organization: from tools used by your employees to your customer-facing apps.
Base64.ai
Base64.ai is an automated document processing API that offers a leading no-code AI solution for understanding documents, photos, and videos. It provides a comprehensive set of features for document processing across various industries, with a strong focus on accuracy, security, and extensibility. Base64.ai is designed to streamline document automation processes and improve data extraction efficiency.
Momentum
Momentum is an AI-powered sales automation tool that extracts and structures data from sales and customer service calls, providing actionable insights that sync directly with CRM systems like Salesforce. It automates note-taking, call summaries, and workflow approvals, enhancing sales efficiency and customer engagement. Momentum integrates with popular tools like Slack and Zoom to streamline communication and provide valuable insights to revenue teams. The platform aims to optimize sales processes, improve data capture, and enhance coaching scalability for high-performing revenue teams.
Enlitic
Enlitic provides healthcare data solutions that leverage artificial intelligence to improve data management, clinical workflows, and create a foundation for real-world evidence medical image databases. Their products, ENDEX and ENCOG, utilize computer vision and natural language processing to standardize, protect, and analyze medical imaging data, enabling healthcare providers to optimize workflows, increase efficiencies, and expand capacity.
PageWise AI
PageWise AI is an AI application that provides instant insight into Confluence data by turning tedious searches and repetitive questions into intuitive AI-powered conversations directly in Slack. It offers effortless integration with Slack and Confluence accounts, eliminates the need for sifting through pages, reduces repetitive queries, connects hidden insights within Confluence pages, boosts communication, and scales with business needs. PageWise AI offers flexible pricing plans catering to different team sizes and requirements, with the option for custom integrations, data sources, and self-hosted solutions for enterprise users.
Coefficient
Coefficient is a data connector tool for Google Sheets and Excel that empowers users to unlock the full potential of their spreadsheets by providing live connections to over 60 business systems. It offers features such as two-way syncing, automatic refreshes, and scheduled alerts, all powered by AI. Coefficient streamlines data import, visualization, and system updates, enabling users to eliminate manual work, improve data accuracy, and save time. Trusted by over 50,000 companies, Coefficient is a user-friendly tool that enhances data analysis and reporting processes.
20 - Open Source AI Tools
LLM-Agents-Papers
A repository that lists papers related to Large Language Model (LLM) based agents. The repository covers various topics including survey, planning, feedback & reflection, memory mechanism, role playing, game playing, tool usage & human-agent interaction, benchmark & evaluation, environment & platform, agent framework, multi-agent system, and agent fine-tuning. It provides a comprehensive collection of research papers on LLM-based agents, exploring different aspects of AI agent architectures and applications.
Webscout
WebScout is a versatile tool that allows users to search for anything using Google, DuckDuckGo, and phind.com. It contains AI models, can transcribe YouTube videos, generate temporary email and phone numbers, has TTS support, webai (terminal GPT and open interpreter), and offline LLMs. It also supports features like weather forecasting, YT video downloading, temp mail and number generation, text-to-speech, advanced web searches, and more.
EasyEdit
EasyEdit is a Python package for edit Large Language Models (LLM) like `GPT-J`, `Llama`, `GPT-NEO`, `GPT2`, `T5`(support models from **1B** to **65B**), the objective of which is to alter the behavior of LLMs efficiently within a specific domain without negatively impacting performance across other inputs. It is designed to be easy to use and easy to extend.
ProX
ProX is a lm-based data refinement framework that automates the process of cleaning and improving data used in pre-training large language models. It offers better performance, domain flexibility, efficiency, and cost-effectiveness compared to traditional methods. The framework has been shown to improve model performance by over 2% and boost accuracy by up to 20% in tasks like math. ProX is designed to refine data at scale without the need for manual adjustments, making it a valuable tool for data preprocessing in natural language processing tasks.
text-to-sql-bedrock-workshop
This repository focuses on utilizing generative AI to bridge the gap between natural language questions and SQL queries, aiming to improve data consumption in enterprise data warehouses. It addresses challenges in SQL query generation, such as foreign key relationships and table joins, and highlights the importance of accuracy metrics like Execution Accuracy (EX) and Exact Set Match Accuracy (EM). The workshop content covers advanced prompt engineering, Retrieval Augmented Generation (RAG), fine-tuning models, and security measures against prompt and SQL injections.
moai
moai is a PyTorch-based AI Model Development Kit (MDK) designed to improve data-driven model workflows, design, and understanding. It offers modularity via monads for model building blocks, reproducibility via configuration-based design, productivity via a data-driven domain modelling language (DML), extensibility via plugins, and understanding via inter-model performance and design aggregation. The tool provides specific integrated actions like play, train, evaluate, plot, diff, and reprod to support heavy data-driven workflows with analytics, knowledge extraction, and reproduction. moai relies on PyTorch, Lightning, Hydra, TorchServe, ONNX, Visdom, HiPlot, Kornia, Albumentations, and the wider open-source community for its functionalities.
Reflection_Tuning
Reflection-Tuning is a project focused on improving the quality of instruction-tuning data through a reflection-based method. It introduces Selective Reflection-Tuning, where the student model can decide whether to accept the improvements made by the teacher model. The project aims to generate high-quality instruction-response pairs by defining specific criteria for the oracle model to follow and respond to. It also evaluates the efficacy and relevance of instruction-response pairs using the r-IFD metric. The project provides code for reflection and selection processes, along with data and model weights for both V1 and V2 methods.
Streamline-Analyst
Streamline Analyst is a cutting-edge, open-source application powered by Large Language Models (LLMs) designed to revolutionize data analysis. This Data Analysis Agent effortlessly automates tasks such as data cleaning, preprocessing, and complex operations like identifying target objects, partitioning test sets, and selecting the best-fit models based on your data. With Streamline Analyst, results visualization and evaluation become seamless. It aims to expedite the data analysis process, making it accessible to all, regardless of their expertise in data analysis. The tool is built to empower users to process data and achieve high-quality visualizations with unparalleled efficiency, and to execute high-performance modeling with the best strategies. Future enhancements include Natural Language Processing (NLP), neural networks, and object detection utilizing YOLO, broadening its capabilities to meet diverse data analysis needs.
sparrow
Sparrow is an innovative open-source solution for efficient data extraction and processing from various documents and images. It seamlessly handles forms, invoices, receipts, and other unstructured data sources. Sparrow stands out with its modular architecture, offering independent services and pipelines all optimized for robust performance. One of the critical functionalities of Sparrow - pluggable architecture. You can easily integrate and run data extraction pipelines using tools and frameworks like LlamaIndex, Haystack, or Unstructured. Sparrow enables local LLM data extraction pipelines through Ollama or Apple MLX. With Sparrow solution you get API, which helps to process and transform your data into structured output, ready to be integrated with custom workflows. Sparrow Agents - with Sparrow you can build independent LLM agents, and use API to invoke them from your system. **List of available agents:** * **llamaindex** - RAG pipeline with LlamaIndex for PDF processing * **vllamaindex** - RAG pipeline with LLamaIndex multimodal for image processing * **vprocessor** - RAG pipeline with OCR and LlamaIndex for image processing * **haystack** - RAG pipeline with Haystack for PDF processing * **fcall** - Function call pipeline * **unstructured-light** - RAG pipeline with Unstructured and LangChain, supports PDF and image processing * **unstructured** - RAG pipeline with Weaviate vector DB query, Unstructured and LangChain, supports PDF and image processing * **instructor** - RAG pipeline with Unstructured and Instructor libraries, supports PDF and image processing. Works great for JSON response generation
tensorzero
TensorZero is an open-source platform that helps LLM applications graduate from API wrappers into defensible AI products. It enables a data & learning flywheel for LLMs by unifying inference, observability, optimization, and experimentation. The platform includes a high-performance model gateway, structured schema-based inference, observability, experimentation, and data warehouse for analytics. TensorZero Recipes optimize prompts and models, and the platform supports experimentation features and GitOps orchestration for deployment.
swirl-search
Swirl is an open-source software that allows users to simultaneously search multiple content sources and receive AI-ranked results. It connects to various data sources, including databases, public data services, and enterprise sources, and utilizes AI and LLMs to generate insights and answers based on the user's data. Swirl is easy to use, requiring only the download of a YML file, starting in Docker, and searching with Swirl. Users can add credentials to preloaded SearchProviders to access more sources. Swirl also offers integration with ChatGPT as a configured AI model. It adapts and distributes user queries to anything with a search API, re-ranking the unified results using Large Language Models without extracting or indexing anything. Swirl includes five Google Programmable Search Engines (PSEs) to get users up and running quickly. Key features of Swirl include Microsoft 365 integration, SearchProvider configurations, query adaptation, synchronous or asynchronous search federation, optional subscribe feature, pipelining of Processor stages, results stored in SQLite3 or PostgreSQL, built-in Query Transformation support, matching on word stems and handling of stopwords, duplicate detection, re-ranking of unified results using Cosine Vector Similarity, result mixers, page through all results requested, sample data sets, optional spell correction, optional search/result expiration service, easily extensible Connector and Mixer objects, and a welcoming community for collaboration and support.
Taiyi-LLM
Taiyi (太一) is a bilingual large language model fine-tuned for diverse biomedical tasks. It aims to facilitate communication between healthcare professionals and patients, provide medical information, and assist in diagnosis, biomedical knowledge discovery, drug development, and personalized healthcare solutions. The model is based on the Qwen-7B-base model and has been fine-tuned using rich bilingual instruction data. It covers tasks such as question answering, biomedical dialogue, medical report generation, biomedical information extraction, machine translation, title generation, text classification, and text semantic similarity. The project also provides standardized data formats, model training details, model inference guidelines, and overall performance metrics across various BioNLP tasks.
LLMs
LLMs is a Chinese large language model technology stack for practical use. It includes high-availability pre-training, SFT, and DPO preference alignment code framework. The repository covers pre-training data cleaning, high-concurrency framework, SFT dataset cleaning, data quality improvement, and security alignment work for Chinese large language models. It also provides open-source SFT dataset construction, pre-training from scratch, and various tools and frameworks for data cleaning, quality optimization, and task alignment.
distilabel
Distilabel is a framework for synthetic data and AI feedback for AI engineers that require high-quality outputs, full data ownership, and overall efficiency. It helps you synthesize data and provide AI feedback to improve the quality of your AI models. With Distilabel, you can: * **Synthesize data:** Generate synthetic data to train your AI models. This can help you to overcome the challenges of data scarcity and bias. * **Provide AI feedback:** Get feedback from AI models on your data. This can help you to identify errors and improve the quality of your data. * **Improve your AI output quality:** By using Distilabel to synthesize data and provide AI feedback, you can improve the quality of your AI models and get better results.
data-juicer
Data-Juicer is a one-stop data processing system to make data higher-quality, juicier, and more digestible for LLMs. It is a systematic & reusable library of 80+ core OPs, 20+ reusable config recipes, and 20+ feature-rich dedicated toolkits, designed to function independently of specific LLM datasets and processing pipelines. Data-Juicer allows detailed data analyses with an automated report generation feature for a deeper understanding of your dataset. Coupled with multi-dimension automatic evaluation capabilities, it supports a timely feedback loop at multiple stages in the LLM development process. Data-Juicer offers tens of pre-built data processing recipes for pre-training, fine-tuning, en, zh, and more scenarios. It provides a speedy data processing pipeline requiring less memory and CPU usage, optimized for maximum productivity. Data-Juicer is flexible & extensible, accommodating most types of data formats and allowing flexible combinations of OPs. It is designed for simplicity, with comprehensive documentation, easy start guides and demo configs, and intuitive configuration with simple adding/removing OPs from existing configs.
oci-data-science-ai-samples
The Oracle Cloud Infrastructure Data Science and AI services Examples repository provides demos, tutorials, and code examples showcasing various features of the OCI Data Science service and AI services. It offers tools for data scientists to develop and deploy machine learning models efficiently, with features like Accelerated Data Science SDK, distributed training, batch processing, and machine learning pipelines. Whether you're a beginner or an experienced practitioner, OCI Data Science Services provide the resources needed to build, train, and deploy models easily.
indexify
Indexify is an open-source engine for building fast data pipelines for unstructured data (video, audio, images, and documents) using reusable extractors for embedding, transformation, and feature extraction. LLM Applications can query transformed content friendly to LLMs by semantic search and SQL queries. Indexify keeps vector databases and structured databases (PostgreSQL) updated by automatically invoking the pipelines as new data is ingested into the system from external data sources. **Why use Indexify** * Makes Unstructured Data **Queryable** with **SQL** and **Semantic Search** * **Real-Time** Extraction Engine to keep indexes **automatically** updated as new data is ingested. * Create **Extraction Graph** to describe **data transformation** and extraction of **embedding** and **structured extraction**. * **Incremental Extraction** and **Selective Deletion** when content is deleted or updated. * **Extractor SDK** allows adding new extraction capabilities, and many readily available extractors for **PDF**, **Image**, and **Video** indexing and extraction. * Works with **any LLM Framework** including **Langchain**, **DSPy**, etc. * Runs on your laptop during **prototyping** and also scales to **1000s of machines** on the cloud. * Works with many **Blob Stores**, **Vector Stores**, and **Structured Databases** * We have even **Open Sourced Automation** to deploy to Kubernetes in production.
xGitGuard
xGitGuard is an AI-based system developed by Comcast Cybersecurity Research and Development team to detect secrets (e.g., API tokens, usernames, passwords) exposed on GitHub repositories. It uses advanced Natural Language Processing to detect secrets at scale and with appropriate velocity. The tool provides workflows for detecting credentials and keys/tokens in both enterprise and public GitHub accounts. Users can set up search patterns, configure API access, run detections with or without ML filters, and train ML models for improved detection accuracy. xGitGuard also supports custom keyword scans for targeted organizations or repositories. The tool is licensed under Apache 2.0.
20 - OpenAI Gpts
👑 Data Privacy for Language & Training Centers 👑
Language and Skill Training Centers collect personal information of learners, including progress tracking and sometimes payment details.
Missing Cluster Identification Program
I analyze and integrate missing clusters in data for coherent structuring.
Mermaid Architect GPT | 💡 -> 👁
Turn your projects' Ideas into Clear Flowcharts(data flow) with Recommended Tech Stack
FAANG.AI
Get into FAANG. Practice with an AI expert in algorithms, data structures, and system design. Do a mock interview and improve.
Palm Reader
Moved to https://chat.openai.com/g/g-KFnF7qssT-palm-reader . Interprets palm readings from user-uploaded hand images. Turned off setting to use data for OpenAi to improve model.
Face Reader
Moved to https://chat.openai.com/g/g-q6GNcOkYx-face-reader. Reads faces to tell fortunes based on Chinese face reading. Turned off setting to use data for OpenAi to improve model.
Deal Architect
Designing Strategic M&A Blueprints for Success in buying, selling or merging companies. Use this GPT to simplify, speed up and improve the quality of the M&A process. With custom data - 100s of creative options in deal flow, deal structuring, financing and more. **Version 2.2 - 28012024**
📩 メール執筆・校正アシスタント【✅セキュリティ強化済み】
メールの作成・添削・返信の提案などを行います。メールの校正は、草案をそのままコピペするだけで可能です。件名・署名・添付ファイルなどを忘れないようリマインドも行います。さらにこのGPTは、特別な設定で「Use conversation data in your GPT to improve our models」の項目をOFFにしています。そのため、うっかり機密情報が含まれるメール文を送信してしまった場合でも、ChatGPTの学習に利用されないと思われます。 ※ただし、今後の仕様変更や他の経路からの情報漏洩などのリスクもありえます。個人情報は決して書き込まないでください。
Coach Gestion Data
Collecte et analyse de données sur la résilience aux catastrophes naturelles.
👑 Data Privacy for Pet Grooming Services 👑
Pet Grooming and Boarding Services store pet owner contact information, pet health data, and service preferences.
Customer Retention Consultant
Analyzes customer churn and provides strategies to improve loyalty and retention.
Data Analysis - SERP
it helps me analyze serp results and data from certain websites in order to create an outline for the writer
Data-Driven Messaging Campaign Generator
Create, analyze & duplicate customized automated message campaigns to boost retention & drive revenue for your website or app