CrackSQL
LLM-based Dialect Translation System
Stars: 63
CrackSQL is a powerful SQL dialect translation tool that integrates rule-based strategies with large language models (LLMs) for high accuracy. It enables seamless conversion between dialects (e.g., PostgreSQL โ MySQL) with flexible access through Python API, command line, and web interface. The tool supports extensive dialect compatibility, precision & advanced processing, and versatile access & integration. It offers three modes for dialect translation and demonstrates high translation accuracy over collected benchmarks. Users can deploy CrackSQL using PyPI package installation or source code installation methods. The tool can be extended to support additional syntax, new dialects, and improve translation efficiency. The project is actively maintained and welcomes contributions from the community.
README:
๐ Unlock seamless SQL translation โ effortless, precise, and efficient across databases~ ๐ฌ
Demo โข Quick Start โข Feature Extension โข FAQ โข Community โข Contributors โข License
English | ็ฎไฝไธญๆ
Star โญ and subscribe ๐ for the latest features and improvements!
CrackSQL is a powerful SQL dialect translation tool that integrates rule-based strategies with LLMs for high accuracy. It enables seamless conversion between dialects (e.g., PostgreSQL โ MySQL) with flexible access through Python API, command line, and web interface.
- 09/2025: Our newly-developed benchmark PARROT (i.e., the first SQL-to-SQL benchmark) has been accepted by NeurIPS 2025 ! Welcome your submission to challenge our leaderboard! โ๏ธ ๐ ๐ฅ
- 06/2025: We have been invited by the Ploutos community to present CrackSQL at their livestream event! ๐ฅ
- 04/2025: We have released our demo paper about this project that can be found online! ๐
- 03/2025: We have refactored the code and released our project across multiple open-source platforms (PyPI). We are currently working on new features and more contributors are welcomed! ๐ ๐ซ
- 02/2025: Our paper "Cracking SQL Barrier: An LLM-based Dialect Translation System" has been accepted by SIGMOD 2025! ๐ ๐ ๐
- ๐ Extensive Dialect Compatibility: Effortlessly translates between PostgreSQL, MySQL, and Oracle with tailored flexible strategies.
- ๐ฏ Precision & Advanced Processing: Achieves flawless translations with function-oriented query handling and cutting-edge model-based syntax matching through an adaptive local-to-global iteration strategy.
- ๐ Versatile Access & Integration: Seamlessly integrates with Python API, command line, and web interface to meet all user requirements.
Currently, CrackSQL has integrated three modes for dialect translation, adopting the rules from SQLGlot and supporting a wide range of large language models (LLMs), including prevalent models like GPT as well as the recent DeepSeek.
| Mode | SQL Dialect | LLM (w/o & w fine-tuned) |
Embedding Model (w/o & w fine-tuned) |
||
|---|---|---|---|---|---|
| Cloud Service (e.g., ๐ฌ GPT series) |
Local Deployed (e.g., ๐ค Hugging Face) |
Cloud Service (e.g., ๐ฌ GPT series) |
Local Deployed (e.g., ๐ค Hugging Face) |
||
| Rule-only | 24 | - | - | - | - |
| LLM-direct | โ | โ | โ | - | - |
| Rule+LLM | 3 (pg/mysql/oracle) |
โ | โ | โ | โ |
Additionally, the prerequisites for each mode are listed below, where SQL Parser (ANTLR) and Dialect Specification have already been provided. Please refer to Feature Extension section to customize and enhance CrackSQL to make it more powerful for your own cases.
| Mode | SQL Dialect | Model Service | |||
|---|---|---|---|---|---|
| SQL Parser | Dialect Specification | Database Connection | LLM | Embedding Model | |
| Rule-only | โ
(SQLGlot) |
- | - | - | - |
| LLM-direct | - | - | - | โ | - |
| Rule+LLM | โ
(ANTLR) |
โ | โ / - | โ | โ / - |
The following table demonstrates the translation accuracy (%) of different methods over our collected benchmark (N/A denotes the dialect translation is not supported in Ora2Pg).
- (1) $Acc_{EX}$ indicates the translated SQL is syntactically correct and executable over the target database.
- (2) $Acc_{RES}$ represents the translated SQL delivers exactly the same result (including the displayed order) as the original ones.
Note that the required translation duration is highly dependent on the SQL complexity (e.g., the number of SQL syntax piece to be translated) and can vary from several seconds to minutes.
| Method | PG โ MySQL | PG โ MySQL | MySQL โ PG | MySQL โ PG | PG โ Oracle | PG โ Oracle |
|---|---|---|---|---|---|---|
| $Acc_{EX}$ | $Acc_{RES}$ | $Acc_{EX}$ | $Acc_{RES}$ | $Acc_{EX}$ | $Acc_{RES}$ | |
| SQLGlot | 74.19 | 70.97 | 60.32 | 60.32 | 55.81 | 53.49 |
| jOOQ | 70.97 | 70.97 | 39.68 | 39.68 | 62.79 | 60.47 |
| Ora2Pg | N/A | N/A | 33.33 | 33.33 | N/A | N/A |
| SQLines | 9.68 | 9.68 | 31.75 | 31.75 | 53.49 | 48.84 |
| GPT-4o | 61.29 | 61.29 | 50.79 | 44.44 | 60.47 | 55.81 |
| CrackSQL (Ours) | 87.1 | 74.19 | 85.71 | 79.37 | 69.77 | 67.44 |
The following showcases the primary pages of the CrackSQL interface service, including the service guidance homepage and detailed translation process.
- Homepage of the deployed translation service:
- Detailed translation process of specific translation pair:
We have currently offered two methods (i.e., PyPI package and source code installation) to deploy CrackSQL.
- Install the PyPI package at the official website.
# create virtual environment
conda create -n CrackSQL python=3.10
conda activate CrackSQL
# install PyPI package
pip install cracksql
- Run with the PyPI package. A running code example using this package is presented below:
from cracksql.cracksql import translate, initkb
def initkb_func():
try:
initkb("./init_config.yaml") # fill the basic configurations in the `.yaml` first
print("Knowledge base initialized successfully")
except Exception as e:
print(f"Knowledge base initialization failed: {str(e)}")
import traceback
traceback.print_exc()
def trans_func():
target_db_config = {
"host": "target database host",
"port": target database number (int type),
"user": "target database username",
"password": "target database password",
"db_name": "target database database name"
}
vector_config = {
"src_kb_name": "source database knowledge base name",
"tgt_kb_name": "target database knowledge base name"
}
try:
print("Starting SQL translation...")
translated_sql, model_ans_list, used_pieces, lift_histories = translate(
model_name="DeepSeek-R1-Distill-Qwen-32B",
src_sql='SELECT DISTINCT "t1"."id" , EXTRACT(YEAR FROM CURRENT_TIMESTAMP) - EXTRACT(YEAR FROM CAST( "t1"."birthday" AS TIMESTAMP )) FROM "patient" AS "t1" INNER JOIN "examination" AS "t2" ON "t1"."id" = "t2"."id" WHERE "t2"."rvvt" = "+"',
src_dialect="postgresql",
tgt_dialect="mysql",
target_db_config=target_db_config,
vector_config=vector_config,
out_dir="./",
retrieval_on=False,
top_k=3
)
print("Translation completed!")
print(f"Translated SQL: {translated_sql}")
print(f"Model answer list: {model_ans_list}")
print(f"Used knowledge pieces: {used_pieces}")
print(f"Lift histories: {lift_histories}")
except Exception as e:
print(f"Error occurred during translation: {str(e)}")
import traceback
traceback.print_exc()
if __name__ == "__main__":
initkb_func()
trans_func()git clone https://github.com/weAIDB/CrackSQL.git# Start backend
cd CrackSQL/backend
# Install dependencies
conda create -n CrackSQL python=3.10
conda activate CrackSQL
pip install -r requirements.txt
# Initialize database
flask db init # Initialize
flask db migrate # Generate version file
flask db upgrade # Synchronize to database
# Initialize knowledge base (Optional, can be done manually in the frontend after starting the frontend project)
# 1. First rename config/init_config.yaml.copy to config/init_config.yaml
# 2. Modify the relevant information in config/init_config.yaml. If you want to initialize the knowledge base, Embedding Model is required
python init_knowledge_base.py --config_file xxxxxxxxx
# Start backend service (The backend service port can also be modified in app.py, currently 30006)
python app.py
# Open a new terminal window, start frontend (requires nodejs, version 20.11.1+)
cd CrackSQL/webui
# Install dependencies
pnpm install
# Start development server
pnpm run dev
# Visit http://localhost:50212 to use the Web interface
# Tips:
# If you want to modify the frontend port number, you can modify it in webui/vite.config.js: port: 50212
# If the backend API port number has been changed, or you want to use the server's IP, you can modify the VITE_APP_BASE_URL parameter in webui/.env.serve-dev file (if the file does not exist, you can rename webui/.env.serve-dev_template to .env.serve-dev).# Initialize knowledge base (Optional, can be done manually in the frontend after starting the frontend project)
# 1. First rename config/init_config.yaml.copy to config/init_config.yaml
# 2. Modify the relevant information in config/init_config.yaml. If you want to initialize the knowledge base, Embedding Model is required
python init_knowledge_base.py --init_all
# Translate
# specify the required configurations displayed by `--help` command
python translate.py --helpTo complement additional syntax, you can modify the .g4 files in ANTLR according to the grammar rules shown below.
In this grammar, each parsing rule is structured recursively and consists of both non-terminal and terminal tokens.
Once your .g4 files are prepared, you can use the official ANTLR tool to generate an updated Python parser for integration into CrackSQL.
sql_script
: sql_plus_command_no_semicolon? (
(sql_plus_command | unit_statement) (SEMICOLON '/'? (sql_plus_command | unit_statement))* SEMICOLON? '/'?
) EOF
;
......
To complement additional specification, you can append new specifications to a .json file in the following format.
[
{
"keyword": "the SQL snippet, REQUIRED",
"type": "function/keyword/type/operator, REQUIRED",
"tree": "syntax tree generated by SQL parser, REQUIRED",
"description": "brief usage description, REQUIRED",
"detail": "detailed usage illustration, REQUIRED (empty string if None)",
"link": ["link1", "link2", "link3"],
"example": ["example1", "example2", "example3"]
},
{
......
}
]Enabling CrackSQL to support new dialects requires two key components: (1) a dialect syntax parser and (2) functionality specifications.
You can start by checking the official ANTLR repository to see if the desired dialect grammar (i.e., ANTLR .g4 files) is already available.
If the required grammar does not exist, you need to compose the corresponding ANTLR grammar files to build the SQL syntax parser.
Once the .g4 files are ready, you can use the official ANTLR tool to generate an updated Python parser.
This parser can then be integrated into CrackSQL.
You need to transform the functionality specifications (e.g., the Oracle function descriptions) into a .json file.
In this file, each item should be organized according to the following format.
[
{
"keyword": "the SQL snippet, REQUIRED",
"type": "snippet type from four options: '(1) function, or (2) keyword, or (3) type, or (4) operator', REQUIRED",
"tree": "syntax tree generated by SQL parser, REQUIRED",
"description": "brief usage description, REQUIRED",
"detail": "detailed usage illustration, REQUIRED (empty string if None)",
"link": ["link1", "link2", "link3"],
"example": ["example1", "example2", "example3"]
},
{
......
}
]Q: How to make CrackSQL support additional syntax or new dialect?
A: To support additional syntax, you need to modify the `.g4` files in ANTLR and then generate an updated Python parser. Moreover, you should provide the corresponding dialect specifications for the newly-added syntax.To support new dialect, two key components (i.e., dialect syntax parser and functionality specifications) are required. Currently, the syntax parser is created based on ANTLR grammar, and the specifications can be derived from processing official documents.
For more detailed information, please refer to the Feature Extension section.
- Effective Artifact Release: We are currently preparing our MoE-based cross-dialect embedding models for practical usage and intend to release them on Hugging Face soon.
- Comprehensive Dialect Support: We will support more dialects with prepared syntax parser and functionality specifications, which is a longstanding work and more contributors are welcomed!
- Translation Efficiency Improvement: We aim to implement the rules discovered by LLMs into rule systems, thus reducing the LLM invocation overhead.
We deeply appreciate the invaluable effort contributed by our dedicated team of developers, supportive users, and esteemed industry partners.
If you like this project, please cite our paper:
@article{zhou2025cracksql,
author = {Wei Zhou and
Yuyang Gao and
Xuanhe Zhou and
Guoliang Li},
title = {{Cracking SQL Barriers:} {An} LLM-based Dialect Transaltion System},
journal = {Proc. {ACM} Manag. Data},
volume = {3},
number = {3 (SIGMOD)},
year = {2025}
}
@article{zhou2025cracksqldemo,
author = {Wei Zhou and
Yuyang Gao and
Xuanhe Zhou and
Guoliang Li},
title = {CrackSQL: A Hybrid SQL Dialect Translation System Powered by Large Language Models},
journal = {arXiv Preprint},
url = {https://arxiv.org/abs/2504.00882},
year = {2025}
}
This project is licensed under the MIT License - see the LICENSE file for details
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for CrackSQL
Similar Open Source Tools
CrackSQL
CrackSQL is a powerful SQL dialect translation tool that integrates rule-based strategies with large language models (LLMs) for high accuracy. It enables seamless conversion between dialects (e.g., PostgreSQL โ MySQL) with flexible access through Python API, command line, and web interface. The tool supports extensive dialect compatibility, precision & advanced processing, and versatile access & integration. It offers three modes for dialect translation and demonstrates high translation accuracy over collected benchmarks. Users can deploy CrackSQL using PyPI package installation or source code installation methods. The tool can be extended to support additional syntax, new dialects, and improve translation efficiency. The project is actively maintained and welcomes contributions from the community.
LTEngine
LTEngine is a free and open-source local AI machine translation API written in Rust. It is self-hosted and compatible with LibreTranslate. LTEngine utilizes large language models (LLMs) via llama.cpp, offering high-quality translations that rival or surpass DeepL for certain languages. It supports various accelerators like CUDA, Metal, and Vulkan, with the largest model 'gemma3-27b' fitting on a single consumer RTX 3090. LTEngine is actively developed, with a roadmap outlining future enhancements and features.
logicstamp-context
LogicStamp Context is a static analyzer that extracts deterministic component contracts from TypeScript codebases, providing structured architectural context for AI coding assistants. It helps AI assistants understand architecture by extracting props, hooks, and dependencies without implementation noise. The tool works with React, Next.js, Vue, Express, and NestJS, and is compatible with various AI assistants like Claude, Cursor, and MCP agents. It offers features like watch mode for real-time updates, breaking change detection, and dependency graph creation. LogicStamp Context is a security-first tool that protects sensitive data, runs locally, and is non-opinionated about architectural decisions.
llm.rb
llm.rb is a zero-dependency Ruby toolkit for Large Language Models that includes various providers like OpenAI, Gemini, Anthropic, xAI (Grok), zAI, DeepSeek, Ollama, and LlamaCpp. It provides full support for chat, streaming, tool calling, audio, images, files, and structured outputs. The toolkit offers features like unified API across providers, pluggable JSON adapters, tool calling, JSON Schema structured output, streaming responses, TTS, transcription, translation, image generation, files API, multimodal prompts, embeddings, models API, OpenAI vector stores, and more.
llm
llm.rb is a zero-dependency Ruby toolkit for Large Language Models that includes OpenAI, Gemini, Anthropic, xAI (Grok), DeepSeek, Ollama, and LlamaCpp. The toolkit provides full support for chat, streaming, tool calling, audio, images, files, and structured outputs (JSON Schema). It offers a single unified interface for multiple providers, zero dependencies outside Ruby's standard library, smart API design, and optional per-provider process-wide connection pool. Features include chat, agents, media support (text-to-speech, transcription, translation, image generation, editing), embeddings, model management, and more.
LightRAG
LightRAG is a repository hosting the code for LightRAG, a system that supports seamless integration of custom knowledge graphs, Oracle Database 23ai, Neo4J for storage, and multiple file types. It includes features like entity deletion, batch insert, incremental insert, and graph visualization. LightRAG provides an API server implementation for RESTful API access to RAG operations, allowing users to interact with it through HTTP requests. The repository also includes evaluation scripts, code for reproducing results, and a comprehensive code structure.
honey
Bee is an ORM framework that provides easy and high-efficiency database operations, allowing developers to focus on business logic development. It supports various databases and features like automatic filtering, partial field queries, pagination, and JSON format results. Bee also offers advanced functionalities like sharding, transactions, complex queries, and MongoDB ORM. The tool is designed for rapid application development in Java, offering faster development for Java Web and Spring Cloud microservices. The Enterprise Edition provides additional features like financial computing support, automatic value insertion, desensitization, dictionary value conversion, multi-tenancy, and more.
factorio-learning-environment
Factorio Learning Environment is an open source framework designed for developing and evaluating LLM agents in the game of Factorio. It provides two settings: Lab-play with structured tasks and Open-play for building large factories. Results show limitations in spatial reasoning and automation strategies. Agents interact with the environment through code synthesis, observation, action, and feedback. Tools are provided for game actions and state representation. Agents operate in episodes with observation, planning, and action execution. Tasks specify agent goals and are implemented in JSON files. The project structure includes directories for agents, environment, cluster, data, docs, eval, and more. A database is used for checkpointing agent steps. Benchmarks show performance metrics for different configurations.
pixeltable
Pixeltable is a Python library designed for ML Engineers and Data Scientists to focus on exploration, modeling, and app development without the need to handle data plumbing. It provides a declarative interface for working with text, images, embeddings, and video, enabling users to store, transform, index, and iterate on data within a single table interface. Pixeltable is persistent, acting as a database unlike in-memory Python libraries such as Pandas. It offers features like data storage and versioning, combined data and model lineage, indexing, orchestration of multimodal workloads, incremental updates, and automatic production-ready code generation. The tool emphasizes transparency, reproducibility, cost-saving through incremental data changes, and seamless integration with existing Python code and libraries.
funcchain
Funcchain is a Python library that allows you to easily write cognitive systems by leveraging Pydantic models as output schemas and LangChain in the backend. It provides a seamless integration of LLMs into your apps, utilizing OpenAI Functions or LlamaCpp grammars (json-schema-mode) for efficient structured output. Funcchain compiles the Funcchain syntax into LangChain runnables, enabling you to invoke, stream, or batch process your pipelines effortlessly.
bee
Bee is an easy and high efficiency ORM framework that simplifies database operations by providing a simple interface and eliminating the need to write separate DAO code. It supports various features such as automatic filtering of properties, partial field queries, native statement pagination, JSON format results, sharding, multiple database support, and more. Bee also offers powerful functionalities like dynamic query conditions, transactions, complex queries, MongoDB ORM, cache management, and additional tools for generating distributed primary keys, reading Excel files, and more. The newest versions introduce enhancements like placeholder precompilation, default date sharding, ElasticSearch ORM support, and improved query capabilities.
mcphub.nvim
MCPHub.nvim is a powerful Neovim plugin that integrates MCP (Model Context Protocol) servers into your workflow. It offers a centralized config file for managing servers and tools, with an intuitive UI for testing resources. Ideal for LLM integration, it provides programmatic API access and interactive testing through the `:MCPHub` command.
req_llm
ReqLLM is a Req-based library for LLM interactions, offering a unified interface to AI providers through a plugin-based architecture. It brings composability and middleware advantages to LLM interactions, with features like auto-synced providers/models, typed data structures, ergonomic helpers, streaming capabilities, usage & cost extraction, and a plugin-based provider system. Users can easily generate text, structured data, embeddings, and track usage costs. The tool supports various AI providers like Anthropic, OpenAI, Groq, Google, and xAI, and allows for easy addition of new providers. ReqLLM also provides API key management, detailed documentation, and a roadmap for future enhancements.
Acontext
Acontext is a context data platform designed for production AI agents, offering unified storage, built-in context management, and observability features. It helps agents scale from local demos to production without the need to rebuild context infrastructure. The platform provides solutions for challenges like scattered context data, long-running agents requiring context management, and tracking states from multi-modal agents. Acontext offers core features such as context storage, session management, disk storage, agent skills management, and sandbox for code execution and analysis. Users can connect to Acontext, install SDKs, initialize clients, store and retrieve messages, perform context engineering, and utilize agent storage tools. The platform also supports building agents using end-to-end scripts in Python and Typescript, with various templates available. Acontext's architecture includes client layer, backend with API and core components, infrastructure with PostgreSQL, S3, Redis, and RabbitMQ, and a web dashboard. Join the Acontext community on Discord and follow updates on GitHub.
open-responses
OpenResponses API provides enterprise-grade AI capabilities through a powerful API, simplifying development and deployment while ensuring complete data control. It offers automated tracing, integrated RAG for contextual information retrieval, pre-built tool integrations, self-hosted architecture, and an OpenAI-compatible interface. The toolkit addresses development challenges like feature gaps and integration complexity, as well as operational concerns such as data privacy and operational control. Engineering teams can benefit from improved productivity, production readiness, compliance confidence, and simplified architecture by choosing OpenResponses.
ai-microcore
MicroCore is a collection of python adapters for Large Language Models and Vector Databases / Semantic Search APIs. It allows convenient communication with these services, easy switching between models & services, and separation of business logic from implementation details. Users can keep applications simple and try various models & services without changing application code. MicroCore connects MCP tools to language models easily, supports text completion and chat completion models, and provides features for configuring, installing vendor-specific packages, and using vector databases.
For similar tasks
CrackSQL
CrackSQL is a powerful SQL dialect translation tool that integrates rule-based strategies with large language models (LLMs) for high accuracy. It enables seamless conversion between dialects (e.g., PostgreSQL โ MySQL) with flexible access through Python API, command line, and web interface. The tool supports extensive dialect compatibility, precision & advanced processing, and versatile access & integration. It offers three modes for dialect translation and demonstrates high translation accuracy over collected benchmarks. Users can deploy CrackSQL using PyPI package installation or source code installation methods. The tool can be extended to support additional syntax, new dialects, and improve translation efficiency. The project is actively maintained and welcomes contributions from the community.
For similar jobs
lollms-webui
LoLLMs WebUI (Lord of Large Language Multimodal Systems: One tool to rule them all) is a user-friendly interface to access and utilize various LLM (Large Language Models) and other AI models for a wide range of tasks. With over 500 AI expert conditionings across diverse domains and more than 2500 fine tuned models over multiple domains, LoLLMs WebUI provides an immediate resource for any problem, from car repair to coding assistance, legal matters, medical diagnosis, entertainment, and more. The easy-to-use UI with light and dark mode options, integration with GitHub repository, support for different personalities, and features like thumb up/down rating, copy, edit, and remove messages, local database storage, search, export, and delete multiple discussions, make LoLLMs WebUI a powerful and versatile tool.
Azure-Analytics-and-AI-Engagement
The Azure-Analytics-and-AI-Engagement repository provides packaged Industry Scenario DREAM Demos with ARM templates (Containing a demo web application, Power BI reports, Synapse resources, AML Notebooks etc.) that can be deployed in a customerโs subscription using the CAPE tool within a matter of few hours. Partners can also deploy DREAM Demos in their own subscriptions using DPoC.
minio
MinIO is a High Performance Object Storage released under GNU Affero General Public License v3.0. It is API compatible with Amazon S3 cloud storage service. Use MinIO to build high performance infrastructure for machine learning, analytics and application data workloads.
mage-ai
Mage is an open-source data pipeline tool for transforming and integrating data. It offers an easy developer experience, engineering best practices built-in, and data as a first-class citizen. Mage makes it easy to build, preview, and launch data pipelines, and provides observability and scaling capabilities. It supports data integrations, streaming pipelines, and dbt integration.
AiTreasureBox
AiTreasureBox is a versatile AI tool that provides a collection of pre-trained models and algorithms for various machine learning tasks. It simplifies the process of implementing AI solutions by offering ready-to-use components that can be easily integrated into projects. With AiTreasureBox, users can quickly prototype and deploy AI applications without the need for extensive knowledge in machine learning or deep learning. The tool covers a wide range of tasks such as image classification, text generation, sentiment analysis, object detection, and more. It is designed to be user-friendly and accessible to both beginners and experienced developers, making AI development more efficient and accessible to a wider audience.
tidb
TiDB is an open-source distributed SQL database that supports Hybrid Transactional and Analytical Processing (HTAP) workloads. It is MySQL compatible and features horizontal scalability, strong consistency, and high availability.
airbyte
Airbyte is an open-source data integration platform that makes it easy to move data from any source to any destination. With Airbyte, you can build and manage data pipelines without writing any code. Airbyte provides a library of pre-built connectors that make it easy to connect to popular data sources and destinations. You can also create your own connectors using Airbyte's no-code Connector Builder or low-code CDK. Airbyte is used by data engineers and analysts at companies of all sizes to build and manage their data pipelines.
labelbox-python
Labelbox is a data-centric AI platform for enterprises to develop, optimize, and use AI to solve problems and power new products and services. Enterprises use Labelbox to curate data, generate high-quality human feedback data for computer vision and LLMs, evaluate model performance, and automate tasks by combining AI and human-centric workflows. The academic & research community uses Labelbox for cutting-edge AI research.




