Best AI tools for< Compiler Scientist >
Infographic
20 - AI tool Sites

Replit
Replit is a software creation platform that provides an integrated development environment (IDE), artificial intelligence (AI) assistance, and deployment services. It allows users to build, test, and deploy software projects directly from their browser, without the need for local setup or configuration. Replit offers real-time collaboration, code generation, debugging, and autocompletion features powered by AI. It supports multiple programming languages and frameworks, making it suitable for a wide range of development projects.

Coddy
Coddy is an AI-powered coding assistant that helps developers write better code faster. It provides real-time feedback, code completion, and error detection, making it the perfect tool for both beginners and experienced developers. Coddy also integrates with popular development tools like Visual Studio Code and GitHub, making it easy to use in your existing workflow.

Anycores
Anycores is an AI tool designed to optimize the performance of deep neural networks and reduce the cost of running AI models in the cloud. It offers a platform that provides automated solutions for tuning and inference consultation, optimized networks zoo, and platform for reducing AI model cost. Anycores focuses on faster execution, reducing inference time over 10x times, and footprint reduction during model deployment. It is device agnostic, supporting Nvidia, AMD GPUs, Intel, ARM, AMD CPUs, servers, and edge devices. The tool aims to provide highly optimized, low footprint networks tailored to specific deployment scenarios.

ONNX
ONNX is an open standard for machine learning interoperability, providing a common format to represent machine learning models. It defines a set of operators and a file format for AI developers to use models across various frameworks, tools, runtimes, and compilers. ONNX promotes interoperability, hardware access, and community engagement.

MacroMicro
MacroMicro is an AI analytics platform that combines technology and research expertise to empower users with valuable insights into global market trends. With over 0k registered users and 0M+ monthly website traffic, MacroMicro offers real-time charts, cycle analysis, and data-driven insights to optimize investment strategies. The platform compiles the MM Global Recession Probability, utilizes OpenAI's Embedding technology, and provides exclusive reports and analysis on key market events. Users can access dynamic and automatically-updated charts, a powerful toolbox for analysis, and engage with a vibrant community of macroeconomic professionals.

Replexica
Replexica is an AI-powered i18n compiler for React that is JSON-free and LLM-backed. It is designed for shipping multi-language frontends fast.

BugFree.ai
BugFree.ai is an AI-powered platform designed to help users practice system design and behavior interviews, similar to Leetcode. The platform offers a range of features to assist users in preparing for technical interviews, including mock interviews, real-time feedback, and personalized study plans. With BugFree.ai, users can improve their problem-solving skills and gain confidence in tackling complex interview questions.

illbeback.ai
illbeback.ai is the #1 site for AI jobs around the world. It provides a platform for both job seekers and employers to connect in the field of Artificial Intelligence. The website features a wide range of AI job listings from top companies, offering opportunities for professionals in the AI industry to advance their careers. With a user-friendly interface, illbeback.ai simplifies the job search process for AI enthusiasts and provides valuable resources for companies looking to hire AI talent.

PseudoEditor
PseudoEditor is a free, fast, and online pseudocode IDE/editor that aims to simplify the process of writing pseudocode. It offers dynamic syntax highlighting, code saving, error highlighting, and a pseudocode compiler feature. The platform allows users to write and debug pseudocode efficiently, with the goal of creating algorithms quickly and easily. PseudoEditor is the first and only pseudocode online editor/IDE available for free, providing a smoother writing environment and faster coding experience.

Narada
Narada is an AI application designed for busy professionals to streamline their work processes. It leverages cutting-edge AI technology to automate tasks, connect favorite apps, and enhance productivity through intelligent automation. Narada's LLM Compiler routes text and voice commands to the right tools in real time, offering seamless app integration and time-saving features.

GetSelected.ai
GetSelected.ai is a personal AI-powered interviewer platform that helps users enhance their interview skills through AI technology. The platform offers features such as mock interviews, personalized feedback, job position customization, AI-driven quizzes, resume optimization, and code compiler for IT roles. Users can practice interview scenarios, improve communication skills, and prepare for recruitment processes with the help of AI tools. GetSelected.ai aims to provide a comprehensive and customizable experience to meet unique career goals and stand out in the competitive job market.

Roadmapped.ai
Roadmapped.ai is an AI-powered platform designed to help users learn various topics efficiently and quickly. By providing a structured roadmap generated in seconds, the platform eliminates the need to navigate through scattered online resources aimlessly. Users can input a topic they want to learn, and the AI will generate a personalized roadmap with curated resources. The platform also offers features like AI-powered YouTube search, saving roadmaps, priority support, and access to a private Discord community.

SoraPrompt
SoraPrompt is an AI model that can create realistic and imaginative scenes from text instructions. It is the latest text-to-video technology from the OpenAI development team. Users can compile text prompts to generate video query summaries for efficient content analysis. SoraPrompt also allows users to share their interests and ideas with others.

Twig AI
Twig AI is an AI tool designed for Customer Experience, offering an AI assistant that resolves customer issues instantly, supporting both users and support agents 24/7. It provides features like converting user requests into API calls, instant responses for user questions, and factual answers cited with trustworthy sources. Twig simplifies data retrieval from external sources, offers personalization options, and includes a built-in knowledge base. The tool aims to drive agent productivity, provide insights to monitor customer experience, and offers various application interfaces for different user roles.

Rargus
Rargus is a generative AI tool that specializes in turning customer feedback into actionable insights for businesses. By collecting feedback from various channels and utilizing custom AI analysis, Rargus helps businesses understand customer needs and improve product development. The tool enables users to compile and analyze feedback efficiently, leading to data-driven decision-making and successful product launches. Rargus also offers solutions for consumer insights, product management, and product marketing, helping businesses enhance customer satisfaction and drive growth.

AI Document Creator
AI Document Creator is an innovative tool that leverages artificial intelligence to assist users in generating various types of documents efficiently. The application utilizes advanced algorithms to analyze input data and create well-structured documents tailored to the user's needs. With AI Document Creator, users can save time and effort in document creation, ensuring accuracy and consistency in their outputs. The tool is user-friendly and accessible, making it suitable for individuals and businesses seeking to streamline their document creation process.

Dokkio
Dokkio is an AI-powered platform that helps users find, organize, and understand all of their online files. By leveraging AI technology, Dokkio enables users to work with their cloud files efficiently and collaboratively. The platform offers tools for managing multiple activities, finding documents and files, compiling research materials, and organizing a content library. Dokkio aims to streamline the process of accessing and utilizing online content, regardless of where it is stored.

Smarty
Smarty is an AI-powered productivity tool that acts as an execution engine for businesses. It combines AI technology with human experts to help users manage tasks, events, scheduling, and productivity. Smarty offers features like natural-language-based console, unified view of tasks and calendar, automatic prioritization, brain dumping, automation shortcuts, and personalized interactions. It helps users work smarter, stay organized, and save time by streamlining workflows and enhancing productivity. Smarty is designed to be a versatile task organizer app suitable for professionals looking to optimize daily planning and task management.

Notis
Notis is an AI voice-powered copilot designed for Notion users. It allows users to break free from their desks by turning their phones into a Notion copilot. Users can capture thoughts, organize them, and get answers from their workspace using voice commands. Notis offers features like transcribing voice notes, managing tasks, writing meeting minutes, content creation for social media, managing customer relationships, tracking expenses, drafting documents, compiling knowledge bases, and more. It integrates seamlessly with Notion, providing a second brain system to manage both professional and personal life efficiently.

PLEASEDONTCODE
PLEASEDONTCODE is an AI Code Generator specifically designed for Arduino and ESP32 embedded systems. It simplifies the coding process for users, helping them overcome common obstacles in programming such as compilation errors, syntax confusion, and logic issues. By automating code generation, the tool ensures error-free, syntactically correct, and logically sound code with clear comments for easy understanding. It aims to streamline project execution, increase productivity, and accelerate the development of prototype projects.
20 - Open Source Tools

llm-course
The LLM course is divided into three parts: 1. 🧩 **LLM Fundamentals** covers essential knowledge about mathematics, Python, and neural networks. 2. 🧑🔬 **The LLM Scientist** focuses on building the best possible LLMs using the latest techniques. 3. 👷 **The LLM Engineer** focuses on creating LLM-based applications and deploying them. For an interactive version of this course, I created two **LLM assistants** that will answer questions and test your knowledge in a personalized way: * 🤗 **HuggingChat Assistant**: Free version using Mixtral-8x7B. * 🤖 **ChatGPT Assistant**: Requires a premium account. ## 📝 Notebooks A list of notebooks and articles related to large language models. ### Tools | Notebook | Description | Notebook | |----------|-------------|----------| | 🧐 LLM AutoEval | Automatically evaluate your LLMs using RunPod |  | | 🥱 LazyMergekit | Easily merge models using MergeKit in one click. |  | | 🦎 LazyAxolotl | Fine-tune models in the cloud using Axolotl in one click. |  | | ⚡ AutoQuant | Quantize LLMs in GGUF, GPTQ, EXL2, AWQ, and HQQ formats in one click. |  | | 🌳 Model Family Tree | Visualize the family tree of merged models. |  | | 🚀 ZeroSpace | Automatically create a Gradio chat interface using a free ZeroGPU. |  |

cia
CIA is a powerful open-source tool designed for data analysis and visualization. It provides a user-friendly interface for processing large datasets and generating insightful reports. With CIA, users can easily explore data, perform statistical analysis, and create interactive visualizations to communicate findings effectively. Whether you are a data scientist, analyst, or researcher, CIA offers a comprehensive set of features to streamline your data analysis workflow and uncover valuable insights.

byteir
The ByteIR Project is a ByteDance model compilation solution. ByteIR includes compiler, runtime, and frontends, and provides an end-to-end model compilation solution. Although all ByteIR components (compiler/runtime/frontends) are together to provide an end-to-end solution, and all under the same umbrella of this repository, each component technically can perform independently. The name, ByteIR, comes from a legacy purpose internally. The ByteIR project is NOT an IR spec definition project. Instead, in most scenarios, ByteIR directly uses several upstream MLIR dialects and Google Mhlo. Most of ByteIR compiler passes are compatible with the selected upstream MLIR dialects and Google Mhlo.

maxtext
MaxText is a high-performance, highly scalable, open-source LLM written in pure Python/Jax and targeting Google Cloud TPUs and GPUs for training and inference. MaxText achieves high MFUs and scales from single host to very large clusters while staying simple and "optimization-free" thanks to the power of Jax and the XLA compiler. MaxText aims to be a launching off point for ambitious LLM projects both in research and production. We encourage users to start by experimenting with MaxText out of the box and then fork and modify MaxText to meet their needs.

Endia
Endia is a dynamic Array library for Scientific Computing, offering automatic differentiation of arbitrary order, complex number support, dual API with PyTorch-like imperative or JAX-like functional interface, and JIT Compilation for speeding up training and inference. It can handle complex valued functions, perform both forward and reverse-mode automatic differentiation, and has a builtin JIT compiler. Endia aims to advance AI & Scientific Computing by pushing boundaries with clear algorithms, providing high-performance open-source code that remains readable and pythonic, and prioritizing clarity and educational value over exhaustive features.

raft
RAFT (Reusable Accelerated Functions and Tools) is a C++ header-only template library with an optional shared library that contains fundamental widely-used algorithms and primitives for machine learning and information retrieval. The algorithms are CUDA-accelerated and form building blocks for more easily writing high performance applications.

nncase
nncase is a neural network compiler for AI accelerators that supports multiple inputs and outputs, static memory allocation, operators fusion and optimizations, float and quantized uint8 inference, post quantization from float model with calibration dataset, and flat model with zero copy loading. It can be installed via pip and supports TFLite, Caffe, and ONNX ops. Users can compile nncase from source using Ninja or make. The tool is suitable for tasks like image classification, object detection, image segmentation, pose estimation, and more.

AiTreasureBox
AiTreasureBox is a versatile AI tool that provides a collection of pre-trained models and algorithms for various machine learning tasks. It simplifies the process of implementing AI solutions by offering ready-to-use components that can be easily integrated into projects. With AiTreasureBox, users can quickly prototype and deploy AI applications without the need for extensive knowledge in machine learning or deep learning. The tool covers a wide range of tasks such as image classification, text generation, sentiment analysis, object detection, and more. It is designed to be user-friendly and accessible to both beginners and experienced developers, making AI development more efficient and accessible to a wider audience.

ByteMLPerf
ByteMLPerf is an AI Accelerator Benchmark that focuses on evaluating AI Accelerators from a practical production perspective, including the ease of use and versatility of software and hardware. Byte MLPerf has the following characteristics: - Models and runtime environments are more closely aligned with practical business use cases. - For ASIC hardware evaluation, besides evaluate performance and accuracy, it also measure metrics like compiler usability and coverage. - Performance and accuracy results obtained from testing on the open Model Zoo serve as reference metrics for evaluating ASIC hardware integration.

AITreasureBox
AITreasureBox is a comprehensive collection of AI tools and resources designed to simplify and accelerate the development of AI projects. It provides a wide range of pre-trained models, datasets, and utilities that can be easily integrated into various AI applications. With AITreasureBox, developers can quickly prototype, test, and deploy AI solutions without having to build everything from scratch. Whether you are working on computer vision, natural language processing, or reinforcement learning projects, AITreasureBox has something to offer for everyone. The repository is regularly updated with new tools and resources to keep up with the latest advancements in the field of artificial intelligence.

npcsh
`npcsh` is a python-based command-line tool designed to integrate Large Language Models (LLMs) and Agents into one's daily workflow by making them available and easily configurable through the command line shell. It leverages the power of LLMs to understand natural language commands and questions, execute tasks, answer queries, and provide relevant information from local files and the web. Users can also build their own tools and call them like macros from the shell. `npcsh` allows users to take advantage of agents (i.e. NPCs) through a managed system, tailoring NPCs to specific tasks and workflows. The tool is extensible with Python, providing useful functions for interacting with LLMs, including explicit coverage for popular providers like ollama, anthropic, openai, gemini, deepseek, and openai-like providers. Users can set up a flask server to expose their NPC team for use as a backend service, run SQL models defined in their project, execute assembly lines, and verify the integrity of their NPC team's interrelations. Users can execute bash commands directly, use favorite command-line tools like VIM, Emacs, ipython, sqlite3, git, pipe the output of these commands to LLMs, or pass LLM results to bash commands.

cl-waffe2
cl-waffe2 is an experimental deep learning framework in Common Lisp, providing fast, systematic, and customizable matrix operations, reverse mode tape-based Automatic Differentiation, and neural network model building and training features accelerated by a JIT Compiler. It offers abstraction layers, extensibility, inlining, graph-level optimization, visualization, debugging, systematic nodes, and symbolic differentiation. Users can easily write extensions and optimize their networks without overheads. The framework is designed to eliminate barriers between users and developers, allowing for easy customization and extension.

Bodo
Bodo is a high-performance Python compute engine designed for large-scale data processing and AI workloads. It utilizes an auto-parallelizing just-in-time compiler to optimize Python programs, making them 20x to 240x faster compared to alternatives. Bodo seamlessly integrates with native Python APIs like Pandas and NumPy, eliminates runtime overheads using MPI for distributed execution, and provides exceptional performance and scalability for data workloads. It is easy to use, interoperable with the Python ecosystem, and integrates with modern data platforms like Apache Iceberg and Snowflake. Bodo focuses on data-intensive and computationally heavy workloads in data engineering, data science, and AI/ML, offering automatic optimization and parallelization, linear scalability, advanced I/O support, and a high-performance SQL engine.

mlir-aie
This repository contains an MLIR-based toolchain for AI Engine-enabled devices, such as AMD Ryzen™ AI and Versal™. This repository can be used to generate low-level configurations for the AI Engine portion of these devices. AI Engines are organized as a spatial array of tiles, where each tile contains AI Engine cores and/or memories. The spatial array is connected by stream switches that can be configured to route data between AI Engine tiles scheduled by their programmable Data Movement Accelerators (DMAs). This repository contains MLIR representations, with multiple levels of abstraction, to target AI Engine devices. This enables compilers and developers to program AI Engine cores, as well as describe data movements and array connectivity. A Python API is made available as a convenient interface for generating MLIR design descriptions. Backend code generation is also included, targeting the aie-rt library. This toolchain uses the AI Engine compiler tool which is part of the AMD Vitis™ software installation: these tools require a free license for use from the Product Licensing Site.

mlc-llm
MLC LLM is a high-performance universal deployment solution that allows native deployment of any large language models with native APIs with compiler acceleration. It supports a wide range of model architectures and variants, including Llama, GPT-NeoX, GPT-J, RWKV, MiniGPT, GPTBigCode, ChatGLM, StableLM, Mistral, and Phi. MLC LLM provides multiple sets of APIs across platforms and environments, including Python API, OpenAI-compatible Rest-API, C++ API, JavaScript API and Web LLM, Swift API for iOS App, and Java API and Android App.

husky
Husky is a research-focused programming language designed for next-generation computing. It aims to provide a powerful and ergonomic development experience for various tasks, including system level programming, web/native frontend development, parser/compiler tasks, game development, formal verification, machine learning, and more. With a strong type system and support for human-in-the-loop programming, Husky enables users to tackle complex tasks such as explainable image classification, natural language processing, and reinforcement learning. The language prioritizes debugging, visualization, and human-computer interaction, offering agile compilation and evaluation, multiparadigm support, and a commitment to a good ecosystem.

llama3.java
Llama3.java is a practical Llama 3 inference tool implemented in a single Java file. It serves as the successor of llama2.java and is designed for testing and tuning compiler optimizations and features on the JVM, especially for the Graal compiler. The tool features a GGUF format parser, Llama 3 tokenizer, Grouped-Query Attention inference, support for Q8_0 and Q4_0 quantizations, fast matrix-vector multiplication routines using Java's Vector API, and a simple CLI with 'chat' and 'instruct' modes. Users can download quantized .gguf files from huggingface.co for model usage and can also manually quantize to pure 'Q4_0'. The tool requires Java 21+ and supports running from source or building a JAR file for execution. Performance benchmarks show varying tokens/s rates for different models and implementations on different hardware setups.

MPLSandbox
MPLSandbox is an out-of-the-box multi-programming language sandbox designed to provide unified and comprehensive feedback from compiler and analysis tools for LLMs. It simplifies code analysis for researchers and can be seamlessly integrated into LLM training and application processes to enhance performance in a range of code-related tasks. The sandbox environment ensures safe code execution, the code analysis module offers comprehensive analysis reports, and the information integration module combines compilation feedback and analysis results for complex code-related tasks.

LLM-Agents-Papers
A repository that lists papers related to Large Language Model (LLM) based agents. The repository covers various topics including survey, planning, feedback & reflection, memory mechanism, role playing, game playing, tool usage & human-agent interaction, benchmark & evaluation, environment & platform, agent framework, multi-agent system, and agent fine-tuning. It provides a comprehensive collection of research papers on LLM-based agents, exploring different aspects of AI agent architectures and applications.

LLM-TPU
LLM-TPU project aims to deploy various open-source generative AI models on the BM1684X chip, with a focus on LLM. Models are converted to bmodel using TPU-MLIR compiler and deployed to PCIe or SoC environments using C++ code. The project has deployed various open-source models such as Baichuan2-7B, ChatGLM3-6B, CodeFuse-7B, DeepSeek-6.7B, Falcon-40B, Phi-3-mini-4k, Qwen-7B, Qwen-14B, Qwen-72B, Qwen1.5-0.5B, Qwen1.5-1.8B, Llama2-7B, Llama2-13B, LWM-Text-Chat, Mistral-7B-Instruct, Stable Diffusion, Stable Diffusion XL, WizardCoder-15B, Yi-6B-chat, Yi-34B-chat. Detailed model deployment information can be found in the 'models' subdirectory of the project. For demonstrations, users can follow the 'Quick Start' section. For inquiries about the chip, users can contact SOPHGO via the official website.
20 - OpenAI Gpts

BioinformaticsManual
Compile instructions from the web and github for bioinformatics applications. Receive line-by-line instructions and commands to get started
Melange Mentor
I'm a tutor for JavaScript and Melange, a compiler for OCaml that targets JavaScript.

Lead Scout
I compile and enrich precise company and professional profiles. Simply provide any name, email address, or company and I'll generate a complete profile.

A Remedy for Everything
Natural remedies for over 220 Ailments Compiled from 5 Years of Extensive Research.

Coloring Book Generator
Crafts full coloring books with a cover and compiled into a downloadable document.

FlutterCraft
FlutterCraft is an AI-powered assistant that streamlines Flutter app development. It interprets user-provided descriptions to generate and compile Flutter app code, providing ready-to-install APK and iOS files. Ideal for rapid prototyping, FlutterCraft makes app development accessible and efficient.

Daily Horoscope
Get your daily horoscope summary, categorized and compiled from various online sources. For entertainment purposes only.

Interview Pro
By combining the expertise of top career coaches with advanced AI, our GPT helps you excel in interviews across various job functions and levels. We've also compiled the most practical tips for you | We value your experience, please contact [email protected] if you need support ❤️!

Linux Kernel Expert
Formal and professional Linux Kernel Expert, adept in technical jargon.
ReScript
Write ReScript code. Trained with versions 10 & 11. Documentation github.com/guillempuche/gpt-rescript