
lerobotdepot
LeRobotDepot is a community-driven repository listing open-source hardware, components, and 3D-printable projects compatible with the LeRobot library. It helps users easily discover, build, and contribute to affordable, accessible robotics solutions powered by state-of-the-art AI.
Stars: 69

LeRobotDepot is a repository listing open-source hardware, components, and 3D-printable projects compatible with the LeRobot library. It helps users discover, build, and contribute to affordable robotics solutions powered by AI. The repository includes various robot arms, grippers, cameras, and accessories, along with detailed information on pricing, compatibility, and additional components. Users can find kits for assembling arms, wrist cameras, haptic sensors, and other modules. The repository also features mobile arms, bi-manual arms, humanoid robots, and task kits for specific tasks like push T task and handling a toaster. Additionally, there are resources for teleoperation, cameras, and common accessories like self-fusing silicone rubber for increasing grip friction.
README:
Welcome to LeRobotDepot. This repository is listing open-source hardware, components, and 3D-printable projects compatible with the LeRobot library. It helps users easily discover, build, and contribute to affordable, accessible robotics solutions powered by state-of-the-art AI.
Hardware in this family uses Feetech motors—specifically, the STS3215 series available in both 7.4V and 12V variants. These motors are popular for their balance between performance and cost:
- 7.4V Version: Typically offers a stall torque of approximately 16.5 kg·cm at 6V. This option is often sufficient for basic robotics applications.
- 12V Version: Delivers around 30 kg·cm of stall torque, providing increased power for more demanding tasks.
By standardizing on the STS3215, projects in the Feetech Family maintain similar power and control characteristics, ensuring compatibility across accessories and modules.
This 5 DOF arm is the recommended arm to get started with LeRobot—especially the 7.4V version.
Price | US | EU | RMB |
---|---|---|---|
Follower and Leader arms | $232 | €244 | ¥1343.16 |
One Arm | $123 | €128 | ¥682.23 |
For detailed information on the various accessories available for the SO-ARM100, including mounting options and additional components, please refer to the SO-ARM100 repository’s hardware documentation.
The SO-ARM100 supports multiple wrist camera options to suit a variety of applications. There are three officially supported options and one community-developed alternative:
Camera Name | Reference Link | Notes |
---|---|---|
Vinmooog Webcam | SO-ARM100 Instructions | |
32x32mm UVC Module | SO-ARM100 Instructions | |
Arducam 5MP Wide Angle | Le Kiwi STL File | It can also be used with 32x32mm UVC modules, but if you don't use a wide-angle camera, the gripper will not appear in the camera view. |
RealSense™ D405 | SO-ARM100 Instructions | |
RealSense™ D435 | STL File |
You can find kits for the SO100 arms here:
Both assembled and non-assembled kits are available, depending on the supplier.
This 5 DOF arm is similar to the SO-ARM100 but uses only the gripper as a 3D printed part. It is recommended to build or purchase the SO100 arm instead. While the Moss v1 robot is still supported, it will be deprecated. Additionally, 3D-printed parts for the SO-ARM100 are now available for purchase if you don't have a printer.
Price | US | EU | RMB |
---|---|---|---|
Follower and Leader arms | $288 | €274 | ¥1631.46 |
One Arm | $159 | €153 | ¥868.13 |
See SO-ARM100 Accessories for compatible components and mounts.
This is a 6 DOF arm, based on the SO-ARM100 leader and follower arms, with an additional joint enabled by one more STS3215 servo motor.
The price is roughly equivalent to the SO-ARM100, plus the cost of one extra Feetech servo motor—either 7.4V or 12V, depending on your chosen configuration.
For wrist cameras, haptic sensors, and other modules, see SO-ARM100 Accessories for compatible components.
This is a 6 DOF arm, developed by the community around the SimpleAutomation repository. It is a refined version of the SO-ARM100, offering enhanced movement precision and a gripper better optimized for handling small objects.
Price | US |
---|---|
Follower and Leader arms | ± $450 |
AB-SO-BOT is built using a combination of 3D-printed parts and standard 4040-T-slot aluminium extrusion to create a flexible and modular body for the SO-ARM100 robotic arm. This modularity allows for easy customization, expansion, and adaptation for different robotic applications.
Mobile version of the SO-ARM100.
Price | US | EU |
---|---|---|
12V | $488.21 | €542.56 |
5V | $524.95 | €525.9 |
Base only (5V) | $251.95 | €306.9 |
Base only (12V) | $257.43 | €305 |
Base only wired | $174 | €233.3 |
Mobile version of the SO-ARM100 with two arms.
Price | US | EU | RMB |
---|---|---|---|
Total | ~ $300 | ~ €300 | ~ ¥2000 |
Miniature version of the BDX Droid by Disney.
~€410
A project for a full body robot—currently featuring the torso and arms.
Parallel-finger gripper compatible with SO-ARM100.
~€25
Precise gripper compatible with SO-ARM100.
It provides an additional axis to the SO-ARM100 robot arm.
Hardware in this family uses Dynamixel servo motors, which are considered more of an industry standard than Feetech motors.
The Koch-v1-1 is a 5 DOF robotic arm. If you want to familiarise yourself with more industry standard Dynamixel servo motors, this project could be a good starting point. Compared to the SO-ARM100, you will have less torque and a more limited range of movement from its base.
Price | US | EU | UK | RMB | JPY |
---|---|---|---|---|---|
Follower and Leader arms | $477 | €673 | £507 | ¥3947 | ¥22439 |
Leader Arm | $278 | €368 | £285 | ¥2251 | ¥15446 |
Follower Arm | $199 | €305 | £222 | ¥1696 | ¥6993 |
The Koch-v1-1 supports 2 wrist camera options:
Camera Name | Reference Link | Notes |
---|---|---|
SVPRO 1080P | Discord Message with STL file | |
N/A | WOWROBO Gripper-Camera Kit | Kit including the gripper, the camera mount and the camera |
- Robotic arm inspired by Kochv-1-1: WOWROBO Twinarm
ALOHA 2 is a bimanual teleoperation system that uses two types of arms—a pair of smaller, ergonomically designed leader arms and two robust follower arms—to support coordinated dual-arm manipulation. Each arm offers 6 degrees of freedom (6 DOF), which provides an extensive range of motion for accessing various positions and orientations.
The system is designed for research in fine-grained bimanual manipulation. Its construction includes enhanced gripper mechanisms, a passive gravity compensation system, and a rigid frame that supports precise and repeatable operations for complex tasks. These advanced features and components are reflected in its higher cost compared to more basic robotic arm solutions.
~$27,000
- "T" for push T task.
- A "toaster" with 2 pieces of "toast".
- A paper towel base & rod + paper towel roll.
- Cube.
- Ring.
To increase friction on gripper.
Hardware that attaches to the back of your hand and fingertips that tracks 16 degrees of freedom. Compatible with HOPEJr hands.
Name | Price Range | Link | Resolution | FPS | Wide Angle | Microphone |
---|---|---|---|---|---|---|
Innomaker 1080P USB2.0 | ± $18, €16 | Innomaker Link | 1920×1080 | 30 | Fov(D) = 130° Fov(H) = 103° |
No |
Innomaker 720p USB2.0 | ± $10, €14 | Innomaker Link | 1280×720 | 30 | FOV (D) = 120° FOV (H) = 102° |
No |
Innomaker OV9281 USB 2.0 | ± $36, €42 | Innomaker Link | 1280×800 | 120 | FOV Up to 148° | No |
Vinmooog Webcam | ± $14, €12 | Amazon Link | 1920×1080 | N/A | N/A | Yes |
- https://www.amazon.co.uk/ELP-Conferencing-Fisheye-0-01Lux-Computer/dp/B08Y1KY5T9?th=1
- https://www.amazon.com/dp/B07CSJN2KH
Interested in contributing? Please take a moment to review our CONTRIBUTING.md for guidelines on how to get started.
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for lerobotdepot
Similar Open Source Tools

lerobotdepot
LeRobotDepot is a repository listing open-source hardware, components, and 3D-printable projects compatible with the LeRobot library. It helps users discover, build, and contribute to affordable robotics solutions powered by AI. The repository includes various robot arms, grippers, cameras, and accessories, along with detailed information on pricing, compatibility, and additional components. Users can find kits for assembling arms, wrist cameras, haptic sensors, and other modules. The repository also features mobile arms, bi-manual arms, humanoid robots, and task kits for specific tasks like push T task and handling a toaster. Additionally, there are resources for teleoperation, cameras, and common accessories like self-fusing silicone rubber for increasing grip friction.

SimAI
SimAI is the industry's first full-stack, high-precision simulator for AI large-scale training. It provides detailed modeling and simulation of the entire LLM training process, encompassing framework, collective communication, network layers, and more. This comprehensive approach offers end-to-end performance data, enabling researchers to analyze training process details, evaluate time consumption of AI tasks under specific conditions, and assess performance gains from various algorithmic optimizations.

TokenPacker
TokenPacker is a novel visual projector that compresses visual tokens by 75%∼89% with high efficiency. It adopts a 'coarse-to-fine' scheme to generate condensed visual tokens, achieving comparable or better performance across diverse benchmarks. The tool includes TokenPacker for general use and TokenPacker-HD for high-resolution image understanding. It provides training scripts, checkpoints, and supports various compression ratios and patch numbers.

HuatuoGPT-II
HuatuoGPT2 is an innovative domain-adapted medical large language model that excels in medical knowledge and dialogue proficiency. It showcases state-of-the-art performance in various medical benchmarks, surpassing GPT-4 in expert evaluations and fresh medical licensing exams. The open-source release includes HuatuoGPT2 models in 7B, 13B, and 34B versions, training code for one-stage adaptation, partial pre-training and fine-tuning instructions, and evaluation methods for medical response capabilities and professional pharmacist exams. The tool aims to enhance LLM capabilities in the Chinese medical field through open-source principles.

IDvs.MoRec
This repository contains the source code for the SIGIR 2023 paper 'Where to Go Next for Recommender Systems? ID- vs. Modality-based Recommender Models Revisited'. It provides resources for evaluating foundation, transferable, multi-modal, and LLM recommendation models, along with datasets, pre-trained models, and training strategies for IDRec and MoRec using in-batch debiased cross-entropy loss. The repository also offers large-scale datasets, code for SASRec with in-batch debias cross-entropy loss, and information on joining the lab for research opportunities.

camel
CAMEL is an open-source library designed for the study of autonomous and communicative agents. We believe that studying these agents on a large scale offers valuable insights into their behaviors, capabilities, and potential risks. To facilitate research in this field, we implement and support various types of agents, tasks, prompts, models, and simulated environments.

ColossalAI
Colossal-AI is a deep learning system for large-scale parallel training. It provides a unified interface to scale sequential code of model training to distributed environments. Colossal-AI supports parallel training methods such as data, pipeline, tensor, and sequence parallelism and is integrated with heterogeneous training and zero redundancy optimizer.

LLaVA-MORE
LLaVA-MORE is a new family of Multimodal Language Models (MLLMs) that integrates recent language models with diverse visual backbones. The repository provides a unified training protocol for fair comparisons across all architectures and releases training code and scripts for distributed training. It aims to enhance Multimodal LLM performance and offers various models for different tasks. Users can explore different visual backbones like SigLIP and methods for managing image resolutions (S2) to improve the connection between images and language. The repository is a starting point for expanding the study of Multimodal LLMs and enhancing new features in the field.

unoplat-code-confluence
Unoplat-CodeConfluence is a universal code context engine that aims to extract, understand, and provide precise code context across repositories tied through domains. It combines deterministic code grammar with state-of-the-art LLM pipelines to achieve human-like understanding of codebases in minutes. The tool offers smart summarization, graph-based embedding, enhanced onboarding, graph-based intelligence, deep dependency insights, and seamless integration with existing development tools and workflows. It provides a precise context API for knowledge engine and AI coding assistants, enabling reliable code understanding through bottom-up code summarization, graph-based querying, and deep package and dependency analysis.

awesome-MLSecOps
Awesome MLSecOps is a curated list of open-source tools, resources, and tutorials for MLSecOps (Machine Learning Security Operations). It includes a wide range of security tools and libraries for protecting machine learning models against adversarial attacks, as well as resources for AI security, data anonymization, model security, and more. The repository aims to provide a comprehensive collection of tools and information to help users secure their machine learning systems and infrastructure.

Q-Bench
Q-Bench is a benchmark for general-purpose foundation models on low-level vision, focusing on multi-modality LLMs performance. It includes three realms for low-level vision: perception, description, and assessment. The benchmark datasets LLVisionQA and LLDescribe are collected for perception and description tasks, with open submission-based evaluation. An abstract evaluation code is provided for assessment using public datasets. The tool can be used with the datasets API for single images and image pairs, allowing for automatic download and usage. Various tasks and evaluations are available for testing MLLMs on low-level vision tasks.

Xwin-LM
Xwin-LM is a powerful and stable open-source tool for aligning large language models, offering various alignment technologies like supervised fine-tuning, reward models, reject sampling, and reinforcement learning from human feedback. It has achieved top rankings in benchmarks like AlpacaEval and surpassed GPT-4. The tool is continuously updated with new models and features.

LlamaV-o1
LlamaV-o1 is a Large Multimodal Model designed for spontaneous reasoning tasks. It outperforms various existing models on multimodal reasoning benchmarks. The project includes a Step-by-Step Visual Reasoning Benchmark, a novel evaluation metric, and a combined Multi-Step Curriculum Learning and Beam Search Approach. The model achieves superior performance in complex multi-step visual reasoning tasks in terms of accuracy and efficiency.

BitBLAS
BitBLAS is a library for mixed-precision BLAS operations on GPUs, for example, the $W_{wdtype}A_{adtype}$ mixed-precision matrix multiplication where $C_{cdtype}[M, N] = A_{adtype}[M, K] \times W_{wdtype}[N, K]$. BitBLAS aims to support efficient mixed-precision DNN model deployment, especially the $W_{wdtype}A_{adtype}$ quantization in large language models (LLMs), for example, the $W_{UINT4}A_{FP16}$ in GPTQ, the $W_{INT2}A_{FP16}$ in BitDistiller, the $W_{INT2}A_{INT8}$ in BitNet-b1.58. BitBLAS is based on techniques from our accepted submission at OSDI'24.

pr-agent
PR-Agent is a tool that helps to efficiently review and handle pull requests by providing AI feedbacks and suggestions. It supports various commands such as generating PR descriptions, providing code suggestions, answering questions about the PR, and updating the CHANGELOG.md file. PR-Agent can be used via CLI, GitHub Action, GitHub App, Docker, and supports multiple git providers and models. It emphasizes real-life practical usage, with each tool having a single GPT-4 call for quick and affordable responses. The PR Compression strategy enables effective handling of both short and long PRs, while the JSON prompting strategy allows for modular and customizable tools. PR-Agent Pro, the hosted version by CodiumAI, provides additional benefits such as full management, improved privacy, priority support, and extra features.

Awesome-AI-GPTs
Awesome AI GPTs is an open repository that collects resources and fun ways to use OpenAI GPTs. It includes databases, search tools, open-source projects, articles, attack and defense strategies, installation of custom plugins, knowledge bases, and community interactions related to GPTs. Users can find curated lists, leaked prompts, and various GPT applications in this repository. The project aims to empower users with AI capabilities and foster collaboration in the AI community.
For similar tasks

lerobotdepot
LeRobotDepot is a repository listing open-source hardware, components, and 3D-printable projects compatible with the LeRobot library. It helps users discover, build, and contribute to affordable robotics solutions powered by AI. The repository includes various robot arms, grippers, cameras, and accessories, along with detailed information on pricing, compatibility, and additional components. Users can find kits for assembling arms, wrist cameras, haptic sensors, and other modules. The repository also features mobile arms, bi-manual arms, humanoid robots, and task kits for specific tasks like push T task and handling a toaster. Additionally, there are resources for teleoperation, cameras, and common accessories like self-fusing silicone rubber for increasing grip friction.

sunone_aimbot
Sunone Aimbot is an AI-powered aim bot for first-person shooter games. It leverages YOLOv8 and YOLOv10 models, PyTorch, and various tools to automatically target and aim at enemies within the game. The AI model has been trained on more than 30,000 images from popular first-person shooter games like Warface, Destiny 2, Battlefield 2042, CS:GO, Fortnite, The Finals, CS2, and more. The aimbot can be configured through the `config.ini` file to adjust various settings related to object search, capture methods, aiming behavior, hotkeys, mouse settings, shooting options, Arduino integration, AI model parameters, overlay display, debug window, and more. Users are advised to follow specific recommendations to optimize performance and avoid potential issues while using the aimbot.

Everything-LLMs-And-Robotics
The Everything-LLMs-And-Robotics repository is the world's largest GitHub repository focusing on the intersection of Large Language Models (LLMs) and Robotics. It provides educational resources, research papers, project demos, and Twitter threads related to LLMs, Robotics, and their combination. The repository covers topics such as reasoning, planning, manipulation, instructions and navigation, simulation frameworks, perception, and more, showcasing the latest advancements in the field.

spear
SPEAR is a Simulator for Photorealistic Embodied AI Research that addresses limitations in existing simulators by offering 300 unique virtual indoor environments with detailed geometry, photorealistic materials, and unique floor plans. It provides an OpenAI Gym interface for interaction via Python, released under an MIT License. The simulator was developed with support from the Intelligent Systems Lab at Intel and Kujiale.

anthrax-ai
AnthraxAI is a Vulkan-based game engine that allows users to create and develop 3D games. The engine provides features such as scene selection, camera movement, object manipulation, debugging tools, audio playback, and real-time shader code updates. Users can build and configure the project using CMake and compile shaders using the glslc compiler. The engine supports building on both Linux and Windows platforms, with specific dependencies for each. Visual Studio Code integration is available for building and debugging the project, with instructions provided in the readme for setting up the workspace and required extensions.
For similar jobs

lerobotdepot
LeRobotDepot is a repository listing open-source hardware, components, and 3D-printable projects compatible with the LeRobot library. It helps users discover, build, and contribute to affordable robotics solutions powered by AI. The repository includes various robot arms, grippers, cameras, and accessories, along with detailed information on pricing, compatibility, and additional components. Users can find kits for assembling arms, wrist cameras, haptic sensors, and other modules. The repository also features mobile arms, bi-manual arms, humanoid robots, and task kits for specific tasks like push T task and handling a toaster. Additionally, there are resources for teleoperation, cameras, and common accessories like self-fusing silicone rubber for increasing grip friction.

weave
Weave is a toolkit for developing Generative AI applications, built by Weights & Biases. With Weave, you can log and debug language model inputs, outputs, and traces; build rigorous, apples-to-apples evaluations for language model use cases; and organize all the information generated across the LLM workflow, from experimentation to evaluations to production. Weave aims to bring rigor, best-practices, and composability to the inherently experimental process of developing Generative AI software, without introducing cognitive overhead.

LLMStack
LLMStack is a no-code platform for building generative AI agents, workflows, and chatbots. It allows users to connect their own data, internal tools, and GPT-powered models without any coding experience. LLMStack can be deployed to the cloud or on-premise and can be accessed via HTTP API or triggered from Slack or Discord.

VisionCraft
The VisionCraft API is a free API for using over 100 different AI models. From images to sound.

kaito
Kaito is an operator that automates the AI/ML inference model deployment in a Kubernetes cluster. It manages large model files using container images, avoids tuning deployment parameters to fit GPU hardware by providing preset configurations, auto-provisions GPU nodes based on model requirements, and hosts large model images in the public Microsoft Container Registry (MCR) if the license allows. Using Kaito, the workflow of onboarding large AI inference models in Kubernetes is largely simplified.

PyRIT
PyRIT is an open access automation framework designed to empower security professionals and ML engineers to red team foundation models and their applications. It automates AI Red Teaming tasks to allow operators to focus on more complicated and time-consuming tasks and can also identify security harms such as misuse (e.g., malware generation, jailbreaking), and privacy harms (e.g., identity theft). The goal is to allow researchers to have a baseline of how well their model and entire inference pipeline is doing against different harm categories and to be able to compare that baseline to future iterations of their model. This allows them to have empirical data on how well their model is doing today, and detect any degradation of performance based on future improvements.

tabby
Tabby is a self-hosted AI coding assistant, offering an open-source and on-premises alternative to GitHub Copilot. It boasts several key features: * Self-contained, with no need for a DBMS or cloud service. * OpenAPI interface, easy to integrate with existing infrastructure (e.g Cloud IDE). * Supports consumer-grade GPUs.

spear
SPEAR (Simulator for Photorealistic Embodied AI Research) is a powerful tool for training embodied agents. It features 300 unique virtual indoor environments with 2,566 unique rooms and 17,234 unique objects that can be manipulated individually. Each environment is designed by a professional artist and features detailed geometry, photorealistic materials, and a unique floor plan and object layout. SPEAR is implemented as Unreal Engine assets and provides an OpenAI Gym interface for interacting with the environments via Python.