
awesome-and-novel-works-in-slam
collecting some new ideas for slam (Semantic, 3DGS, BEV, Nav, LLM, Multi-session) and will update this repo weekly, for both engineering and academic usings.
Stars: 53

This repository contains a curated list of cutting-edge works in Simultaneous Localization and Mapping (SLAM). It includes research papers, projects, and tools related to various aspects of SLAM, such as 3D reconstruction, semantic mapping, novel algorithms, large-scale mapping, and more. The repository aims to showcase the latest advancements in SLAM technology and provide resources for researchers and practitioners in the field.
README:
This repo contains a mostly cutting edge (new derives from old) list of awesome and novel works in slam
If you find this repository useful, please consider STARing this list. Feel free to share this list with others! More comments will be found below, yet just opinions of mine. Let's make slamer great again.
- 3DGSNeRF
- v2x
- Semantic
- novel
- Largemodel
- leaders
- journals
- conferences
- Datasets
- Competition
- Tools
- Sim
- BuySensors
- FindIntern
- WeMedia
- WhySDMap
- WhyTraversability
- WhyLongTerm
Follow [Jiheng Yang] for more information on NeRF/3DGS/EVER, He will update news weekly, best Chinese community for NeRF/3DGS, I will do supplement when I have time.
Follow [ai kwea] for more practical experience on NeRF, I will do supplement when I have time.
Follow [Chongjie Ye] for more practical experience on 3DGS, I will do supplement when I have time.
-
EVER: "Exact Volumetric Ellipsoid Rendering for Real-time View Synthesis", arxiv 2024. [Code]
-
STORM: "STORM: Spatio-Temporal Reconstruction Model for Large-Scale Outdoor Scenes", arxiv 2024. [Project]
-
NaVILA: "NaVILA: Legged Robot Vision-Language-Action Model for Navigation", arxiv 2024. [Paper]
-
BeNeRF: "BeNeRF: Neural Radiance Fields from a Single Blurry Image and Event Stream", arxiv 2024. [Paper]
-
SLGaussian: "SLGaussian: Fast Language Gaussian Splatting in Sparse Views", arxiv 2024. [Paper]
-
HUGSIM: "HUGSIM: A Real-Time, Photo-Realistic and Closed-Loop Simulator for Autonomous Driving", arxiv 2024. [Code]
-
GaussianRPG: "GaussianRPG: 3D Gaussian Rendering PlayGround", ECCV 2024. [Code]
-
HSFM: "Reconstructing People, Places, and Cameras", arxiv 2024. [Project]
-
Gaussian Splatting: "3D Gaussian Splatting for Real-Time Radiance Field Rendering", ACM Transactions on Graphics 2023. [Paper] [Code]
-
Neural-Sim: "Learning to Generate Training Data with NeRF", ECCV 2022. [Paper] [Code] [Webpage]
-
iNeRF: "Inverting Neural Radiance Fields for Pose Estimation", IROS, 2021. [Paper] [Code] [Website] [Dataset]
-
iMAP: "Implicit Mapping and Positioning in Real-Time", ICCV, 2021. [Paper] [Code]
-
SHINE-Mapping: "Large-Scale 3D Mapping Using Sparse Hierarchical Implicit Neural Representations", ICRA, 2023. [Paper] [Code]
-
H2-Mapping: "Real-time Dense Mapping Using Hierarchical Hybrid Representation", RA-L, 2023. [Paper] [Code]
-
LATITUDE: Robotic Global Localization with Truncated Dynamic Low-pass Filter in City-scale NeRF, ICRA, 2023. [Paper] [Code]
-
NeuSE: "Neural SE(3)-Equivariant Embedding for Consistent Spatial Understanding with Objects", arXiv. [Paper] [Code]
-
ObjectFusion: "Accurate object-level SLAM with neural object priors", Graphical Models, 2022. [Paper]
-
NDF_Change: "Robust Change Detection Based on Neural Descriptor Fields", IROS, 2022. [Paper]
-
LNDF: "Local Neural Descriptor Fields: Locally Conditioned Object Representations for Manipulation", ICRA, 2023. [Paper] [Webpage]
- NeRF-LOAM: Neural Implicit Representation for Large-Scale Incremental LiDAR Odometry and Mapping, arXiv. [Paper] [Code]
-
"Implicit Map Augmentation for Relocalization", ECCV Workshop, 2022. [Paper]
-
Co-SLAM: Joint Coordinate and Sparse Parametric Encodings for Neural Real-Time SLAM, CVPR, 2023. [Paper] [Website]
-
Neural Implicit Dense Semantic SLAM, arXiv, 2023. [Paper]
-
NeRF-Navigation: "Vision-Only Robot Navigation in a Neural Radiance World", ICRA, 2022. [Paper] [Code] [Website]
-
ESDF: "Sampling-free obstacle gradients and reactive planning in Neural Radiance Fields", arXiv. [Paper]
-
Normal-NeRF: "Normal-NeRF: Ambiguity-Robust Normal Estimation for Highly Reflective Scenes", arXiv 2025. [Code]
-
GaussianProperty: "Integrating Physical Properties to 3D Gaussians with LMMs", arXiv 2025. [Project]
-
InfiniCube: "InfiniCube: Unbounded and Controllable Dynamic 3D Driving Scene Generation with World-Guided Video Models", arXiv 2024. [Paper]
-
GS-LIVOM: "Real-time Gaussian Splatting Assisted LiDAR-Inertial-Visual Odometry and Dense Mappings", arXiv2024. [Code]
-
OpenGS-SLAM: "OpenGS-SLAM: Open-Set Dense Semantic SLAM with 3D Gaussian Splatting for Object-Level Scene Understanding", arXiv2024. [Code]
Follow [Siheng Chen] for more information on V2X Perception, I will do supplement when I have time.
Follow [Runsheng Xu] for more information on V2X Perception, I will do supplement when I have time.
-
HighwayEnv: "An Environment for Autonomous Driving Decision-Making", GitHub. [Code]
-
OpenLane-V2: "The World's First Perception and Reasoning Benchmark for Scene Structure in Autonomous Driving.", GitHub. [Code]
-
OPen3d: "A Modern Library for 3D Data Processing", arxiv. [Paper] [Code]
-
SuMa++: "SuMa++: Efficient LiDAR-based Semantic SLAM", IEEE/RSJ International Conference on Intelligent Robots and Systems 2019. [Paper] [Code]
-
Khronos: "Khronos: A Unified Approach for Spatio-Temporal Metric-Semantic SLAM in Dynamic Environments", Robotics: Science and Systems. [Paper] [Code]
-
DRLSV: "Dense reinforcement learning for safety validation of autonomous vehicles", Nature 2023. [Paper]
-
Liquid AI: "Closed-form continuous-time neural networks", Nature 2023. [Paper]
-
UniDistill: "A Universal Cross-Modality Knowledge Distillation Framework for 3D Object Detection in Bird's-Eye View", CVPR 2023. [Code]
-
Let Occ Flow: "Let Occ Flow: Self-Supervised 3D Occupancy Flow Prediction", CoRL 2024. [Code]
-
UniLoc: "UniLoc: Towards Universal Place Recognition Using Any Single Modality", arxiv 2024. [Paper]
-
FAST-LIEO: "FAST-LIEO", github 2024. [Code]
-
Awesome-Robotics-Diffusion: "Awesome-Robotics-Diffusion", github 2024. [Code]
-
Diffusion Planner: "Diffusion-Based Planning for Autonomous Driving with Flexible Guidance", 2025 International Conference on Learning Representation (ICLR). [Code]
-
Learning More With Less: "Learning More With Less: Sample Efficient Dynamics Learning and Model-Based RL for Loco-Manipulation", arxiv 2025. [Paper]
-
SNN-VPR: "Applications of Spiking Neural Networks in Visual Place Recognition", IEEE Transactions on Robotics 2025. [Paper]
-
DriveLikeAHuman: "Rethinking Autonomous Driving with Large Language Models", CVPR, 2023. [Code]
-
ReAct: Synergizing Reasoning and Acting in Language Models[Code]
-
GraphRag[Code]
-
DreamerV3[Code]
-
DinoV2[Code]
-
CLIP[Code]
-
llama[Code]
-
Gato[Paper]
-
Open the Black Box of Transformers[Paper]
-
DeepSeek-V3[Code]
-
DeepSeek-R1[Code]
-
Lotus[Code]
-
VLN-CE-Isaac[Code]
-
DeepMind [Homepage]
-
NVlabs [Homepage]
-
NVIDIA Isaac [Homepage]
-
NVIDIA AI IOT [Homepage]
-
ROS-device-drivers [Homepage]
-
dora-rs [Homepage]
-
Eclipse zenoh [Homepage]
-
ADLINK-ROS [Homepage]
-
eProsima [Homepage]
-
ANYbotics [Homepage]
-
Idiap Research Institute [Homepage]
-
Waymo Research [Homepage]
-
Bosch Research [Homepage]
-
Motional [Homepage]
-
valeo.ai [Homepage]
-
wayveai [Homepage]
-
Foxglove [Homepage]
-
rerun [Homepage]
-
Udacity [Homepage]
-
Nature Robots [Homepage]
-
DeepSeek [Homepage]
-
livox [Homepage]
-
The International Digital Economy Academy [Homepage]
-
ADG@PJLab [Homepage]
-
OpenDriveLab [Homepage]
-
OpenRobotLab [Homepage]
-
Alibaba AD Lab [Homepage]
-
Sense-GVT [Homepage]
-
Megvii Robot [Homepage]
-
Megvii Research [Homepage]
-
Baidu Apollo Auto [Homepage]
-
Alibaba TongYi Vision Intelligence Lab [Homepage]
-
qcraftai [Homepage]
-
haomoai [Homepage]
-
HUAWEI Noah's Ark Lab [Homepage]
-
stepfun-ai [Homepage]
-
Toyota Research Institute Machine Learning [Homepage]
-
Oxford Control Group [Homepage]
-
Oxford Dynamic Robot Systems Group [Homepage]
-
Imperial College London Dyson Robotics Laboratory [Homepage]
-
UCL CDT in Foundational AI [Homepage]
-
UCL Robot Perception and Learning Lab [Homepage]
-
School of Informatics University of Edinburgh [Homepage]
-
ETHZ Autonomous Systems Lab [Homepage]
-
ETHZ Robotic Systems Lab [Homepage]
-
ETHZ Computer-Vision-and-Geometry-Lab [Homepage]
-
ETHZ Visual Intelligence and Systems Group [Homepage]
-
ETHZ University of Cyprus Vision for Robotics Lab [Homepage]
-
ETHZ Learning and Adaptive Systems Group [Homepage]
-
ETHZ Photogrammetry and Remote Sensing Lab [Homepage]
-
UZH Robotics and Perception Group [Homepage]
-
EPFL CV LAB [Homepage]
-
TUM-Institue-of-Auotomative-Technology [Homepage]
-
TUM-Computer-Vision-Group [Homepage]
-
TUM Smart Robotics Lab [Homepage]
-
Stuttgart Flight Robotics and Perception Group [Homepage]
-
Heidelberg Computer Vision and Learning Lab [Homepage]
-
UFR Robot Learning [Homepage]
-
University of Tübingen [Homepage]
-
Bonn Photogrammetry Robotics [Homepage]
-
Karlsruhe Institute of Technology Institute of Measurement and Control Systems [Homepage]
-
IGMR - RWTH Aachen University [Homepage]
-
RWTH Aachen Institute of Automatic Control [Homepage]
-
RWTH Aachen Institut für Kraftfahrzeuge ika [Homepage]
-
Goettingen Data Fusion Group [Homepage]
-
Albert-Ludwigs-Universität Freiburg Robot Learning Lab [Homepage]
-
KTH Robotics Perception and Learning [Homepage]
-
University of Turku Turku Intelligent Embedded and Robotic Systems Lab [Homepage]
-
Istituto Italiano di Tecnologia Robotics [Homepage]
-
NMBU Robotics [Homepage]
-
TU Delft 3D geoinformation research group [Homepage]
-
TU Delft Intelligent Vehicles [Homepage]
-
TU Delft Autonomous Multi-Robots Lab [Homepage]
-
Poznan University of Technology Mobile Robots Lab [Homepage]
-
Sapienza University of Rome Robots Vision and Perception [Homepage]
-
LTU Robotics & AI Engineering Group [Homepage]
-
Control of Networked Systems Group at University of Klagenfurt [Homepage]
-
Vision for Robotics and Autonomous Systems Czech Technical University in Prague (CTU) [Homepage]
-
Stanford Autonomous Systems Lab [Homepage]
-
Stanford Robotics Embodied Artificial Intelligence Lab REAL [Homepage]
-
Stanford Vision and Learning Lab [Homepage]
-
Stanford Computational Vision and Geometry Lab [Homepage]
-
Stanford Center for Research on Foundation Models [Homepage]
-
Stanford Open Virtual Assistant Lab [Homepage]
-
MIT-SPARKs [Homepage]
-
MIT Driverless Perception Team [Homepage]
-
MIT CSAIL Computer Vision [Homepage]
-
MIT's Marine Robotics Group [Homepage]
-
MIT HAN Lab [Homepage]
-
MIT Aerospace Controls Laboratory [Homepage]
-
MIT Robust Robotics Group [Homepage]
-
MIT Urban Mobility Lab + Transit Lab [Homepage]
-
CMU Robotic Exploration Lab [Homepage]
-
CMU Robot Perception Lab [Homepage]
-
Learning and Control for Agile Robotics Lab at CMU [Homepage]
-
Princeton Computational Imaging [Homepage]
-
Princeton Safe Robotics Lab [Homepage]
-
UPenn Perception Action Learning Group [Homepage]
-
UPenn Kumar Robotics [Homepage]
-
UCB Model Predictive Control Laboratory [Homepage]
-
Berkeley Automation Lab [Homepage]
-
UCLA Verifiable & Control-Theoretic Robotics Laboratory [Homepage]
-
UCLA Mobility Lab [Homepage]
-
Cornell Tech Kuleshov Group [Homepage]
-
UCSD Existential Robotics Lab [Homepage]
-
UCSD Hao Su's Lab [Homepage]
-
Umich Ford Center for Autonomous Vehicles FCAV [Homepage]
-
Umich The Autonomous Robotic Manipulation Lab studies motion planning manipulation human-robot collaboration [Homepage]
-
Umich Dynamic Legged Locomotion Robotics Lab [Homepage]
-
Umich Computational Autonomy and Robotics Laboratory [Homepage]
-
NYU AI4CE Lab [Homepage]
-
GaTech BORG Lab [Homepage]
-
GaTech Intelligent Vision and Automation Laboratory IVA Lab [Homepage]
-
Special Interest Group for Robotics Enthusiasts at UIUC [Homepage]
-
UIUC-Robotics [Homepage]
-
Texas Austin Autonomous Mobile Robotics Laboratory [Homepage]
-
Texas Austin Visual Informatics Group [Homepage]
-
University of Delaware Robot Perception Navigation Group RPNG [Homepage]
-
Virginia Tech Transportation Institute [Homepage]
-
SIT Robust Field Autonomy Lab [Homepage]
-
University at Buffalo Spatial AI & Robotics Lab [Homepage]
-
Toronto STARS Laboratory [Homepage]
-
Toronto Autonomous Space Robotics Lab [Homepage]
-
UBC Computer Vision Group [Homepage]
-
UWaterloo CL2 [Homepage]
-
École Polytechnique de Montréal Making Innovative Space Technology Lab [Homepage]
-
Université Laval Northern Robotics Laboratory [Homepage]
-
KAIST Urban Robotics Lab [Homepage]
-
KAIST Cognitive Learning for Vision and Robotics CLVR lab [Homepage]
-
Yonsei Computational Intelligence Laboratory [Homepage]
-
SNU RPM [Homepage]
-
SNU Machine Perception and Reasoning Lab [Homepage]
-
DGIST APRL [Homepage]
-
Japan National Institute of Advanced Industrial Science and Technology [Homepage]
-
Japan Nagoya [Homepage]
-
NUS showlab [Homepage]
-
NTU ARIS [Homepage]
-
CUHK OpenMMLab [Homepage]
-
HKUST Aerial Robotics Group [Homepage]
-
HKUST XRIM-Lab [Homepage]
-
HKU Mars Lab [Homepage]
-
HKU CVMI Lab [Homepage]
-
CUHK Deep Vision Lab [Homepage]
-
City University of Hong Kong MetaSLAM [Homepage]
-
HK PolyU Visual Learning and Reasoning Group [Homepage]
-
NTNU Autonomous Robots Lab [Homepage]
-
Tsinghua IIIS MARS Lab [Homepage]
-
Tsinghua Institute for AI Industry Research [Homepage]
-
SJTU Vision and Intelligent System Group [Homepage]
-
SJTU Intelligent Robotics and Machine Vision Lab [Homepage]
-
SJTU Thinklab [Homepage]
-
ZJU Advanced-Perception-on-Robotics-and-Intelligent-Learning-Lab [Homepage]
-
ZJU CAD CG [Homepage]
-
ZJU Advanced Intelligent Machines AIM [Homepage]
-
ZJU Robotics Lab [Homepage]
-
ZJU FAST Lab [Homepage]
-
ZJU OM-AI Lab [Homepage]
-
Fudan Zhang Vision Group [Homepage]
-
Chongxuan Li's research group @ Renmin University of China [Homepage]
-
Tongji Intelligent Electric Vehicle Research Group [Homepage]
-
Tongji Intelligent Sensing Perception and Computing Group [Homepage]
-
NUDT NuBot [Homepage]
-
HUST EIC Vision Lab [Homepage]
-
ShanghaiTech Vision and Intelligent Perception Lab [Homepage]
-
HITSZ nROS-Lab [Homepage]
-
GAP-LAB-CUHK-SZ [Homepage]
-
Westlake University Audio Signal and Information Processing Lab [Homepage]
-
Wuhan University Urban Spatial Intelligence Research Group at LIESMARS [Homepage]
-
Wuhan University Integrated and Intelligent Navigation Group [Homepage]
-
WHU GREAT (GNSS+ Research, Application and Teaching) group [Homepage]
-
SYSU STAR Smart Aerial Robotics Group [Homepage]
-
SYSU RAPID Lab [Homepage]
-
SYSU Pengyu Team [Homepage]
-
NKU Robot Autonomy and Human-AI Collaboration Group [Homepage]
-
HKUSTGZ Research group of visual generative models [Homepage]
-
HNU Neuromorphic Automation and Intelligence Lab [Homepage]
-
NEU REAL [Homepage]
-
Israel Autonomous Navigation and Sensor Fusion Lab [Homepage]
-
CSIRORobotics Brisbane, Queensland, Australia [Homepage]
-
Robotics Institute, University of Technology Sydney, Sydney, Australia [Homepage]
-
Robotic Perception Group at the Australian Centre For Robotics, Sydney, Australia [Homepage]
Find more on this link in case of u r looking for PhD positions[I] [II].
-
Science Robotics [Registeration]
-
TRO IEEE Transactions on Robotics [Registeration]
-
IJRR International Journal of Robotics Research [Registeration]
-
JFR Journal of Field Robotics [Registeration]
-
Transportation Research Part C [Registeration]
-
TII IEEE Transactions on Industrial Informatics [Registeration]
-
TITS IEEE Transactions on Intelligent Transportation Systems [Registeration]
-
TASE IEEE Transactions on Automation Science and Engineering [Registeration]
-
TMech IEEE/ASME Transactions on Mechatronics [Registeration]
-
TGRS IEEE Transactions on Geoscience and Remote Sensing [Registeration]
-
TCSVT IEEE Transactions on Circuits and Systems for Video Technology [Registeration]
-
TIV IEEE Transactions on Intelligent Vehicles [Registeration]
-
TIM IEEE Transactions on Instrumentation & Measurement [Registeration]
-
RA-L IEEE Robotics and Automation Letters [Registeration]
-
TVT IEEE Transactions on Vehicular Technology [Registeration]
-
TTE IEEE Transactions on Transportation Electrification [Registeration]
-
TIE IEEE Transactions on Industrial Electronics [Registeration]
-
Communications in Transportation Research [Registeration]
-
IEEE Open Journal of Intelligent Transportation Systems [Registeration]
-
Mar. RSS Robotics: Science and Systems [Registeration]
-
Nov. CVPR IEEE Conference on Computer Vision and Pattern Recognition [Registeration]
-
Sept. ICRA IEEE International Conference on Robotics and Automation [Registeration]
-
Mar. IROS IEEE/RSJ lnternational Conference onlntelligent Robots and Systems [Registeration]
-
Mar. ICCV International Conference on Computer Vision [Registeration]
-
Mar. ECCV European Conference on Computer Vision [Registeration]
-
Jan. SIGGRAPH Special Interest Group for Computer GRAPHICS [Registeration]
-
Jul. CoRL Conference on Robot Learning [Registeration]
-
Sept. ICLR International Conference on Learning Representations [Registeration]
-
NYU MARS [Homepage]
-
KAIST Urban [Homepage]
-
EuRoC [Homepage]
-
MCD Dataset [Homepage]
-
MSU-4S [Homepage]
-
HoloVIC [Homepage]
-
Oxford Robotcar [Homepage]
-
NCLT [Homepage]
-
Newer College [Homepage]
-
Mulran [Homepage]
-
KITTI [Homepage]
-
Argoverse [Homepage]
-
Onpenlanev2 [Homepage]
-
M2DGR [Homepage]
-
GEODE [Homepage]
-
SubT [Homepage]
-
darpa_subt_worlds [Homepage]
-
R3LIVE [Homepage]
-
WildScenes [Homepage]
-
RELLIS-3D [Homepage]
-
Rubble [Homepage]
-
TIERS [Homepage]
-
HeLiPR [Homepage]
-
GEODE [Homepage]
-
CVGlobal [Homepage]
-
mapillary [Homepage]
-
SLABIM [Homepage]
-
BorealTC [Homepage]
-
fusionportable [Homepage]
find more in paperwithcode
Practice makes perfect, though lack of innovation.
- ICRA 2023 Robodepth [Link]
- ICRA 2024 RoboDrive [Link]
- ICRA 2023 Sim2Real [Link]
- nerfbaselines [Link]
Libs:
NeRF/3DGS:
Calibration:
- Kalibr [CamCam/CamIMU/IMUIMU/Intrinsic]
- iKalibr [IMU/Cam/Radar/LiDAR/RGBD]
- SensorsCalibration [LiDARCam/LiDARIMU/LiDARLiDAR/RadarCam/RadarLiDAR/CamCam/Intrinsic]
- livox_camera_calib [LiDARCam]
- direct_visual_lidar_calibration [LiDARCam]
- Calibration-Is-All-You-Need [LiDARCam/LiDARIMU/CamIMU/CamCam]
Evaluation:
- EVO [EVO]
- CloudMapEvaluation [CloudMapEvaluation]
- SLAMHIVE [SLAMHIVE]
- Robustness_Metric [SLAMHIVE]
Communication:
-
Message Queuing Telemetry Transport [Video]
-
Hypertext Transfer Protocol Secure [Video]
-
Controller Area Network [Video]
Deep Learning Framework:
Writing:
- NSFC-Latex [Video]
Books:
-
learnLLM[Code]
-
Foundations-of-LLMs[Code]
-
MathRL[Code]
-
SLAM-Handbook[Code]
-
slambook2[Code]
-
TransferLearning[Code]
- AAI [Sim now]
- Airsim [Sim now]
- aiSim [Sim now]
- Apollo [Sim now]
- CARLA [Sim now]
- CarMaker [Sim now]
- CarSim [Sim now]
- Cognata [Sim now]
- ESI Pro-Sivic [Sim now]
- Metamoto [Sim now]
- NVDIA Drive Sim [Sim now]
- NVDIA Isaac Sim [Sim now]
- OnSite [Sim now]
- PanoSim [Sim now]
- ParallelDomain [Sim now]
- Pilot-D GaiA [Sim now]
- PreScan [Sim now]
- rFpro [Sim now]
- RightHook [Sim now]
- RoadRunner [Sim now]
- SCANeR [Sim now]
- SILAB [Sim now]
- TruckMaker [Sim now]
- Vires VTD [Sim now]
- VI Worldsim [Sim now]
- ZJL-VTS [Sim now]
- MineSim [Sim now]
- 51Sim-one [Sim now]
- Mujoco [Sim now]
- Genesis [Sim now]
- Cosmos [Sim now]
Multi-sensors:RGB, RGBD, Infrared, Polarization, Event-based(Future), Motion, Capture(GT), 2D Spinning LiDAR, 3D Spinning LiDAR, 3D Solid-State LiDAR, 4D High Resolution Radar(Future), mmWave Radar, UWB, RTK(GT), Wheel Odom, IMU, Ultrasonic Sensor.
For bridging college and society, mainly base in Shanghai, Beijing, Hangzhou, Suzhou, Canton, Shenzhen, HongKong
Shanghai: AI Lab, DiDi, Horizon, Li Auto, BOSCH, Huawei, SenseTime,
Beijing: Xiaomi, Baidu, Pony.ai,
Hangzhou: Unitree, Damo,
Suzhou: Momenta,
Canton: ZhuoYu, WeRide, Xiaopeng,
interview experience like Find More
BTW, RA/TA Opportunities:
CMU AirLab
A STAR
CityU MetaSLAM
ETH-RSS
Follow [XueLingFeiHua] for general understanding of sensors used by autonomous vehicle, I will do supplement when I have time.
For autonomous driving vechile and outdoor robotics, use more light-weight map instead of no map, specifically more like maplite [demo]
Especially for indoor environments, the prior is BIM [BIM]
-
RoadNet: "Translating Images to Road Network: A Non-Autoregressive Sequence-to-Sequence Approach", ICCV 2023 (Oral). [Paper] [Code]
-
OrienterNet: "OrienterNet: Visual Localization in 2D Public Maps with Neural Matching", CVPR 2023. [Paper] [Code]
-
OSMLoc: "OSMLoc: Single Image-Based Visual Localization in OpenStreetMap with Semantic and Geometric Guidances", arxiv 2024. [Paper] [Code]
-
svo-dt: " Drift-free Visual SLAM using Digital Twins", arxiv 2024. [Paper] [Code]
-
CityNav: "CityNav: Language-Goal Aerial Navigation Dataset with Geographic Information", arxiv 2024. [Paper]
-
DFVDT: "Drift-free Visual SLAM using Digital Twins", arxiv 2024. [Paper]
-
BEVPlace2: "BEVPlace++: Fast, Robust, and Lightweight LiDAR Global Localization for Unmanned Ground Vehicles", arxiv 2024. [Code]
-
TripletLoc: "TripletLoc: One-Shot Global Localization using Semantic Triplet in Urban Environment", arxiv2024. [Code]
-
Reliable-loc: "Reliable LiDAR global localization using spatial verification and pose uncertainty", arxiv 2024. [Code]
-
Render2Loc: "Render-and-Compare: Cross-view 6-DoF Localization from Noisy Prior", 2023 IEEE International Conference on Multimedia and Expo (ICME). [Code]
-
CityWalker: "CityWalker: Learning Embodied Urban Navigation from Web-Scale Videos", 2025 CVPR. [Code]
-
EI-Nav: "OPEN: Openstreetmap-enhanced oPen-air sEmantic Navigation", 2025 arxiv. [Code]
-
DeepGPS: "DeepGPS: deep learning enhanced GPS positioning in urban canyons", 2024 IEEE Transactions on Mobile Computing (TMC). [Code]
Follow [TUMFTM] for more information on HDMap & SDMap aided localization, I will do supplement when I have time.
Follow [Tong Qin] for more information on SDMap aided localization, I will do supplement when I have time.
Follow [Yujiao Shi] and [tudelft-iv] for more information on Satellite images aided localization, I will do supplement when I have time.
For all terrain navigation, because We're tired of filtering ground points...
-
MonoForce: "MonoForce: Self-supervised Learning of Physics-aware Model for Predicting Robot-terrain Interaction", arXiv 2022. [Paper] [Code]
-
TML: "A Global Traversability Mapping Library easily integratable with any SLAM System". [Code]
-
TMMP: "Bayesian Generalized Kernel Inference for Terrain Traversability Mapping", the 2nd Annual Conference on Robot Learning. [Code]
-
STEPP: "STEPP: Semantic Traversability Estimation using Pose Projected features", not yet. [Code]
-
HFHWC: "Autonomous Driving in Unstructured Environments: How Far Have We Come?", [Code]
-
EcSemMap: "Uncertainty-aware Evidential Bayesian Semantic Mapping (EBS)", [Code]
-
LoD: "Learning-on-the-Drive: Self-supervised Adaptive Long-range Perception for High-speed Offroad Driving", [Paper]
-
tapdo: "tadpo", [Code]
-
ROLO-SLAM: "ROLO-SLAM: Rotation-Optimized LiDAR-Only SLAM in Uneven Terrain with Ground Vehicle", [Code]
-
MGGPlanner: "Multi-robot Grid Graph Exploration Planner", [Code]
-
GroundGrid: "GroundGrid: LiDAR Point Cloud Ground Segmentation and Terrain Estimation", [Code]
For multi-session mapping and updating (change detection), dynamic objects filter.
-
Present and Future of SLAM in Extreme Environments: "Present and Future of SLAM in Extreme Environments: The DARPA SubT Challenge", 2022 IEEE Transactions on Robotics. [Paper]
-
LT-Mapper: "Lt-mapper: A modular framework for lidar-based lifelong mapping", 2022 International Conference on Robotics and Automation (ICRA). [Paper] [Code]
-
RoLL: "ROLL: Long-Term Robust LiDAR-based Localization With Temporary Mapping in Changing Environments", arxiv 2025. [Code]
-
Elite: "Ephemerality meets LiDAR-based Lifelong Mapping", arxiv 2025. [Code]
-
SOLiD: "SOLiD: Spatially Organized and Lightweight Global Descriptor for FOV-constrained LiDAR Place Recognition", IEEE ROBOTICS AND AUTOMATION LETTERS 2024. [Code]
-
HeLiOS: "HeLiOS: Heterogeneous LiDAR Place Recognition via Overlap-based Learning and Local Spherical Transformer", 2025 International Conference on Robotics and Automation (ICRA). [Code]
-
HHRM: " Lifelong 3D Mapping Framework for Hand-held & Robot-mounted LiDAR Mapping Systems", IEEE ROBOTICS AND AUTOMATION LETTERS 2024. [Paper]
-
LiLoc: " LiLoc: Lifelong Localization using Adaptive Submap Joining and Egocentric Factor Graph", 2025 International Conference on Robotics and Automation (ICRA). [Code]
-
KISS-Matcher: "KISS-Matcher: Fast and Robust Point Cloud Registration Revisited", arxiv 2024. [Paper] [Code]
-
G3Reg: "Pyramid Semantic Graph-based Global Point Cloud Registration with Low Overlap", IEEE/RSJ International Conference on Intelligent Robots and Systems 2023. [Code]
-
SG-Reg: "SG-Reg: Generalizable and Efficient Scene Graph Registration", TRO under review. [Code]
-
DeepPointMap: "DeepPointMap: Advancing LiDAR SLAM with Unified Neural Descriptors", AAAI Conference on Artificial Intelligence 2024. [Code]
-
GLoc3D: "Global Localization in Large-scale Point Clouds via Roll-pitch-yaw Invariant Place Recognition and Low-overlap Global Registration", *IEEE Transactions on Circuits and Systems for Video Technology *. [Code]
-
maplalb2.0: "maplab 2.0 – A Modular and Multi-Modal Mapping Framework", IEEE Robotics and Automation Letters. [Code]
-
MR-SLAM: "Disco: Differentiable scan context with orientation", IEEE Robotics and Automation Letters. [Code]
-
MS-Mapping: "MS-Mapping: An Uncertainty-Aware Large-Scale Multi-Session LiDAR Mapping System", arxiv 2024. [Paper] [Code]
-
NF-Atlas: "Multi-Volume Neural Feature Fields for Large Scale LiDAR Mapping", IEEE Robotics and Automation Letters 2023. [Paper] [Code]
-
DiSCo-SLAM: "DiSCo-SLAM: Distributed Scan Context-Enabled Multi-Robot LiDAR SLAM With Two-Stage Global-Local Graph Optimization", IEEE International Conference on Robotics and Automation 2022. [Paper] [Code]
-
imesa: "{iMESA}: Incremental Distributed Optimization for Collaborative Simultaneous Localization and Mapping", Robotics: Science and Systems 2024. [Paper] [Code]
-
DCL-SLAM: "Swarm-lio: Decentralized swarm lidar-inertial odometry", IEEE International Conference on Robotics and Automation 2023. [Paper] [Code]
-
ROAM: "Riemannian Optimization for Active Mapping with Robot Teams", IEEE Transactions on Robotics 2025. [Project]
-
BlockMap: "Block-Map-Based Localization in Large-Scale Environment", IEEE International Conference on Robotics and Automation 2024. [Paper] [Code]
-
SLIM: "SLIM: Scalable and Lightweight LiDAR Mapping in Urban Environments", arxiv maybe tro 2024. [Paper] [Code]
reduce z-axis drift
-
Norlab-icp: "A 2-D/3-D mapping library relying on the "Iterative Closest Point" algorithm.", [Code]
Follow [Ji Zhang] for more information on Robust Navigation, I will do supplement when I have time.
Follow [Xiang Gao] for more information on SLAM, I will do supplement when I have time.
Follow [Tong Qin] for more information on V-SLAM, I will do supplement when I have time.
Follow [Gisbi Kim] for more information on Map Maintaining, I will do supplement when I have time.
Follow [Kin Zhang] for more information on Dynamic Objects Removing, I will do supplement when I have time.
Follow [Zhijian Qiao] for more information on Multi-Session Mapping, I will do supplement when I have time.
Follow [koide] for more information on Life-Long LiDAR Mapping, I will do supplement when I have time.
If you find this repository useful, please consider citing this repo:
@misc{runjtu2024slamrepo,
title = {awesome-and-novel-works-in-slam},
author = {Runheng Zuo},
howpublished = {\url{https://github.com/runjtu/awesome-and-novel-works-in-slam}},
year = {2024},
note = "[Online; accessed 04-October-2024]"
}
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for awesome-and-novel-works-in-slam
Similar Open Source Tools

awesome-and-novel-works-in-slam
This repository contains a curated list of cutting-edge works in Simultaneous Localization and Mapping (SLAM). It includes research papers, projects, and tools related to various aspects of SLAM, such as 3D reconstruction, semantic mapping, novel algorithms, large-scale mapping, and more. The repository aims to showcase the latest advancements in SLAM technology and provide resources for researchers and practitioners in the field.

Awesome-LLM-Interpretability
Awesome-LLM-Interpretability is a curated list of materials related to LLM (Large Language Models) interpretability, covering tutorials, code libraries, surveys, videos, papers, and blogs. It includes resources on transformer mechanistic interpretability, visualization, interventions, probing, fine-tuning, feature representation, learning dynamics, knowledge editing, hallucination detection, and redundancy analysis. The repository aims to provide a comprehensive overview of tools, techniques, and methods for understanding and interpreting the inner workings of large language models.

Awesome_papers_on_LLMs_detection
This repository is a curated list of papers focused on the detection of Large Language Models (LLMs)-generated content. It includes the latest research papers covering detection methods, datasets, attacks, and more. The repository is regularly updated to include the most recent papers in the field.

Awesome_Test_Time_LLMs
This repository focuses on test-time computing, exploring various strategies such as test-time adaptation, modifying the input, editing the representation, calibrating the output, test-time reasoning, and search strategies. It covers topics like self-supervised test-time training, in-context learning, activation steering, nearest neighbor models, reward modeling, and multimodal reasoning. The repository provides resources including papers and code for researchers and practitioners interested in enhancing the reasoning capabilities of large language models.

awesome-LLM-game-agent-papers
This repository provides a comprehensive survey of research papers on large language model (LLM)-based game agents. LLMs are powerful AI models that can understand and generate human language, and they have shown great promise for developing intelligent game agents. This survey covers a wide range of topics, including adventure games, crafting and exploration games, simulation games, competition games, cooperation games, communication games, and action games. For each topic, the survey provides an overview of the state-of-the-art research, as well as a discussion of the challenges and opportunities for future work.

Awesome-Embodied-Agent-with-LLMs
This repository, named Awesome-Embodied-Agent-with-LLMs, is a curated list of research related to Embodied AI or agents with Large Language Models. It includes various papers, surveys, and projects focusing on topics such as self-evolving agents, advanced agent applications, LLMs with RL or world models, planning and manipulation, multi-agent learning and coordination, vision and language navigation, detection, 3D grounding, interactive embodied learning, rearrangement, benchmarks, simulators, and more. The repository provides a comprehensive collection of resources for individuals interested in exploring the intersection of embodied agents and large language models.

LogChat
LogChat is an open-source and free AI chat client that supports various chat models and technologies such as ChatGPT, 讯飞星火, DeepSeek, LLM, TTS, STT, and Live2D. The tool provides a user-friendly interface designed using Qt Creator and can be used on Windows systems without any additional environment requirements. Users can interact with different AI models, perform voice synthesis and recognition, and customize Live2D character models. LogChat also offers features like language translation, AI platform integration, and menu items like screenshot editing, clock, and application launcher.

ABigSurveyOfLLMs
ABigSurveyOfLLMs is a repository that compiles surveys on Large Language Models (LLMs) to provide a comprehensive overview of the field. It includes surveys on various aspects of LLMs such as transformers, alignment, prompt learning, data management, evaluation, societal issues, safety, misinformation, attributes of LLMs, efficient LLMs, learning methods for LLMs, multimodal LLMs, knowledge-based LLMs, extension of LLMs, LLMs applications, and more. The repository aims to help individuals quickly understand the advancements and challenges in the field of LLMs through a collection of recent surveys and research papers.

Awesome-World-Models
This repository is a curated list of papers related to World Models for General Video Generation, Embodied AI, and Autonomous Driving. It includes foundation papers, blog posts, technical reports, surveys, benchmarks, and specific world models for different applications. The repository serves as a valuable resource for researchers and practitioners interested in world models and their applications in robotics and AI.

Awesome-Embodied-AI-Job
Awesome Embodied AI Job is a curated list of resources related to jobs in the field of Embodied Artificial Intelligence. It includes job boards, companies hiring, and resources for job seekers interested in roles such as robotics engineer, computer vision specialist, AI researcher, machine learning engineer, and data scientist.

awesome_LLM-harmful-fine-tuning-papers
This repository is a comprehensive survey of harmful fine-tuning attacks and defenses for large language models (LLMs). It provides a curated list of must-read papers on the topic, covering various aspects such as alignment stage defenses, fine-tuning stage defenses, post-fine-tuning stage defenses, mechanical studies, benchmarks, and attacks/defenses for federated fine-tuning. The repository aims to keep researchers updated on the latest developments in the field and offers insights into the vulnerabilities and safeguards related to fine-tuning LLMs.

MedLLMsPracticalGuide
This repository serves as a practical guide for Medical Large Language Models (Medical LLMs) and provides resources, surveys, and tools for building, fine-tuning, and utilizing LLMs in the medical domain. It covers a wide range of topics including pre-training, fine-tuning, downstream biomedical tasks, clinical applications, challenges, future directions, and more. The repository aims to provide insights into the opportunities and challenges of LLMs in medicine and serve as a practical resource for constructing effective medical LLMs.

Paper-Reading-ConvAI
Paper-Reading-ConvAI is a repository that contains a list of papers, datasets, and resources related to Conversational AI, mainly encompassing dialogue systems and natural language generation. This repository is constantly updating.

Awesome-Robotics-3D
Awesome-Robotics-3D is a curated list of 3D Vision papers related to Robotics domain, focusing on large models like LLMs/VLMs. It includes papers on Policy Learning, Pretraining, VLM and LLM, Representations, and Simulations, Datasets, and Benchmarks. The repository is maintained by Zubair Irshad and welcomes contributions and suggestions for adding papers. It serves as a valuable resource for researchers and practitioners in the field of Robotics and Computer Vision.

OpenRedTeaming
OpenRedTeaming is a repository focused on red teaming for generative models, specifically large language models (LLMs). The repository provides a comprehensive survey on potential attacks on GenAI and robust safeguards. It covers attack strategies, evaluation metrics, benchmarks, and defensive approaches. The repository also implements over 30 auto red teaming methods. It includes surveys, taxonomies, attack strategies, and risks related to LLMs. The goal is to understand vulnerabilities and develop defenses against adversarial attacks on large language models.
For similar tasks

awesome-and-novel-works-in-slam
This repository contains a curated list of cutting-edge works in Simultaneous Localization and Mapping (SLAM). It includes research papers, projects, and tools related to various aspects of SLAM, such as 3D reconstruction, semantic mapping, novel algorithms, large-scale mapping, and more. The repository aims to showcase the latest advancements in SLAM technology and provide resources for researchers and practitioners in the field.
For similar jobs

Qwen-TensorRT-LLM
Qwen-TensorRT-LLM is a project developed for the NVIDIA TensorRT Hackathon 2023, focusing on accelerating inference for the Qwen-7B-Chat model using TRT-LLM. The project offers various functionalities such as FP16/BF16 support, INT8 and INT4 quantization options, Tensor Parallel for multi-GPU parallelism, web demo setup with gradio, Triton API deployment for maximum throughput/concurrency, fastapi integration for openai requests, CLI interaction, and langchain support. It supports models like qwen2, qwen, and qwen-vl for both base and chat models. The project also provides tutorials on Bilibili and blogs for adapting Qwen models in NVIDIA TensorRT-LLM, along with hardware requirements and quick start guides for different model types and quantization methods.

dl_model_infer
This project is a c++ version of the AI reasoning library that supports the reasoning of tensorrt models. It provides accelerated deployment cases of deep learning CV popular models and supports dynamic-batch image processing, inference, decode, and NMS. The project has been updated with various models and provides tutorials for model exports. It also includes a producer-consumer inference model for specific tasks. The project directory includes implementations for model inference applications, backend reasoning classes, post-processing, pre-processing, and target detection and tracking. Speed tests have been conducted on various models, and onnx downloads are available for different models.

joliGEN
JoliGEN is an integrated framework for training custom generative AI image-to-image models. It implements GAN, Diffusion, and Consistency models for various image translation tasks, including domain and style adaptation with conservation of semantics. The tool is designed for real-world applications such as Controlled Image Generation, Augmented Reality, Dataset Smart Augmentation, and Synthetic to Real transforms. JoliGEN allows for fast and stable training with a REST API server for simplified deployment. It offers a wide range of options and parameters with detailed documentation available for models, dataset formats, and data augmentation.

ai-edge-torch
AI Edge Torch is a Python library that supports converting PyTorch models into a .tflite format for on-device applications on Android, iOS, and IoT devices. It offers broad CPU coverage with initial GPU and NPU support, closely integrating with PyTorch and providing good coverage of Core ATen operators. The library includes a PyTorch converter for model conversion and a Generative API for authoring mobile-optimized PyTorch Transformer models, enabling easy deployment of Large Language Models (LLMs) on mobile devices.

awesome-RK3588
RK3588 is a flagship 8K SoC chip by Rockchip, integrating Cortex-A76 and Cortex-A55 cores with NEON coprocessor for 8K video codec. This repository curates resources for developing with RK3588, including official resources, RKNN models, projects, development boards, documentation, tools, and sample code.

cl-waffe2
cl-waffe2 is an experimental deep learning framework in Common Lisp, providing fast, systematic, and customizable matrix operations, reverse mode tape-based Automatic Differentiation, and neural network model building and training features accelerated by a JIT Compiler. It offers abstraction layers, extensibility, inlining, graph-level optimization, visualization, debugging, systematic nodes, and symbolic differentiation. Users can easily write extensions and optimize their networks without overheads. The framework is designed to eliminate barriers between users and developers, allowing for easy customization and extension.

TensorRT-Model-Optimizer
The NVIDIA TensorRT Model Optimizer is a library designed to quantize and compress deep learning models for optimized inference on GPUs. It offers state-of-the-art model optimization techniques including quantization and sparsity to reduce inference costs for generative AI models. Users can easily stack different optimization techniques to produce quantized checkpoints from torch or ONNX models. The quantized checkpoints are ready for deployment in inference frameworks like TensorRT-LLM or TensorRT, with planned integrations for NVIDIA NeMo and Megatron-LM. The tool also supports 8-bit quantization with Stable Diffusion for enterprise users on NVIDIA NIM. Model Optimizer is available for free on NVIDIA PyPI, and this repository serves as a platform for sharing examples, GPU-optimized recipes, and collecting community feedback.

depthai
This repository contains a demo application for DepthAI, a tool that can load different networks, create pipelines, record video, and more. It provides documentation for installation and usage, including running programs through Docker. Users can explore DepthAI features via command line arguments or a clickable QT interface. Supported models include various AI models for tasks like face detection, human pose estimation, and object detection. The tool collects anonymous usage statistics by default, which can be disabled. Users can report issues to the development team for support and troubleshooting.