Starred repositories
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
Collection of reinforcement learning algorithms
verl: Volcano Engine Reinforcement Learning for LLMs
[CVPR 2025] UniGoal: Towards Universal Zero-shot Goal-oriented Navigation
Heterogeneous Pre-trained Transformer (HPT) as Scalable Policy Learner.
Unleashing the Power of VLMs in Autonomous Driving via Reinforcement Learning and Reasoning
Official repository of General Scene Adaptation for Vision-and-Language Navigation (ICLR'2025)
[NeurIPS 2024] CLOVER: Closed-Loop Visuomotor Control with Generative Expectation for Robotic Manipulation
Vision-Language Navigation Benchmark in Isaac Lab
Low-level locomotion policy training in Isaac Lab
Unitree robot sdk version 2. https://support.unitree.com/home/zh/developer
[NeurIPS 2024] SG-Nav: Online 3D Scene Graph Prompting for LLM-based Zero-shot Object Navigation
[NeurIPS 2023] Official Implementation of A Generic Active Learning Baseline for LiDAR Semantic Segmentation
[ECCV'24] Official Implementation of SemiVL: Semi-Supervised Semantic Segmentation with Vision-Language Guidance
Orocos Kinematics and Dynamics C++ library
[CVPR23] Official Implementation of MIC: Masked Image Consistency for Context-Enhanced Domain Adaptation
[ICCV23] EDAPS: Enhanced Domain-Adaptive Panoptic Segmentation
Qwen2.5-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.
[CVPR24] Volumetric Environment Representation for Vision-Language Navigation
The Most Faithful Implementation of Segment Anything (SAM) in 3D
[CVPR2024 Highlight]GLEE: General Object Foundation Model for Images and Videos at Scale
[3DV'25] 3D Reconstruction with Spatial Memory
[ECCV 2024] Official implementation of the paper "Semantic-SAM: Segment and Recognize Anything at Any Granularity"