Stars
InstantSplat: Sparse-view SfM-free Gaussian Splatting in Seconds
3D-Aware Vision-Language Models Fine-Tuning with Geometric Distillation
Code release for paper "Reconstructing People, Places, and Cameras", In CVPR 2025 (Highlight)
[SIGGRAPH Asia 2023 (Technical Communications)] EasyVolcap: Accelerating Neural Volumetric Video Research
The official implementation of Flexible Motion In-betweening with Diffusion Models, SIGGRAPH 2024
[NeurIPS D&B Track 2024] Official implementation of HumanVid
UniSkill: Imitating Human Videos via Cross-Embodiment Skill Representations
Parameterizing Everyday Home Activities Towards 3D Generative Modeling of Human-Object Interactions
[CVPR'25 Oral] MoGe: Unlocking Accurate Monocular Geometry Estimation for Open-Domain Images with Optimal Training Supervision
[EMNLP'24 Oral] Official PyTorch implementation of "Text2Chart31: Instruction Tuning for Chart Generation with Automatic Feedback"
CLoSD: Closing the Loop between Simulation and Diffusion for multi-task character control
The official PyTorch implementation of the paper "Human Motion Diffusion Model"
This is the official repository of SIGGRAPH Asia 2024 Paper: Autonomous Character-Scene Interaction Synthesis from Text Instruction
Taming Video Diffusion Prior with Scene-Grounding Guidance for 3D Gaussian Splatting from Sparse Inputs (CVPR2025 Highlight)
OmniControl: Control Any Joint at Any Time for Human Motion Generation, ICLR 2024
DNO: Optimizing Diffusion Noise Can Serve As Universal Motion Priors
Wan: Open and Advanced Large-Scale Video Generative Models
Enjoy the magic of Diffusion models!
Latent Space Super-Resolution for Higher-Resolution Image Generation with Diffusion Models (CVPR 2025)
[CVPR 2025 Best Paper Award] VGGT: Visual Geometry Grounded Transformer
🌟A curated list of DUSt3R-related papers and resources, tracking recent advancements using this geometric foundation model.
Official implementation of CVPR24 highlight paper "Move as You Say, Interact as You Can: Language-guided Human Motion Generation with Scene Affordance"
Open source impl of **MV-DUSt3R+ Single-Stage Scene Reconstruction from Sparse Views In 2 Seconds** from Meta Reality Labs. Project page https://mv-dust3rp.github.io/
[CVPR 2025 Highlight] GEN3C: 3D-Informed World-Consistent Video Generation with Precise Camera Control
[CVPR 2025] MASt3R-SLAM: Real-Time Dense SLAM with 3D Reconstruction Priors
[CVPR 2025] RelationField: Relate Anything in Radiance Fields