8000 GitHub - PKU-EPIC/GraspVLA: GraspVLA: a Grasping Foundation Model Pre-trained on Billion-scale Synthetic Action Data
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

PKU-EPIC/GraspVLA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 

Repository files navigation

GraspVLA: a Grasping Foundation Model Pre-trained on Billion-scale Synthetic Action Data

arXiv Static Badge

We present a cost-effective pretraining paradigm for VLA models using only synthetic data, achieving direct sim-to-real transfer and strong zero-shot generalizability for robotic grasping. Key contributions include:

  • SynGrasp-1B: a billion-frame synthetic grasping dataset, spanning 240 object categories and 10,000+ objects.

  • GraspVLA: a VLA model pretrained on SynGrasp-1B that achieves zero-shot generalization to real-world grasping without fine-tuning.

  • Unified CoT Framework: GraspVLA integrates autoregressive perception and flow-matching-based action generation into a single reasoning process, enabling joint training on synthetic action data and internet-scale semantic data for open-vocabulary grasping.

teaser

TODO List:

  • Release the supplementary material
  • Release model weights
  • Release SynGrasp-1B dataset

License

About

GraspVLA: a Grasping Foundation Model Pre-trained on Billion-scale Synthetic Action Data

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published
0