Hundt et al., 2020 - Google Patents
“good robot!”: Efficient reinforcement learning for multi-step visual tasks with sim to real transferHundt et al., 2020
View PDF- Document ID
- 9549600562725600363
- Author
- Hundt A
- Killeen B
- Greene N
- Wu H
- Kwon H
- Paxton C
- Hager G
- Publication year
- Publication venue
- IEEE Robotics and Automation Letters
External Links
Snippet
Current Reinforcement Learning (RL) algorithms struggle with long-horizon tasks where time can be wasted exploring dead ends and task progress may be easily reversed. We develop the SPOT framework, which explores within action safety zones, learns about …
- 230000002787 reinforcement 0 title abstract description 15
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/62—Methods or arrangements for recognition using electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Hundt et al. | “good robot!”: Efficient reinforcement learning for multi-step visual tasks with sim to real transfer | |
Yang et al. | A deep learning approach to grasping the invisible | |
Ibarz et al. | How to train your robot with deep reinforcement learning: lessons we have learned | |
Sadeghi et al. | Sim2real viewpoint invariant visual servoing by recurrent control | |
Stüber et al. | Let's push things forward: A survey on robot pushing | |
Xu et al. | Efficient learning of goal-oriented push-grasping synergy in clutter | |
Robinson et al. | Robotic vision for human-robot interaction and collaboration: A survey and systematic review | |
Rajvanshi et al. | Saynav: Grounding large language models for dynamic planning to navigation in new environments | |
Ivaldi et al. | Learning to recognize objects through curiosity-driven manipulation with the iCub humanoid robot | |
Karamcheti et al. | Learning visually guided latent actions for assistive teleoperation | |
Wang et al. | Generalization through hand-eye coordination: An action space for learning spatially-invariant visuomotor control | |
Bharadhwaj et al. | Track2Act: Predicting Point Tracks from Internet Videos enables Generalizable Robot Manipulation | |
Akinola et al. | Visionary: Vision architecture discovery for robot learning | |
Hundt et al. | The costar block stacking dataset: Learning with workspace constraints | |
Deng et al. | A learning framework for semantic reach-to-grasp tasks integrating machine learning and optimization | |
Rohrer | BECCA: Reintegrating AI for natural world interaction | |
Ma et al. | A learning from demonstration framework for adaptive task and motion planning in varying package-to-order scenarios | |
Ziaeetabar et al. | Prediction of manipulation action classes using semantic spatial reasoning | |
Ding et al. | Preafford: Universal affordance-based pre-grasping for diverse objects and environments | |
Nguyen et al. | Lightweight language-driven grasp detection using conditional consistency model | |
Liao et al. | Human hand motion prediction in disassembly operations | |
Lin et al. | Action recognition for human-marionette interaction | |
Aslan et al. | End-to-end learning from demonstation for object manipulation of robotis-Op3 humanoid robot | |
Hundt | Effective Visual Robot Learning: Reduce, Reuse, Recycle | |
Hu et al. | Learning from visual demonstrations via replayed task-contrastive model-agnostic meta-learning |