Norkham et al., 2022 - Google Patents
Reinforcement Learning Agent for Path Planning with Expert DemonstrationNorkham et al., 2022
- Document ID
- 16384752627887266531
- Author
- Norkham A
- Chalupa M
- Gardner N
- Khan M
- Zhang X
- Hung C
- Publication year
- Publication venue
- 2022 IEEE 46th Annual Computers, Software, and Applications Conference (COMPSAC)
External Links
Snippet
The problem of path planning is a challenging task for mobile robots. A practical example can be seen in the robots commonly employed in warehouses: they must navigate to pick up goods and move them to certain locations. Therefore, the robot needs a method of moving …
- 230000002787 reinforcement 0 title description 8
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/08—Learning methods
- G06N3/082—Learning methods modifying the architecture, e.g. adding or deleting nodes or connections, pruning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/04—Architectures, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/06—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
- G06N5/04—Inference methods or devices
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
- G06N5/02—Knowledge representation
- G06N5/022—Knowledge engineering, knowledge acquisition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/004—Artificial life, i.e. computers simulating life
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N7/00—Computer systems based on specific mathematical models
- G06N7/005—Probabilistic networks
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course or altitude of land, water, air, or space vehicles, e.g. automatic pilot
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0268—Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course or altitude of land, water, air, or space vehicles, e.g. automatic pilot
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0287—Control of position or course in two dimensions specially adapted to land vehicles involving a plurality of land vehicles, e.g. fleet or convoy travelling
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Jiang et al. | Path planning for intelligent robots based on deep Q-learning with experience replay and heuristic knowledge | |
Liu et al. | A lifelong learning approach to mobile robot navigation | |
Cao et al. | Target search control of AUV in underwater environment with deep reinforcement learning | |
Zhu et al. | Biologically inspired self-organizing map applied to task assignment and path planning of an AUV system | |
Guo et al. | Research progress of path planning methods for autonomous underwater vehicle | |
Xie et al. | Learning with stochastic guidance for robot navigation | |
Jiang et al. | A brief survey: Deep reinforcement learning in mobile robot navigation | |
Guo et al. | A fusion method of local path planning for mobile robots based on LSTM neural network and reinforcement learning | |
CN113139656A (en) | Brain-like speed dual-channel unmanned autonomous decision making method | |
Huang et al. | The USV path planning based on an improved DQN algorithm | |
Zhang et al. | Intelligent vector field histogram based collision avoidance method for auv | |
Sun et al. | Event-triggered reconfigurable reinforcement learning motion-planning approach for mobile robot in unknown dynamic environments | |
Xing et al. | An Algorithm of Complete Coverage Path Planning for Deep‐Sea Mining Vehicle Clusters Based on Reinforcement Learning | |
Sinkar et al. | Multi-agent path finding using dynamic distributed deep learning model | |
Huang et al. | The USV path planning of Dueling DQN algorithm based on tree sampling mechanism | |
Fang et al. | Quadrotor navigation in dynamic environments with deep reinforcement learning | |
Norkham et al. | Reinforcement Learning Agent for Path Planning with Expert Demonstration | |
Khan et al. | Learning sample-efficient target reaching for mobile robots | |
Yue et al. | A new search scheme using multi‐bee‐colony elite learning method for unmanned aerial vehicles in unknown environments | |
Luis et al. | A sample-efficiency comparison between evolutionary algorithms and deep reinforcement learning for path planning in an environmental patrolling mission | |
Chai et al. | Adaptive unknown object rearrangement using low-cost tabletop robot | |
Yan et al. | Autonomous exploration through deep reinforcement learning | |
Kala et al. | Robotic path planning using multi neuron heuristic search | |
Wen et al. | A Hybrid Technique for Active SLAM Based on RPPO Model with Transfer Learning | |
Zhao et al. | 3D Path planning for UAV with improved double deep Q-network |