[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

Norkham et al., 2022 - Google Patents

Reinforcement Learning Agent for Path Planning with Expert Demonstration

Norkham et al., 2022

Document ID
16384752627887266531
Author
Norkham A
Chalupa M
Gardner N
Khan M
Zhang X
Hung C
Publication year
Publication venue
2022 IEEE 46th Annual Computers, Software, and Applications Conference (COMPSAC)

External Links

Snippet

The problem of path planning is a challenging task for mobile robots. A practical example can be seen in the robots commonly employed in warehouses: they must navigate to pick up goods and move them to certain locations. Therefore, the robot needs a method of moving …
Continue reading at ieeexplore.ieee.org (other versions)

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/02Computer systems based on biological models using neural network models
    • G06N3/08Learning methods
    • G06N3/082Learning methods modifying the architecture, e.g. adding or deleting nodes or connections, pruning
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/02Computer systems based on biological models using neural network models
    • G06N3/04Architectures, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/02Computer systems based on biological models using neural network models
    • G06N3/06Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N99/00Subject matter not provided for in other groups of this subclass
    • G06N99/005Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computer systems utilising knowledge based models
    • G06N5/04Inference methods or devices
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computer systems utilising knowledge based models
    • G06N5/02Knowledge representation
    • G06N5/022Knowledge engineering, knowledge acquisition
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/004Artificial life, i.e. computers simulating life
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N7/00Computer systems based on specific mathematical models
    • G06N7/005Probabilistic networks
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course or altitude of land, water, air, or space vehicles, e.g. automatic pilot
    • G05D1/02Control of position or course in two dimensions
    • G05D1/021Control of position or course in two dimensions specially adapted to land vehicles
    • G05D1/0268Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course or altitude of land, water, air, or space vehicles, e.g. automatic pilot
    • G05D1/02Control of position or course in two dimensions
    • G05D1/021Control of position or course in two dimensions specially adapted to land vehicles
    • G05D1/0287Control of position or course in two dimensions specially adapted to land vehicles involving a plurality of land vehicles, e.g. fleet or convoy travelling

Similar Documents

Publication Publication Date Title
Jiang et al. Path planning for intelligent robots based on deep Q-learning with experience replay and heuristic knowledge
Liu et al. A lifelong learning approach to mobile robot navigation
Cao et al. Target search control of AUV in underwater environment with deep reinforcement learning
Zhu et al. Biologically inspired self-organizing map applied to task assignment and path planning of an AUV system
Guo et al. Research progress of path planning methods for autonomous underwater vehicle
Xie et al. Learning with stochastic guidance for robot navigation
Jiang et al. A brief survey: Deep reinforcement learning in mobile robot navigation
Guo et al. A fusion method of local path planning for mobile robots based on LSTM neural network and reinforcement learning
CN113139656A (en) Brain-like speed dual-channel unmanned autonomous decision making method
Huang et al. The USV path planning based on an improved DQN algorithm
Zhang et al. Intelligent vector field histogram based collision avoidance method for auv
Sun et al. Event-triggered reconfigurable reinforcement learning motion-planning approach for mobile robot in unknown dynamic environments
Xing et al. An Algorithm of Complete Coverage Path Planning for Deep‐Sea Mining Vehicle Clusters Based on Reinforcement Learning
Sinkar et al. Multi-agent path finding using dynamic distributed deep learning model
Huang et al. The USV path planning of Dueling DQN algorithm based on tree sampling mechanism
Fang et al. Quadrotor navigation in dynamic environments with deep reinforcement learning
Norkham et al. Reinforcement Learning Agent for Path Planning with Expert Demonstration
Khan et al. Learning sample-efficient target reaching for mobile robots
Yue et al. A new search scheme using multi‐bee‐colony elite learning method for unmanned aerial vehicles in unknown environments
Luis et al. A sample-efficiency comparison between evolutionary algorithms and deep reinforcement learning for path planning in an environmental patrolling mission
Chai et al. Adaptive unknown object rearrangement using low-cost tabletop robot
Yan et al. Autonomous exploration through deep reinforcement learning
Kala et al. Robotic path planning using multi neuron heuristic search
Wen et al. A Hybrid Technique for Active SLAM Based on RPPO Model with Transfer Learning
Zhao et al. 3D Path planning for UAV with improved double deep Q-network