Stowers et al., 2011 - Google Patents
Quadrotor Helicopter Flight Control Using Hough Transform and Depth Map from a Microsoft Kinect Sensor.Stowers et al., 2011
View PDF- Document ID
- 2536077750736007503
- Author
- Stowers J
- Hayes M
- Bainbridge-Smith A
- Publication year
- Publication venue
- MVA
External Links
Snippet
Reliable depth estimation is important to many autonomous robotic systems and visual control algorithms. The Microsoft Kinect is a new, low cost game controller peripheral that calculates a depth map of the environment with good accuracy and high rate. In this paper …
- 238000004805 robotic 0 abstract description 4
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/00221—Acquiring or recognising human faces, facial parts, facial sketches, facial expressions
- G06K9/00268—Feature extraction; Face representation
- G06K9/00281—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/00624—Recognising scenes, i.e. recognition of a whole field of perception; recognising scene-specific objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/00201—Recognising three-dimensional objects, e.g. using range or tactile information
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/36—Image preprocessing, i.e. processing the image information without deciding about the identity of the image
- G06K9/46—Extraction of features or characteristics of the image
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course or altitude of land, water, air, or space vehicles, e.g. automatic pilot
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0231—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11668571B2 (en) | Simultaneous localization and mapping (SLAM) using dual event cameras | |
US12008778B2 (en) | Information processing apparatus, control method for same, non-transitory computer-readable storage medium, and vehicle driving support system | |
TWI827649B (en) | Apparatuses, systems and methods for vslam scale estimation | |
KR102126513B1 (en) | Apparatus and method for determining the pose of the camera | |
Loianno et al. | Cooperative localization and mapping of MAVs using RGB-D sensors | |
Rusu et al. | Laser-based perception for door and handle identification | |
EP3782119B1 (en) | Detection, tracking and 3d modeling of objects with sparse rgb-d slam and interactive perception | |
Shen et al. | Localization through fusion of discrete and continuous epipolar geometry with wheel and IMU odometry | |
Nüchter et al. | A man-portable, IMU-free mobile mapping system | |
Zhang et al. | Building a partial 3D line-based map using a monocular SLAM | |
Zhang et al. | An indoor navigation aid for the visually impaired | |
Stowers et al. | Quadrotor Helicopter Flight Control Using Hough Transform and Depth Map from a Microsoft Kinect Sensor. | |
Hakim et al. | Goal location prediction based on deep learning using RGB-D camera | |
Wietrzykowski et al. | PlaneLoc: Probabilistic global localization in 3-D using local planar features | |
Kakillioglu et al. | 3D sensor-based UAV localization for bridge inspection | |
Krombach et al. | Evaluation of stereo algorithms for obstacle detection with fisheye lenses | |
Bergeon et al. | Low cost 3D mapping for indoor navigation | |
JP2020149186A (en) | Position attitude estimation device, learning device, mobile robot, position attitude estimation method, and learning method | |
Hidalgo | ORBSLAM2 and point cloud processing towards autonomous underwater robot navigation | |
Ribacki et al. | Vision-based global localization using ceiling space density | |
Shi et al. | Vision-based real-time 3D mapping for UAV with laser sensor | |
Peñalver et al. | Multi-view underwater 3D reconstruction using a stripe laser light and an eye-in-hand camera | |
Hernández-García et al. | 3d city models: Mapping approach using lidar technology | |
Yang et al. | Moving target tracking and measurement with a binocular vision system | |
Gallegos et al. | Appearance-based slam relying on a hybrid laser/omnidirectional sensor |