Abstract
3D data have been used for robotics tasks in the last years. These data provide valuable information about the robot environment. Traditionally, stereo cameras has been used to obtain 3D data, but these kind of cameras do not provide information in the lack of texture. There is a new camera, SR4000, which uses infrared light in order to get richer information. In this paper we first analyze this camera. Then, we detail an efficient ICP-like method to build complete 3D models combining Growing Neural Gas (GNG) and visual features. First, we adapt the GNG to the 3D cloud points. Then, we propose the calculation of visual features and its registration to the elements of the GNG. Finally, we use correspondences between frames, an ICP-like method to calculate egomotion. Results of mapping from the egomotion are shown.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Besl, P.J., McKay, N.D.: A method for registration of 3-d shapes. IEEE Trans. on Pattern Analysis and Machine Intelligence 14(2), 239–256 (1992)
Cazorla, M., Viejo, D., Hernandez, A., Nieto, J., Nebot, E.: Large scale egomotion and error analysis with visual features. Journal of Physical Agents 4, 19–24 (2010)
Chen, Y., Medioni, G.: Object modeling by registration of multiple range images. In: Medioni, G. (ed.) Proceedings of IEEE International Conference on Robotics and Automation, vol. 3, pp. 2724–2729 (1991)
Chiabrando, F., Chiabrando, R., Piatti, D., Rinaudo, F.: Sensors for 3d imaging: Metric evaluation and calibration of a ccd/cmos time-of-flight camera. Sensors 9(12), 10080–10096 (2009)
Dissanayake, M.W.M.G., Newman, P., Clark, S., Durrant-Whyte, H.F., Csorba, M.: A solution to the simultaneous localization and map building (slam) problem. IEEE Transactions on Robotics and Automation 17(3), 229–241 (2001)
Fischler, M.A., Bolles, R.C.: Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 24(6), 381–395 (1981)
Fritzke, B.: A Growing Neural Gas Network Learns Topologies, vol. 7, pp. 625–632. MIT Press, Cambridge (1995)
Burak Gokturk, S., Yalcin, H., Bamji, C.: A time-of-flight depth sensor - system description, issues and solutions. In: CVPRW 2004: Proceedings of the 2004 Conference on Computer Vision and Pattern Recognition Workshop (CVPRW 2004), vol. 3, p. 35. IEEE Computer Society, Washington, DC, USA (2004)
David, G.L.: Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision 60, 91–110 (2004)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2011 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Viejo, D., Garcia, J., Cazorla, M. (2011). Visual Features Extraction Based Egomotion Calculation from a Infrared Time-of-Flight Camera. In: Cabestany, J., Rojas, I., Joya, G. (eds) Advances in Computational Intelligence. IWANN 2011. Lecture Notes in Computer Science, vol 6692. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-21498-1_2
Download citation
DOI: https://doi.org/10.1007/978-3-642-21498-1_2
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-21497-4
Online ISBN: 978-3-642-21498-1
eBook Packages: Computer ScienceComputer Science (R0)