Abstract
This paper investigates the integration of multiple time-of-flight (ToF) depth sensors for the purposes of general 3D tracking and specifically of the hands. The advantage of using a network with multiple sensors is in the increased viewing coverage as well as being able to capture a more complete 3D point cloud representation of the object. Given an ideal point cloud representation, tracking can be accomplished without having to first reconstruct a mesh representation of the object. In utilizing a network of depth sensors, calibration between the sensors and the subsequent data alignment of the point clouds poses key challenges. While there has been research on the merging and alignment of scenes with larger objects such as the human body, there is little research available focusing on a smaller and more complicated object such as the human hand. This paper presents a study on ways to merge and align the point clouds from a network of sensors for object and feature tracking from the combined point clouds.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Oikonomidis, I., Kyriazis, N., Argyros, A.A.: Efficient model-based 3d tracking of hand articulations using kinect. In: Proceedings of the 2011 British Machine Vision Conference, pp. 101.1–101.11 (2011)
Kıraç, F., Kara, Y.E., Akarun, L.: Hierarchically constrained 3D hand pose estimation using regression forests from single frame depth data. Pattern Recogn. Lett. Spec. Issue Depth Image Anal. (2013)
Baak, A., Müller, M., Bharaj, G., Seidel, H.-P., Theobalt, C.: A data-driven approach for real-time full body pose reconstruction from a depth camera. In: Proceedings of the 2011 IEEE International Conference on Computer Vision, pp. 1092–1099 (2011)
Bay, H., Ess, A., Tuytelaars, T., Van Gool, L.: SURF: speeded up robust features. Comput. Vis. Image Underst. (CVIU) 110(3), 346–359 (2008)
Rosten, E., Drummond, T.: Machine learning for high-speed corner detection. In: Proceedings of the 2005 IEEE International Conference on Computer Vision, pp. 1508–1515 (2010)
Newcombe, R., Izadi, S., Hilliges, O., Molyneaux, D., Kim, D., Davison, A., Kohli, P., Shotton, J., Hodges, S., Fitzgibbon, A.: KinectFusion: real-time dense surface mapping and tracking. In: Proceedings of the 2011 International Symposium on Mixed and Augmented Reality, pp. 127–136 (2011)
Newcombe, R., Davison, A.J.: Live dense reconstruction with a single moving camera. In: Proceedings of the 2010 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1498–1505 (2010)
Tanskanen, P., Kolev, K., Meier, L., Camposeco, F., Saurer, O., Pollefeys, M.: Live metric 3D reconstruction on mobile phones. In: Proceedings of 2013 IEE International Conference on Computer Vision, pp. 65–72 (2013)
Marton, Z., Rusu, R., Beetz, M.: On fast surface reconstruction methods for large and noisy point clouds. In: Proceedings of 2009 IEEE International Conference on Robotics and Automation, pp. 3218–3233 (2009)
SoftKinectic DepthSense Cameras. http://www.softkinetic.com/en-us/products/depthsensecameras.aspx
Berger, K., Ruhl, K., Brümmer, C., Schröder, Y., Scholz, A., Magnor, M.: Markerless motion capture using multiple color-depth sensors. In: Proceedings of the 2011 Vision, Modeling and Visualization, pp. 317–324 (2011)
Xu, G., Payandeh, S.: Sensitivity study for object reconstruction using a network of time-of-flight depth sensors. In: Proceedings of the 2015 IEEE International Conference on Robotics and Automation (2015)
Kim, Y., Theobalt, C., Diebel, J., Kosecka, J., Miscusik, B., Thrun, S.: Multi-view Image and ToF sensor fusion for dense 3D reconstruction. In: Proceedings of the 2009 IEEE 12th International Conference on Computer Vision Workshops, pp. 1542–1549 (2009)
Zhang, Z.: A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 22(11), 1330–1334 (2000)
Fischler, M., Bolles, R.: Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 24(6), 381–395 (1981)
SoftKinetic DS325 Datasheet. http://www.softkinetic.com/Portals/0/Documents/PDF/WEB_20130527_SK_DS325_Datasheet_V4.0.pdf
Agarwal, A., Jawahar, C.V., Narayanan, P.J.: A survey of planar homography estimation techniques. Technical Reports: International Institute of Information Technology Hyderabad (2005)
Kim, Y., Chan, D., Theobalt, C., Thrun, S.: Design and calibration of a multi-view TOF sensor fusion system. In: Proceedings of the 2009 IEEE Computer Vision and Pattern Recognition Workshops, pp. 1–7, June 2009
Dhawan, A., Honrao, V.: Implementation of hand detection based techniques for human computer interaction. Int. J. Comput. Appl. 72(17), June 2013
Rusu, R., Cousins, S.: 3D is here: point cloud library (PCL). In: Proceedings of the 2011 IEEE International Conference on Robotics and Automation, May 2011
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Xu, G., Payandeh, S. (2015). On Calibration and Alignment of Point Clouds in a Network of RGB-D Sensors for Tracking. In: Bebis, G., et al. Advances in Visual Computing. ISVC 2015. Lecture Notes in Computer Science(), vol 9475. Springer, Cham. https://doi.org/10.1007/978-3-319-27863-6_52
Download citation
DOI: https://doi.org/10.1007/978-3-319-27863-6_52
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-27862-9
Online ISBN: 978-3-319-27863-6
eBook Packages: Computer ScienceComputer Science (R0)