Abstract
This paper presents a novel approach to localize a robot equipped with an omnidirectional camera within a given 3D map. The pose estimate builds upon the synthesis of panoramic depth images, which are compared to the current view of the camera. We present an algorithmic approach to compute the similarity between these synthetic depth images and visual images, and show how to utilize this image matching for mobile robot navigation tasks, i.e. heading estimation, global localization, and navigation towards a target position. The presented method requires neither additional colour nor laser intensity information in the map. We provide a first evaluation of the involved image processing pipeline and a set of proof-of-concept experiments on a mobile robot. The presented approach supports different use cases like map sharing for heterogeneous robotics teams, or the usage of external sources of 3D maps like extruded floor plans.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
A full description of the robot including hardware and software setup can be found in [3].
References
Möller, R.: A model of ant navigation based on visual prediction. J. Theor. Biol. 305, 118–130 (2012)
Möller, R.: Local visual homing by warping of two-dimensional images. Robot. Auton. Syst. 57(1), 87–101 (2009)
Lange, S., Wunschel, D., Schubert, S., Pfeifer, T., Weissig, P., Uhlig, A., Truschzinski, M., Protzel, P.: Two autonomous robots for the dlr spacebot cup - lessons learned from 60 minutes on the moon. In: Proceedings of ISR 2016: 47th International Symposium on Robotics, pp. 1–8, June 2016
Caron, G., Dame, A., Marchand, E.: Direct model based visual tracking and pose estimation using mutual information. Image Vis. Comput. 32(1), 54–63 (2014)
Caselitz, T., Steder, B., Ruhnke, M., Burgard, W.: Monocular camera localization in 3D lidar maps. In: 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 1926–1931, October 2016
Caselitz, T., Steder, B., Ruhnke, M., Burgard, W.: Matching geometry for long-term monocular camera localization. In: ICRA Workshop: AI for Long-Term Autonomy (2016)
Forster, C., Pizzoli, M., Scaramuzza, D.: Air-ground localization and map augmentation using monocular dense reconstruction. In: 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 3971–3978, November 2013
Gawel, A., Cieslewski, T., Dubé, R., Bosse, M., Siegwart, R., Nieto, J.: Structure-based vision-laser matching. In: 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 182–188, October 2016
Pascoe, G., Maddern, W., Newman, P.: Direct visual localisation and calibration for road vehicles in changing city environments. In: 2015 IEEE International Conference on Computer Vision Workshop (ICCVW), pp. 98–105, December 2015
Pascoe, G., Maddern, W., Stewart, A.D., Newman, P.: Farlap: fast robust localisation using appearance priors. In: 2015 IEEE International Conference on Robotics and Automation (ICRA), pp. 6366–6373, May 2015
Stewart, A.D., Newman, P.: Laps - localisation using appearance of prior structure: 6-dof monocular camera localisation using prior pointclouds. In: IEEE International Conference on Robotics and Automation, pp. 2625–2632, May 2012
Wolcott, R.W., Eustice, R.M.: Visual localization within lidar maps for automated urban driving. In: 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 176–183, September 2014
Napier, A., Corke, P., Newman, P.: Cross-calibration of push-broom 2d lidars and cameras in natural scenes. In: 2013 IEEE International Conference on Robotics and Automation, pp. 3679–3684, May 2013
Pandey, G., McBride, J.R., Savarese, S., Eustice, R.M.: Automatic extrinsic calibration of vision and lidar by maximizing mutual information. J. Field Robot. 32(5), 696–722 (2015)
Mur-Artal, R., Montiel, J.M.M., Tardós, J.D.: ORB-SLAM: a versatile and accurate monocular SLAM system. IEEE Trans. Robot. 31(5), 1147–1163 (2015)
Schubert, S., Neubert, P., Protzel, P.: How to build and customize a high-resolution 3D laserscanner using off-the-shelf components. In: Alboul, L., Damian, D., Aitken, J.M.M. (eds.) TAROS 2016. LNCS, vol. 9716, pp. 314–326. Springer, Cham (2016). doi:10.1007/978-3-319-40379-3_33
Scaramuzza, D., Martinelli, A., Siegwart, R.: A toolbox for easily calibrating omnidirectional cameras. In: 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 5695–5701, October 2006
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Schubert, S., Neubert, P., Protzel, P. (2017). Towards Camera Based Navigation in 3D Maps by Synthesizing Depth Images. In: Gao, Y., Fallah, S., Jin, Y., Lekakou, C. (eds) Towards Autonomous Robotic Systems. TAROS 2017. Lecture Notes in Computer Science(), vol 10454. Springer, Cham. https://doi.org/10.1007/978-3-319-64107-2_49
Download citation
DOI: https://doi.org/10.1007/978-3-319-64107-2_49
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-64106-5
Online ISBN: 978-3-319-64107-2
eBook Packages: Computer ScienceComputer Science (R0)