Skip to main content
    • Aa
    • Aa

Homing with stereovision

  • Paramesh Nirmal (a1) and Damian M. Lyons (a1)

Visual Homing is a navigation method based on comparing a stored image of a goal location to the current image to determine how to navigate to the goal location. It is theorized that insects such as ants and bees employ visual homing techniques to return to their nest or hive, and inspired by this, several researchers have developed elegant robot visual homing algorithms. Depth information, from visual scale, or other modality such as laser ranging, can improve the quality of homing. While insects are not well equipped for stereovision, stereovision is an effective robot sensor. We describe the challenges involved in using stereovision derived depth in visual homing and our proposed solutions. Our algorithm, Homing with Stereovision (HSV), utilizes a stereo camera mounted on a pan-tilt unit to build composite wide-field stereo images and estimate distance and orientation from the robot to the goal location. HSV is evaluated in a set of 200 indoor trials using two Pioneer 3-AT robots showing it effectively leverages stereo depth information when compared to a depth from scale approach.

Corresponding author
*Corresponding author. E-mail:
Hide All
1. Cartweight B. and Collet T., “Landmark learning in bees,” J. Comparative Physiol. 151, 521543 (1983).
2. Choi D., Shim I., Bok Y., Oh T. and Kweon I., “Autonomous homing based on laser-camera fusion system,” Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, (2012).
3. Churchill D. and Vardy A., “Homing in scale space,” Proceedings of the IEEE/RSJ Conference on Intelligent Robots and Systems (IROS), (2008).
4. Churchill D. and Vardy A., “An orientation invarient visual homing algorithm,” J. Intell. Robot. Syst. 17 (1), 329 (2012).
5. Dudek G. and Jenkin M., Computational Principles of Mobile Robotics (Cambridge University Press, Cambridge, 2000).
6. Feng W., Zhang B., Roning J., Zong X. and Yi T., “Panoramic Stereo Vision,” Proceedings of the SPIE Conference on Intelligent Robotics and Computer Vision XXX: Algorithms and Techniques, Burlingame CA (2013).
7. Fischler M. and Bolles R., “Random sample concensus: A paradigm for model fitting with applications to image abalysis and automated cartography,” Comm. ACM 24 (6), 381395 (1981).
8. Franz M., Scholkopf B., Mallot M. and Bulthoff H., Where did I take that snapshot? Scene-based homing by image matching. Biol. Cybern. (79), 191202 (1998).
9. Jin Y. and Xie M., “Vision guided Homing for Humanoid Service Robot,” Proceedings of the 15th International Conference on Pattern Recognition (ICPR), Vol. 4 (2000).
10. Liu M., Pradalier C., Pomerleau F. and Siegwart R., “The Role of Homing in Visual Topological Navigation,” Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), (2012).
11. Lowe D., “Distinctive image features from scaleinvarient keypoints,” J. Comput. Vis. 60 (2), 91110 (2004).
12. Möller R., “Do insects use templates or parameters for landmark navigation?,” J. Theor. Biol. 210, 3345 (2001).
13. Möller R., “Local visual homing by warping of two-dimensional images,” Robot. Auton. Syst. 31 (1), 87101 (2009).
14. Möller R. and Vardy A., “Local visual homing by matched-filter descent in image databases,” Biol. Cybern. 95, 413430 (2006).
15. Möller R., Lambrinos D., Pfeifer R. and Wehner R., “Insect strategies of visual homing in mobile robots,” Comput. Vis. Mobile Robot. Workshop (1998).
16. Nirmal P. and Lyons D., “Visual Homing with a Pan-Tilt Based stereo camera,” Proceedings of the SPIE Conference on Intelligent Robots and Computer Vision XXX: Algorithms and Techniques, Burlingame CA (2013).
17. Pons J., Huhner W., Dahmen J. and Mallot H., “Vision-Based Robot Homing in Dynamic Environments,” Proceedings of the 13th IASTED International Conference on Robotics and Applications (2007).
18. Srinivasan M., “Insects as gibsonian animals,” Ecol. Psychol. 10 (3–4), 251270 (1998).
19. Sturzl W. and Mallot H., “Vision-Based Homing with a Panoramic Stereovision Sensor,” Proceedings of the British Machine Vision Conference 2002 LNCS 2525 (2002).
20. Thrun S., Burgard W. and Fox D., Probabilistic Robotics, (MIT Press, Cambridge, MA, 2005).
21. Vardy A. and Möller R., “Biologically plausible visual homing methods based on optical flow techniques,” Connect. Sci. 17, 4790 (2005).
22. Vardy A. and Oppacher F., “Low-level visual homing,” Advances in artificial life - Proceedings, 7th European Concerence on Artificial Life (vol. 2801 Lecture Notes in Artificial Intelligence) (2003).
23. Wehner R., “Spatial Vision in Anthropods,” In: Handbook of Sensory Physiology VII/6C, Comparative physiology and evolution of vision in vertebrates (1981).
24. Zeil J., Hoffman H. and Chal J., “Catchment areas of panoramic images in outdoor scenes,” J. Opt. Soc. Am. 20 (3), 450469 (2003).
Recommend this journal

Email your librarian or administrator to recommend adding this journal to your organisation's collection.

  • ISSN: 0263-5747
  • EISSN: 1469-8668
  • URL: /core/journals/robotica
Please enter your name
Please enter a valid email address
Who would you like to send this to? *



Full text views

Total number of HTML views: 0
Total number of PDF views: 44 *
Loading metrics...

Abstract views

Total abstract views: 177 *
Loading metrics...

* Views captured on Cambridge Core between September 2016 - 20th October 2017. This data will be updated every 24 hours.