Skip to main content

Introduction

  • Chapter
  • First Online:
Visual Perception for Humanoid Robots

Part of the book series: Cognitive Systems Monographs ((COSMOS,volume 38))

  • 454 Accesses

Abstract

The principal research objective of this book is to discuss, understand and implement the essential environmental vision-model coupling for humanoid robots. Concretely, the coupling of model-based visual environmental object recognition with model-based visual global self-localization and its interrelated data handling are the ultimate goals to be achieved. These skills require novel methods to overcome the current sensing and matching limitations. The proposed algorithms and their efficient implementation in a platform independent software system have enabled humanoid robots to attain the bidirectional vision-model coupling. The developed methods apply active stereo vision and use geometric CAD models of real made-for-humans environments. The core applications are tasks involving self-localization and object recognition for grasping and manipulation.

A journey of a thousand miles begins with a single step

Lao Tzu

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 109.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    The use of multiple heterogeneous input and output channels in a system.

  2. 2.

    The pose of an object in 3D space is uniquely defined by a 3D vector location and three orientation angles.

  3. 3.

    A lambertian surface is characterized by its isotropic luminance. This restriction is partially relaxed by metal and glass elements in the environment (see Chap. 5).

  4. 4.

    The collaborative research center 588 “Humanoid Robots - Learning and Cooperating Multimodal Robots” founded by the Deutsche Forschungs-Gemeinschaft DFG.

  5. 5.

    Off-the-shelf cameras providing range-quantized (8-bit), low-dynamic-range (LDR) based on color filter array CFA (Bayer pattern) sensors, such as the IEEE-1394a DragonFly cameras [9] in the humanoid robots ARMAR-IIIa and ARMAR-IIIb.

  6. 6.

    Due to the raw images nature, the density property of real numbers is not hold.

References

  1. Baxter, P., and W. Browne. 2009. Perspectives on Robotic Embodiment from a Developmental Cognitive Architecture. In International Conference on Adaptive and Intelligent Systems, 3–8.

    Google Scholar 

  2. Brooks, R., and L. Stein. 1994. Building Brains for Bodies. Autonomous Robots 1: 7–25.

    Google Scholar 

  3. Patnaik, S. 2007. Robot Cognition and Navigation. Berlin: Springer. ISBN 978-3-540-68916-4.

    Google Scholar 

  4. Sinapov, J., T. Bergquist, C. Schenck, U. Ohiri, S. Griffith, and A. Stoytchev. 2011. Interactive Object Recognition using Proprioceptive and Auditory Feedback. International Journal of Robotics Research 30(10): 1250–1262. ISBN 978-0387853482.

    Google Scholar 

  5. Bar-Cohen, A.M.Y., and D. Hanson. 2009. The Coming Robot Revolution: Expectations and Fears about Emerging Intelligent Humanlike Machines. Berlin: Springer. ISBN 978-0387853482.

    Google Scholar 

  6. Markovich, S. 2002. Amodal Completion in Visual Perception. In Visual Mathematics. vol. 4, ed. Slavik Jablan. ISBN 978-3-642-34273-8.

    Google Scholar 

  7. Asfour, T., K. Regenstein, P. Azad, J. Schröder, A. Bierbaum, N. Vahrenkamp, and R. Dillmann. 2006. ARMAR-III: An Integrated Humanoid Platform for Sensory-Motor Control. In IEEE-RAS International Conference on Humanoid Robots, 169–175.

    Google Scholar 

  8. Asfour, T., K. Welke, P. Azad, A. Ude, and R. Dillmann. 2008. The Karlsruhe Humanoid Head. In IEEE-RAS International Conference on Humanoid Robots, 447–453.

    Google Scholar 

  9. PTGrey. 2008. Dragonfly Technical Reference Manual, 5-8 2008.

    Google Scholar 

  10. Gonzalez-Aguirre, D., T. Asfour, and R. Dillmann. 2011. Robust Image Acquisition for Vision-Model Coupling by Humanoid Robots. In IAPR-Conference on Machine Vision Applications, 557–561.

    Google Scholar 

  11. Gonzalez-Aguirre, D., T. Asfour, and R. Dillmann. 2010. Eccentricity Edge-Graphs from HDR Images for Object Recognition by Humanoid Robots. In IEEE-RAS International Conference on Humanoid Robots, 144 –151.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to David Israel González Aguirre .

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

González Aguirre, D.I. (2019). Introduction. In: Visual Perception for Humanoid Robots. Cognitive Systems Monographs, vol 38. Springer, Cham. https://doi.org/10.1007/978-3-319-97841-3_1

Download citation

Publish with us

Policies and ethics