Skip to main content

Gaze Control of Humanoid Robot for Learning from Demonstration

  • Conference paper
  • First Online:
Robot Intelligence Technology and Applications 4

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 447))

Abstract

Robots can learn knowledge by observing demonstration of humans. As tutees, robots need to not only observe human behaviors, but also make proper feedbacks for human tutors because learning is an interactive process in which information is delivered in bidirectional ways between humans and robots. Gaze is an adequate method for robots to provide human tutors with feedbacks that robots are concentrating on current learning because gaze directly represents where they are paying attention to. This paper proposes a gaze control algorithm with a state machine in learning from demonstration. A human tutor shows demonstration in front of a robot tutee, and the robot tutee observes the demonstration for learning. The robot tutee perceives external environment through its camera, recognizes a human and objects, and figures out a state at which the robot tutee is situated. Then, the robot tutee gazes at proper targets that are predefined by the state machine. The human tutor also adjusts the demonstration to make learning more effectively according to the robot tuteeā€™s feedbacks. The effectiveness of the proposed method is demonstrated through the experiments with a robotic head with 17 degrees of freedom, developed in the RIT Lab., KAIST.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Kim J-.H., Choi S-.H., Park I-.W., Zaheer S.A.: Intelligence technology for robots that think. IEEE Comput. Intell. Mag. 8(3), 70ā€“84 (2013)

    Google ScholarĀ 

  2. Breazeal, C., Scassellati, B.: Robots that Imitate Humans. Trends Cogn. Sci. 6(11), 481ā€“487 (2002)

    ArticleĀ  Google ScholarĀ 

  3. Thomaz, A.L., Cakmak, M.: Learning about objects with human teachers. In: ACM/IEEE International Conference on Human Robot Interaction, LA Jolla, CA, USA, pp. 15ā€“22 (2009)

    Google ScholarĀ 

  4. Pitsch, K., Volmer, A.L., Fritsch, J., Wrede, B., Rohlfing, K., Sagerer, G.: On the loop of action modification and the recipientā€™s gaze in adult-child-interaction. Gesture and Speech in Interaction, Poznan, Poland, pp. 1ā€“17 (2009)

    Google ScholarĀ 

  5. Huang C.M., Thomaz A.L.: Effects of responding to, initiating and ensuring joint attention in human-robot interaction. In: IEEE International Sympoisum on Robots and Human Interactive Communications, Atlanta, GA, USA, pp. 65ā€“71 (2011)

    Google ScholarĀ 

  6. Mora A., Gals D.F., Kanda T., Hagita N.: A teleoperation approach for mobile social robots incorporating automatic gaze control and three-dimensional spatial visualization. IEEE Trans. Syst. Man Cybern. 43(3), 630ā€“642 (2013)

    Google ScholarĀ 

  7. Flanagan, J.R., Johansson, R.S.: Action plans used in action observation. Nature 424(6950), 769ā€“771 (2003)

    ArticleĀ  Google ScholarĀ 

  8. Tatler, B.W., Hayhoe, M.M., Land, M.F., Ballard, D.H.: Eye guidance in natural vision: reinterpreting salience. J. Vis. 11(5), 1ā€“23 (2011)

    Google ScholarĀ 

  9. Senju, A., Csibra, G.: Gaze following in human infants depends on communicative signal. Curr. Bio. 18(9), 668ā€“671 (2008)

    ArticleĀ  Google ScholarĀ 

  10. Cerf, M., Harel, J., EinhƤuser W., Koch, C.: Predicting human gaze using low-level saliency combined with face detection. Adv. Neur. In. 241ā€“248 (2008)

    Google ScholarĀ 

  11. LeCun, Y., Bottou, L., Bengio, Y., Haffer, P.: Gradient-Based learning applied to document recognition. Proc. IEEE 86(11), 2278ā€“2324 (1998)

    ArticleĀ  Google ScholarĀ 

  12. Yoo, B.-S., Kim, J.-H.: Fuzzy integral-based gaze control of a robotic head for human robot interaction. IEEE Trans. Cybern. 45(9), 1769ā€“1783 (2015)

    ArticleĀ  Google ScholarĀ 

Download references

Acknowledgments

This work was partly supported by Institute for Information communications Technology Promotion (IITP) grant funded by the Korea government (MSIP) (No. B0101-15-0551, Virtual Creatures with Digital DNA), the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIP) (No. NRF-2014R1A2A1A10051551), and the Technology Innovation Program, 10045252, Development of robot task intelligence technology, supported by the Ministry of Trade, Industry, and Energy (MOTIE, Korea).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Bum-Soo Yoo .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

Ā© 2017 Springer International Publishing Switzerland

About this paper

Cite this paper

Yoo, BS., Kim, JH. (2017). Gaze Control of Humanoid Robot for Learning from Demonstration. In: Kim, JH., Karray, F., Jo, J., Sincak, P., Myung, H. (eds) Robot Intelligence Technology and Applications 4. Advances in Intelligent Systems and Computing, vol 447. Springer, Cham. https://doi.org/10.1007/978-3-319-31293-4_21

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-31293-4_21

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-31291-0

  • Online ISBN: 978-3-319-31293-4

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics