Skip to main content

The Role of Gaze as a Deictic Cue in Human Robot Interaction

  • Conference paper
  • First Online:
Augmented Cognition. Human Cognition and Behavior (HCII 2020)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 12197))

Included in the following conference series:

  • 1502 Accesses

Abstract

Gaze has a major role in social interaction. As a deictic reference, gaze aims at attracting visual attention of a communication partner to a referred entity in the environment. Gaze direction in natural faces is a well-investigated domain of research at behavioral and neurophysiological levels. However, our knowledge about deictic role of gaze in Human Robot Interaction is limited. The present study focuses on a comparative analysis of the deictic role of gaze direction in alternative face morphologies. We report an experimental study that investigated deictic gaze in a virtual reality environment. Human participants identified object locations by utilizing deictic gaze cues provided by avatar faces, as well as natural human faces. Our findings reveal a facilitating role in the accuracy of objection detection in favor of gaze embedded in natural faces compared to gaze embedded in synthetic avatar face morphologies.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Perrett, D., Smith, P., Potter, D., et al.: Visual cells in the temporal cortex sensitive to face view and gaze direction. In: Proceedings of the Royal Society of London, Series B, Biological Sciences, vol. 223, no. 1232, pp. 293–317. Royal Society (1985)

    Google Scholar 

  2. Stiefelhagen, R., Zhu, J.: Head orientation and gaze direction in meetings. In: CHI 2002 Extended Abstracts on Human Factors in Computing Systems (CHI EA 2002), pp. 858–859. Association for Computing Machinery, New York City (2002)

    Google Scholar 

  3. Katzenmaier, M., Stiefelhagen, B., Schultz, T.: Identifying the addressee in human-human-robot interactions based on head pose and speech. In: ICMI 2004 - Sixth International Conference on Multimodal Interfaces, pp. 144–151. Association for Computing Machinery, New York City (2004)

    Google Scholar 

  4. Tan, Z., Thomsen, N.B., Duan, X., et al.: iSocioBot: a multimodal interactive social robot. Int. J. Soc. Robot. 10(1), 5–19 (2018)

    Article  Google Scholar 

  5. Galantucci, B.: An experimental study of the emergence of human communication systems. Cogn. Sci. 29(5), 737–767 (2005)

    Article  Google Scholar 

  6. Sebanz, N., Bekkering, H., Knoblich, G.: Joint action: bodies and minds moving together. Trends Cogn. Sci. 10(2), 70–76 (2006)

    Article  Google Scholar 

  7. Haddadin, S., Croft, E.: Erratum to: Physical human–robot interaction. In: Siciliano, B., Khatib, O. (eds.) Springer Handbook of Robotics, p. E1. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-32552-1_81

    Chapter  Google Scholar 

  8. Fang, R., Doering, M., Chai, J.Y.: Embodied collaborative referring expression generation in situated human-robot interaction. In: Proceedings of the 10th Annual ACM/IEEE International Conference on Human-Robot Interaction, pp. 271–278. Association for Computing Machinery, New York City (2015)

    Google Scholar 

  9. Lemaignan, S., Warnier, M., Sisbot, E.A., et al.: Artificial cognition for social human–robot interaction: an implementation. Artif. Intell. 247(1), 45–69 (2017)

    Article  MathSciNet  Google Scholar 

  10. Piwek, P.: Salience in the generation of multimodal referring acts. In: Proceedings of the 2009 International Conference on Multimodal Interfaces, pp. 207–210 (2009)

    Google Scholar 

  11. Admoni, H., Scassellati, B.: Social eye gaze in human-robot Interaction: a review. J. Hum.-Robot Interact. 6(1), 25–63 (2017)

    Article  Google Scholar 

  12. Ruhland, K., Peters, C.E., Andrist, S., et al.: A review of eye gaze in virtual agents, social robotics and HCI: behaviour generation, user interaction and perception. Comput. Graph. Forum 34(6), 299–326 (2015)

    Article  Google Scholar 

  13. Yücel, Z., Salah, A.A., Meriçli, Ç., et al.: Joint attention by gaze interpolation and saliency. IEEE Trans. Cybern. 43(3), 829–842 (2013)

    Article  Google Scholar 

  14. Breazeal, C., Dautenhahn, K., Kanda, T.: Social robots. In: Siciliano, B., Khatib, O. (eds.), Springer Handbook of Robotics. 2nd edn. pp. 1935–1961. Springer Handbooks, Heidelberg (2016).https://doi.org/10.1007/978-3-319-32552-1

  15. Devault, D., Kariaeva, N., Kothari, A., et al.: An information-state approach to collaborative reference. In: Proceedings of the ACL Interactive Poster and Demonstration Sessions, pp. 1–4 (2005)

    Google Scholar 

  16. Mutlu, B., Shiwa, T., Kanda, T., et al.: Footing in human-robot conversations: how robots might shape participant roles using gaze cues. Hum. Factors 2(1), 61–68 (2009)

    Google Scholar 

  17. Shannon, C.E.: A mathematical theory of communication. Bell Syst. Tech. J. 1(1), 379–423 (1948)

    Article  MathSciNet  Google Scholar 

  18. Holmqvist, K., Andersson, R.: Paradigms - visual search. In: Eye Tracking: A Comprehensive Guide to Methods and Measures, pp. 409–413. Oxford University Press, Oxford (2011)

    Google Scholar 

Download references

Acknowledgments

This project has been supported by TÜBİTAK 117E021 “A gaze-mediated framework for multimodal Human Robot Interaction”.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Cengiz Acartürk .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Yılmaz, E., Fal, M., Acartürk, C. (2020). The Role of Gaze as a Deictic Cue in Human Robot Interaction. In: Schmorrow, D., Fidopiastis, C. (eds) Augmented Cognition. Human Cognition and Behavior. HCII 2020. Lecture Notes in Computer Science(), vol 12197. Springer, Cham. https://doi.org/10.1007/978-3-030-50439-7_32

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-50439-7_32

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-50438-0

  • Online ISBN: 978-3-030-50439-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics