Abstract
This paper proposes a night time call system using a wearable camera for patients. The proposed system consists of a wearable camera, computer, relay controller, and nurse call. The user wears the wearable camera. All captured eye images are fed to the convolutional neural network to detect the pupil center. When the detected pupil center exceeds a preset threshold value, the computer sends the signal to operate the nurse call via each relay controller. Two experiments were conducted to evaluate the proposed system: verification of the accuracy of pupil center detection and quantitative evaluation of call success. In the former experiment, we collected 2,800 eye images from seven people and conducted a pupil center detection experiment on several training conditions. As a result, an average error of 1.17 pixels was obtained. In the latter experiment, a call experiment was conducted on five healthy people. The experiment time for each subject was about five minutes. The subject experimented while lying on the bed. Twelve audio stimuli were given in one experiment, after getting the stimuli, the subject moved his eye. The correct call in response to the audio stimulus was considered successful, and the precision, recall, and F-measure were calculated. As a result, we obtained the precision, recall, and F-measure of 0.83, 1.00, and 0.91, respectively. These experimental results show the effectiveness of the proposed system.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
eyeSwitch. http://www.emfasys.co.jp/index8f.html
Gazo GPE3. https://www.gazo.co.jp/gaze_point_estimation
Tobii Pro Fusion. https://www.tobiipro.com/product-listing/fusion/
Tobii Pro Glasses 3. https://www.tobiipro.com/product-listing/tobii-pro-glasses-3/
Tobii Pro Nano. https://www.tobiipro.com/product-listing/nano/
Chinsatit, W., Saitoh, T.: CNN-based pupil center detection for wearable gaze estimation system. Applied Computational Intelligence and Soft Computin 2017 (2017). https://doi.org/10.1155/2017/8718956
Gou, C., Zhang, H., Wang, K., Wang, F.Y., Ji, Q.: Cascade learning from adversarial synthetic images for accurate pupil detection. Pattern Recogn. 88, 584–594 (2019). https://doi.org/10.1016/j.patcog.2018.12.014
Klemets, J., Toussaint, P.: Does revealing contextual knowledge of the patient fs intention help nurses’ handling of nurse calls? Int. J. Med. Inform. 86, 1–9 (2016). https://doi.org/10.1016/j.ijmedinf.2015.11.010
Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. In: 26th Annual Conference on Neural Information Processing Systems (NIPS2012), pp. 1097–1105 (2012)
Ohya, T., Morita, K., Yamashita, Y., Egami, C., Ishii, Y., Nagamitsu, S., Matsuishi, T.: Impaired exploratory eye movements in children with Asperger fs syndrome. Brain Dev. 36(3), 241–247 (2014). https://doi.org/10.1016/j.braindev.2013.04.005
Ongenae, F., Claeys, M., Kerckhove, W., Dupont, T., Verhoeve, P., Turck, F.: A self-learning nurse call system. Comput. Biol. Med. 44, 110–123 (2014). https://doi.org/10.1016/j.compbiomed.2013.10.014
Pfister, T., Simonyan, K., Charles, J., Zisserman, A.: Deep convolutional neural networks for efficient pose estimation in gesture videos. In: Cremers, D., Reid, I., Saito, H., Yang, M.-H. (eds.) ACCV 2014. LNCS, vol. 9003, pp. 538–552. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-16865-4_35
Unluturk, M.S., Ozcanhan, M.H., Dalkilic, G.: Improving communication among nurses and patients. Comput. Methods Programs Biomed. 120(2), 102–12 (2015). https://doi.org/10.1016/j.cmpb.2015.04.004
Wang, C.C., Hung, J.C.: Comparative analysis of advertising attention to Facebook social network: evidence from eye-movement data. Comput. Hum. Behav. 100, 192–208 (2019). https://doi.org/10.1016/j.chb.2018.08.007
Yiu, Y.H., et al.: DeepVOG: open-source pupil segmentation and gaze estimation in neuroscience using deep learning. J. Neurosci. Methods 324, 108307 (2019). https://doi.org/10.1016/j.jneumeth.2019.05.016
Acknowledgment
This work was supported by JSPS KAKENHI Grant Numbers 19KT0029.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Sakamoto, K., Saitoh, T., Itoh, K. (2020). Development of Night Time Calling System by Eye Movement Using Wearable Camera. In: Stephanidis, C., Antona, M., Gao, Q., Zhou, J. (eds) HCI International 2020 – Late Breaking Papers: Universal Access and Inclusive Design. HCII 2020. Lecture Notes in Computer Science(), vol 12426. Springer, Cham. https://doi.org/10.1007/978-3-030-60149-2_27
Download citation
DOI: https://doi.org/10.1007/978-3-030-60149-2_27
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-60148-5
Online ISBN: 978-3-030-60149-2
eBook Packages: Computer ScienceComputer Science (R0)