Abstract
Accessible Digital Musical instruments (ADMIs) dedicated to people with motor disabilities represent a relevant niche in accessibility research. The designer is often required to exploit unconventional physical interaction channels, different from hands and fingers. Although comprehensive evaluation methods for Digital Musical Instruments in general are found in literature, little has been done both in ADMIs evaluation and the analysis of suitable interaction channels from a Human-Computer Interaction perspective. In this work the performance of breath, gaze pointing and head movements is analyzed, in terms of movement speed and stability, through a simple experiment. These interaction channels could be exploited in the design of ADMIs dedicated to quadriplegic musicians. The proposed experiment has similarities with past Fitts Law evaluation tests. Results are discussed proposing possible mappings between channels and musical performance parameters. These results could also be useful to inform the design of different interface types.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Source code available under a GNU-GPLv3 license at https://github.com/Neeqstock/HanDMIs-TestSuite.
References
Bååth, R., Strandberg, T., Balkenius, C.: Eye tapping: how to beat out an accurate rhythm using eye movements. In: Proceedings of 11th International Conference on New Interfaces for Musical Expression (NIME 2011), NIME 2011, Oslo (May 2011)
Cross, I.: The nature of music and its evolution. In: Hallam, S., Cross, I., Taut, M. (eds.) The Oxford Handbook of Music Psychology, pp. 3–13. Oxford University Press, Oxford (2009)
Davanzo, N., Dondi, P., Mosconi, M., Porta, M.: Playing music with the eyes through an isomorphic interface. In: Proceedings of the Workshop on Communication by Gaze Interaction - COGAIN 2018, pp. 1–5. ACM Press, Warsaw (2018)
Frid, E.: Accessible digital musical instruments—a review of musical interfaces in inclusive music practice. Multimodal Technol. Interact. 3(3), 57 (2019)
Hornof, A.J.: The prospects for eye-controlled musical performance. In: Proceedings 14th International Conference on New Interfaces for Musical Expression (NIME 2014), NIME 2014, Goldsmiths, University of London (July 2014)
Housemate: Magic Flute (nd). http://housemate.ie/magic-flute/. Accessed 8 June 2019
Hunt, A., Wanderley, M.M., Paradis, M.: The importance of parameter mapping in electronic instrument design. J. New Music Res. 32(4), 429–440 (2003)
Jamboxx: Jamboxx (nd). https://www.jamboxx.com/. Accessed 8 June 2019
Jones, M., Grogg, K., Anschutz, J., Fierman, R.: A sip-and-puff wireless remote control for the Apple iPod. Assist. Technol. 20(2), 107–110 (2008)
Lausted, C.G., Johnson, A.T., Scott, W.H., Johnson, M.M., Coyne, K.M., Coursey, D.C.: Maximum static inspiratory and expiratory pressures with different lung volumes. BioMed. Eng. OnLine 5, 29 (2006)
Lubet, A.: Music, Disability, and Society. Temple University Press, Philadelphia (2011)
McKenzie, S.: Fitts’ law. In: Norman, K.L. Kirakowski, J. (eds.) Handbook of Human-Computer Interaction, vol. 1, pp. 349–370. Hoboken (2018)
Mougharbel, I., El-Hajj, R., Ghamlouch, H., Monacelli, E.: Comparative study on different adaptation approaches concerning a sip and puff controller for a powered wheelchair. In: Proceedings of the 2013 Science and Information Conference, pp. 597–603. London (October 2013)
O’Modhrain, S.: A framework for the evaluation of digital musical instruments. Comput. Music J. 35(1), 28–42 (2011)
Ooms, K., Dupont, L., Lapon, L., Popelka, S.: Accuracy and precision of fixation locations recorded with the low-cost eye tribe tracker in different experimental setups. J. Eye Mov. Res. 8(1) (2015)
Sarig-Bahat, H.: Evidence for exercise therapy in mechanical neck disorders. Man. Ther. 8(1), 10–20 (2003)
Shima, K., Tsuji, T., Kandori, A., Yokoe, M., Sakoda, S.: Measurement and evaluation of finger tapping movements using log-linearized Gaussian mixture networks. Sensors 9(3), 2187–2201 (2009)
Sibert, L.E., Jacob, R.J.K.: Evaluation of eye gaze interaction. In: Proceedings of SIGCHI Conference on Human Factors in Computing Systems, CHI 2000, pp. 281–288. Association for Computing Machinery, The Hague (April 2000)
Vamvakousis, Z., Ramirez, R.: The EyeHarp: a gaze-controlled digital musical instrument. Front. Psychol. 7, article 906 (2016)
Wang, F., Ren, X.: Empirical evaluation for finger input properties in multi-touch interaction. In: Proceedings of 27th International Conference on Human Factors in Computing Systems (CHI 2009), pp. 1063–1072. ACM Press, Boston (2009)
Williams, K.E., Berthelsen, D., Nicholson, J.M., Walker, S., Abad, V.: The effectiveness of a short-term group music therapy intervention for parents who have a child with a disability. J. Music Ther. 49(1), 23–44 (2012)
Acknowledgements
We would like to offer our special thanks to Daniele Zago (from University of Padova, Italy) for advising us on statistical analysis, and to Mattia Galante (from University of Milan, Italy) for the great help given by assisting the testers.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Davanzo, N., Avanzini, F. (2020). Experimental Evaluation of Three Interaction Channels for Accessible Digital Musical Instruments. In: Miesenberger, K., Manduchi, R., Covarrubias Rodriguez, M., Peňáz, P. (eds) Computers Helping People with Special Needs. ICCHP 2020. Lecture Notes in Computer Science(), vol 12377. Springer, Cham. https://doi.org/10.1007/978-3-030-58805-2_52
Download citation
DOI: https://doi.org/10.1007/978-3-030-58805-2_52
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-58804-5
Online ISBN: 978-3-030-58805-2
eBook Packages: Computer ScienceComputer Science (R0)