UCL Discovery
UCL home » Library Services » Electronic resources » UCL Discovery

New Interfaces for Classifying Performance Gestures in Music

Rhodes, Chris; Allmendinger, Richard; Climent, Ricardo; (2019) New Interfaces for Classifying Performance Gestures in Music. In: Yin, H and Camacho, D and Tino, P and Tallón-Ballesteros, AJ and Menezes, R and Allmendinger, R, (eds.) Intelligent Data Engineering and Automated Learning – IDEAL 2019. (pp. pp. 31-42). Springer, Cham: Cham, Switzerland. Green open access

[thumbnail of Rhodes_New Interfaces for Classifying Performance Gestures in Music.pdf]
Preview
Text
Rhodes_New Interfaces for Classifying Performance Gestures in Music.pdf

Download (756kB) | Preview

Abstract

Interactive machine learning (ML) allows a music performer to digitally represent musical actions (via gestural interfaces) and affect their musical output in real-time. Processing musical actions (termed performance gestures) with ML is useful because it predicts and maps often-complex biometric data. ML models can therefore be used to create novel interactions with musical systems, game-engines, and networked analogue devices. Wekinator is a free open-source software for ML (based on the Waikato Environment for Knowledge Analysis – WEKA - framework) which has been widely used, since 2009, to build supervised predictive models when developing real-time interactive systems. This is because it is accessible in its format (i.e. a graphical user interface – GUI) and simplified approach to ML. Significantly, it allows model training via gestural interfaces through demonstration. However, Wekinator offers the user several models to build predictive systems with. This paper explores which ML models (in Wekinator) are the most useful for predicting an output in the context of interactive music composition. We use two performance gestures for piano, with opposing datasets, to train available ML models, investigate compositional outcomes and frame the investigation. Our results show ML model choice is important for mapping performance gestures because of disparate mapping accuracies and behaviours found between all Wekinator ML models.

Type: Proceedings paper
Title: New Interfaces for Classifying Performance Gestures in Music
Event: Intelligent Data Engineering and Automated Learning (IDEAL)
Location: Univ Manchester, Manchester, ENGLAND
Dates: 14 Nov 2019 - 16 Nov 2019
ISBN-13: 978-3-030-33616-5
Open access status: An open access version is available from UCL Discovery
DOI: 10.1007/978-3-030-33617-2_4
Publisher version: https://doi.org/10.1007/978-3-030-33617-2_4
Language: English
Additional information: Interactive machine learning, Wekinator, Myo, HCI, Performance gestures, Interactive music, Gestural interfaces
Keywords: Science & Technology, Technology, Computer Science, Artificial Intelligence, Computer Science, Information Systems, Computer Science, Theory & Methods, Computer Science, Interactive machine learning, Wekinator, Myo, HCI, Performance gestures, Interactive music, Gestural interfaces
UCL classification: UCL
UCL > Provost and Vice Provost Offices > School of Education
UCL > Provost and Vice Provost Offices > School of Education > UCL Institute of Education
UCL > Provost and Vice Provost Offices > School of Education > UCL Institute of Education > IOE - Culture, Communication and Media
URI: https://discovery.ucl.ac.uk/id/eprint/10216491
Downloads since deposit
3Downloads
Download activity - last month
Download activity - last 12 months
Downloads by country - last 12 months

Archive Staff Only

View Item View Item