gSPAT is a interactive system able to map gestural data to spatialisation parameters using a commercial gestural controller such that the results are musically meaningful to both performer and audience.
This system is currently under development. Until today, it has been developed using the Myo armband as input device, which is able t o send OSC and MIDI data through the MyoMusic application. The Myo data is mapped using a Support Vector Machine classification system (ml-lib) and consequentially used to drive a sound spatialiser.
Recently the gSPAT has been involved in the realisation of VoicErutseG v0.1, an interactive performance of Sequenza III by L.Berio and Stripsody by C.Berberian; performed at Frontiers Festival and electroAQustica.
Moreover, it has been utilised to develop touch-less Virtual Instruments and Mixed Reality interaction system prototypes.
From experiments and applications of the gSPAT, it has been usable to perform different pieces without adding any choreographic indications to the performer; resulted in a ‘natural’ experience for the performer and applicable different contexts such as Mixed Reality and touch-free music interaction.