Users collaborate with this system to form an improvised musical ensemble. The group of human users access the software via Apple iPads, where they tap a GUI button to make musical sounds. After a predetermined set of time, the system changes the GUI to a new selection of sounds.
The system ‘triggers an interface update when it calculates that the amount of gestural change in the ensemble has exceeded a predetermined threshold. These moments can correspond to natural “segments” of ensemble improvisations, and so are appropriate times to intervene in the performance… This interaction results in a musical experience similar to “structured improvisation” or “game pieces” except with an ensemble-tracking agent director, rather than a human.’