Touch Interaction for Corpus-based Audio–Visual Synthesis

Diemo Schwarz

Proceedings of the International Conference on New Interfaces for Musical Expression

Abstract:

Audio–visual corpus-based synthesis extends the principle of concatenative sound synthesis to the visual domain, where, in addition to the sound corpus (i.e. a collection of segments of recorded sound with a perceptual description of their sound character), the artist uses a corpus of still images with visual perceptual description (colour, texture, brightness, entropy), in order to create an audio–visual musical performance by navigating in real-time through these descriptor spaces, i.e. through the collection of sound grains in a space of perceptual audio descriptors, and at the same time through the visual descriptor space, i.e. selecting images from the visual corpus for rendering, and thus navigate in parallel through both corpora interactively with gestural control via movement sensors. The artistic–scientific question that is explored here is how to control at the same time the navigation through the audio and the image descriptor spaces with gesture sensors, in other words, how to link the gesture sensing to both the image descriptors and the sound descriptors in order to create a symbiotic multi-modal embodied audio–visual experience.

Citation:

Diemo Schwarz. 2023. Touch Interaction for Corpus-based Audio–Visual Synthesis. Proceedings of the International Conference on New Interfaces for Musical Expression. DOI: 10.5281/zenodo.11189220

BibTeX Entry:

  @inproceedings{nime2023_55,
 abstract = {Audio–visual corpus-based synthesis extends the principle of concatenative sound synthesis to the visual domain, where, in addition to the sound corpus (i.e. a collection of segments of recorded sound with a perceptual description of their sound character), the artist uses a corpus of still images with visual perceptual description (colour, texture, brightness, entropy), in order to create an audio–visual musical performance by navigating in real-time through these descriptor spaces, i.e. through the collection of sound grains in a space of perceptual audio descriptors, and at the same time through the visual descriptor space, i.e. selecting images from the visual corpus for rendering, and thus navigate in parallel through both corpora interactively with gestural control via movement sensors.
The artistic–scientific question that is explored here is how to control at the same time the navigation through the audio and the image descriptor spaces with gesture sensors, in other words, how to link the gesture sensing to both the image descriptors and the sound descriptors in order to create a symbiotic multi-modal embodied audio–visual experience.},
 address = {Mexico City, Mexico},
 articleno = {55},
 author = {Diemo Schwarz},
 booktitle = {Proceedings of the International Conference on New Interfaces for Musical Expression},
 doi = {10.5281/zenodo.11189220},
 editor = {Miguel Ortiz and Adnan Marquez-Borbon},
 issn = {2220-4806},
 month = {May},
 numpages = {8},
 pages = {394--401},
 title = {Touch Interaction for Corpus-based Audio–Visual Synthesis},
 track = {Papers},
 url = {http://nime.org/proceedings/2023/nime2023_55.pdf},
 year = {2023}
}