Exploring the potential of interactive Machine Learning for Sound Generation: A preliminary study with sound artists

Gerardo Meza

Proceedings of the International Conference on New Interfaces for Musical Expression

Abstract:

Interactive Machine Learning (IML) is an approach previously explored in music discipline. However, its adaptation in sound synthesis as an algorithmic method of creation has not been examined. This article presents the prototype ASCIML, an Assistant for Sound Creation with Interactive Machine Learning, that allows musicians to use IML to create personalized datasets and generate new sounds. Additionally, a preliminary study is presented which aims to evaluate the potential of ASCIML as a tool for sound synthesis and to gather feedback and suggestions for future improvements. The prototype can be used in Google Colaboratory and is divided into four main stages: Data Design, Training, Evaluation and Audio Creation. Results from the study, which involved 27 musicians with no prior knowledge of Machine Learning (ML), showed that most participants preferred using microphone recording and synthesis to design their dataset and that the Envelopegram visualization was found to be particularly meaningful to understand sound datasets. It was also found that the majority of participants preferred to implement a pre-trained model on their data and relied on hearing the audio reconstruction provided by the interface to evaluate the model performance. Overall, the study demonstrates the potential of ASCIML as a tool for hands-on neural audio sound synthesis and provides valuable insights for future developments in the field.

Citation:

Gerardo Meza. 2023. Exploring the potential of interactive Machine Learning for Sound Generation: A preliminary study with sound artists. Proceedings of the International Conference on New Interfaces for Musical Expression. DOI: 10.5281/zenodo.11189313

BibTeX Entry:

  @inproceedings{nime2023_88,
 abstract = {Interactive Machine Learning (IML) is an approach previously explored in music discipline. However, its adaptation in sound synthesis as an algorithmic method of creation has not been examined. This article presents the prototype ASCIML, an Assistant for Sound Creation with Interactive Machine Learning, that allows musicians to use IML to create personalized datasets and generate new sounds. Additionally, a preliminary study is presented which aims to evaluate the potential of ASCIML as a tool for sound synthesis and to gather feedback and suggestions for future improvements. The prototype can be used in Google Colaboratory and is divided into four main stages: Data Design, Training, Evaluation and Audio Creation. Results from the study, which involved 27 musicians with no prior knowledge of Machine Learning (ML), showed that most participants preferred using microphone recording and synthesis to design their dataset and that the Envelopegram visualization was found to be particularly meaningful to understand sound datasets. It was also found that the majority of participants preferred to implement a pre-trained model on their data and relied on hearing the audio reconstruction provided by the interface to evaluate the model performance. Overall, the study demonstrates the potential of ASCIML as a tool for hands-on neural audio sound synthesis and provides valuable insights for future developments in the field.},
 address = {Mexico City, Mexico},
 articleno = {88},
 author = {Gerardo  Meza},
 booktitle = {Proceedings of the International Conference on New Interfaces for Musical Expression},
 doi = {10.5281/zenodo.11189313},
 editor = {Miguel Ortiz and Adnan Marquez-Borbon},
 issn = {2220-4806},
 month = {May},
 numpages = {4},
 pages = {604--607},
 title = {Exploring the potential of interactive Machine Learning for Sound Generation: A preliminary study with sound artists},
 track = {Work in Progress},
 url = {http://nime.org/proceedings/2023/nime2023_88.pdf},
 year = {2023}
}