Towards user-friendly audio creation

  • Authors:
  • Cécile Picard;Christian Frisson;Jean Vanderdonckt;Damien Tardieu;Thierry Dutoit

  • Affiliations:
  • Research Program on Digital, Art Technologies, Belgium;Université catholique de Louvain (UCL), Louvain-la-Neuve, Belgium;Université catholique de Louvain (UCL), Louvain-la-Neuve, Belgium;Université de Mons, Mons, Belgium;Université de Mons, Mons, Belgium

  • Venue:
  • Proceedings of the 5th Audio Mostly Conference: A Conference on Interaction with Sound
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper presents a new approach to sound composition for soundtrack composers and sound designers. We propose a tool for usable sound manipulation and composition that targets sound variety and expressive rendering of the composition. We first automatically segment audio recordings into atomic grains which are displayed on our navigation tool according to signal properties. To perform the synthesis, the user selects one recording as model for rhythmic pattern and timbre evolution, and a set of audio grains. Our synthesis system then processes the chosen sound material to create new sound sequences based on onset detection on the recording model and similarity measurements between the model and the selected grains. With our method, we can create a large variety of sound events such as those encountered in virtual environments or other training simulations, but also sound sequences that can be integrated in a music composition. We present a usability-minded interface that allows to manipulate and tune sound sequences in an appropriate way for sound design.