This paper presents a new approach to sound composition for soundtrack composers and sound designers. We propose a tool for usable sound manipulation and composition that targets sound variety and expressive rendering of the composition. We first automatically segment audio recordings into atomic grains which are displayed on our navigation tool ac- cording to signal properties. To perform the synthesis, the user selects one recording as model for rhythmic pattern and timbre evolution, and a set of audio grains. Our synthesis system then processes the chosen sound material to create new sound sequences based on onset detection on the record- ing model and similarity measurements between the model and the selected grains. With our method, we can create a large variety of sound events such as those encountered in virtual environments or other training simulations, but also sound sequences that can be integrated in a music composition. We present a usability-minded interface that allows to manipulate and tune sound sequences in an appropriate way for sound design.
Picard, C., Frisson, Ch., Vanderdonckt, J., Tardieu, D., Dutoit, Th.
ACM Press, New York
Towards User-friendly Audio Creation, Proc. of 5th ACM Conference on Interaction with Sound AudioMostly'2010
2010
21
Downloads
Towards User-friendly Audio Creation (power point)