Skip to Main content Skip to Navigation
Book sections

Interactive Sound Texture Synthesis Through Semi-Automatic User Annotations

Diemo Schwarz 1 Baptiste Caramiaux 1
1 Equipe Interactions musicales temps-réel
STMS - Sciences et Technologies de la Musique et du Son
Abstract : We present a way to make environmental recordings controllable again by the use of continuous annotations of the high-level semantic parameter one wishes to control, e.g. wind strength or crowd excitation level. A partial annotation can be propagated to cover the entire recording via cross-modal analysis between gesture and sound by canonical time warping (CTW). The annotations serve as a descriptor for lookup in corpus-based concatenative synthesis in order to invert the sound/annotation relationship. The workflow has been evaluated by a preliminary subject test and results on canonical correlation analysis (CCA) show high consistency between annotations and a small set of audio descriptors being well correlated with them. An experiment of the propagation of annotations shows the superior performance of CTW over CCA with as little as 20 s of annotated material.
Complete list of metadata

Cited literature [19 references]  Display  Hide  Download
Contributor : Ircam Ircam <>
Submitted on : Monday, June 8, 2015 - 2:27:31 PM
Last modification on : Friday, January 8, 2021 - 2:04:40 PM
Long-term archiving on: : Tuesday, April 25, 2017 - 4:22:29 AM


Files produced by the author(s)


  • HAL Id : hal-01161076, version 1


Diemo Schwarz, Baptiste Caramiaux. Interactive Sound Texture Synthesis Through Semi-Automatic User Annotations. Springer International Publishing; Aramaki, M., Derrien, O., Kronland-Martinet, R., Ystad, S. Sound, Music, and Motion, Lecture Notes in Computer Science, Vol. 8905, pp.372-392, 2014. ⟨hal-01161076⟩



Record views


Files downloads