Skip to Main content Skip to Navigation
Book sections

Interactive Sound Texture Synthesis Through Semi-Automatic User Annotations

Diemo Schwarz 1 Baptiste Caramiaux 1
1 Equipe Interactions musicales temps-réel
STMS - Sciences et Technologies de la Musique et du Son
Abstract : We present a way to make environmental recordings controllable again by the use of continuous annotations of the high-level semantic parameter one wishes to control, e.g. wind strength or crowd excitation level. A partial annotation can be propagated to cover the entire recording via cross-modal analysis between gesture and sound by canonical time warping (CTW). The annotations serve as a descriptor for lookup in corpus-based concatenative synthesis in order to invert the sound/annotation relationship. The workflow has been evaluated by a preliminary subject test and results on canonical correlation analysis (CCA) show high consistency between annotations and a small set of audio descriptors being well correlated with them. An experiment of the propagation of annotations shows the superior performance of CTW over CCA with as little as 20 s of annotated material.
Complete list of metadatas

Cited literature [19 references]  Display  Hide  Download

https://hal.archives-ouvertes.fr/hal-01161076
Contributor : Ircam Ircam <>
Submitted on : Monday, June 8, 2015 - 2:27:31 PM
Last modification on : Wednesday, October 14, 2020 - 4:07:45 AM
Long-term archiving on: : Tuesday, April 25, 2017 - 4:22:29 AM

File

index.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : hal-01161076, version 1

Citation

Diemo Schwarz, Baptiste Caramiaux. Interactive Sound Texture Synthesis Through Semi-Automatic User Annotations. Springer International Publishing; Aramaki, M., Derrien, O., Kronland-Martinet, R., Ystad, S. Sound, Music, and Motion, Lecture Notes in Computer Science, Vol. 8905, pp.372-392, 2014. ⟨hal-01161076⟩

Share

Metrics

Record views

376

Files downloads

292