F. Pachet and D. Cazaly, A classification of musical genre, Proc. RIAO Content-Based Multimedia Information Access Conference, 2000.

G. Tzanetakis and P. Cook, Musical genre classification of audio signals, IEEE Transactions on Acoustics, Speech and Signal Processing, vol.10, pp.293-302, 2002.

J. Aucouturier and F. Pachet, Improving timbre similarity: How high is the sky?, Journal of Negative Results in Speech and Audio Sciences, vol.1, issue.1, 2004.

A. Ghias, J. Logan, D. Chamberlin, and B. Smith, Query by Humming: Musical Information Retrieval in an Audio Database, pp.213-236, 1995.

R. B. Dannenberg, W. P. Birmingham, G. Tzanetakis, C. Meek, N. Hu et al., The MUSART testbed for query-by-humming evaluation, Computer Music Journal, vol.28, issue.2, pp.34-48, 2004.

Y. Kim, D. Williamson, and S. Pilli, Towards quantifying the album effect in artist identification, Proc. of Int.Conf. on Music Information Retrieval (ISMIR), 2006.

Y. Li and D. Wang, Singing voice separation from monaural recordings, Proc. of Int.Conf. on Music Information Retrieval, 2006.

A. Ozerov, P. Philippe, R. Gribonval, and F. Bimbot, One microphone singing voice separation using source-adapted models, Proc. IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), 2005.
URL : https://hal.archives-ouvertes.fr/inria-00564491

S. Vembu and S. Baumann, Separation of vocals from polyphonic audio recordings, Proc. Int. Conf. on Music Information Retrieval (ISMIR), 2005.

G. Poliner, D. Ellis, A. Ehmann, E. Gomez, S. Streich et al., Melody transcription from music audio: Approaches and evaluation, IEEE Transactions on Audio, Speech and Language Processing, vol.15, issue.4, 2007.

A. Bregman, Auditory Scene Analysis: The Perceptual Organization of Sound, 1990.

M. Lagrange and G. Tzanetakis, Sound source tracking and formation using normalized cuts, Proc. Int. Conf. on Acoustics, Speech and Signal Processing (ICASSP), 2007.

, Computational Auditory Scene Analysis, 1998.

, Computational Auditory Scene Analysis: Principles, Algorithms and Applications, 2006.

E. Vincent, Musical source separation using time-frequency priors, IEEE Trans. on Audio, Speech and Language Processing, vol.14, issue.1, pp.91-98, 2006.
URL : https://hal.archives-ouvertes.fr/inria-00544269

S. T. Roweis, One microphone source separation, Proceedings of the Neural Information Processing Systems(NIPS), pp.793-799, 2000.

J. Shi and J. Malik, Normalized cuts and image segmentation, IEEE Trans. on Pattern Analysis and Machine Intelligence, vol.22, issue.8, pp.888-905, 2000.

D. Ellis and K. Lee, Minimal-impact audio-based personal archives, Proc. ACM Workshop on Continuous Archival and Retrieval of Personal Experience (CARPE), 2004.

R. Cai, L. Lu, and A. Hanjalic, Unsupervised content discovery in composite audio, Proc. ACM Multimedia, 2005.

S. Dubnov and T. Appel, Audio segmentation by singular value clustering, Proc. of Int.Conf. on Computer Music (ICMC), 2004.

F. Bach and M. I. Jordan, Blind one-microphone speech separation: A spectral learning approach, Proc. Neural Information Processing Systems (NIPS), 2004.

F. R. Bach and M. I. Jordan, Learning spectral clustering, with application to speech separation, Journal of Machine Learning Research, vol.7, 1963.

S. Srinivasan and M. Kankanhalli, Harmonicity and dynamics based audio separation, Proc. Int. Conf. on Acoustics, Speech, and Signal Processing (ICASSP'03), vol.5, pp.640-643, 2003.
DOI : 10.1109/icassp.2003.1200052

URL : http://www.comp.nus.edu.sg/~mohan/papers/audio_sep.pdf

S. Srinivasan, Auditory blobs, Proc. Int. Conf. on Acoustics, Speech, and Signal Processing (ICASSP'04), vol.4, pp.313-316, 2004.
DOI : 10.1109/icassp.2004.1326826

R. Mcaulay and T. Quatieri, Speech analysis/synthesis based on sinusoidal representation, IEEE Trans. on Acoustics, Speech, and Signal Processing, vol.34, issue.4, pp.744-754, 1986.

M. Lagrange, S. Marchand, and J. Rault, Enhancing the tracking of partials for the sinusoidal modeling of polyphonic sounds, IEEE Transactions on Acoustics, Speech and Signal Processing, 2007.
URL : https://hal.archives-ouvertes.fr/hal-00308191

M. S. Puckette and J. C. Brown, Accuracy of frequency estimates using the phase vocoder, IEEE Trans. on Audio and Speech Processing, vol.6, issue.2, 1998.

S. Marchand and M. Lagrange, On the equivalence of phase-based methods for the estimation of instantaneous frequency, Proc. European Conference on Signal Processing, 2006.
URL : https://hal.archives-ouvertes.fr/hal-00308038

M. Lagrange and S. Marchand, Estimating the instantaneous frequency of sinusoidal components using phase-based methods, Journal of the Audio Engineering Society, 2007.
URL : https://hal.archives-ouvertes.fr/hal-00308193

M. Abe and J. O. Smith, Design Criteria for Simple Sinusoidal Parameter Estimation based on Quadratic Interpolation of FFT Magnitude Peaks, 117th Convention of the Audio Engineering Society, 2004.

R. Badeau, B. David, and G. Richard, High resolution spectral analysis of mixtures of complex exponentials modulated by polynomials, IEEE Trans. on Signal Processing, vol.54, issue.4, pp.1341-1350, 2006.
URL : https://hal.archives-ouvertes.fr/hal-00479781

T. Virtanen and A. Klapuri, Separation of harmonic sound sources using sinusoidal modeling, Proc. ICASSP, vol.2, pp.765-768, 2000.
DOI : 10.1109/icassp.2000.859072

URL : http://www.icsi.berkeley.edu/~dpwe/research/etc/icassp2000/pdf/2718_159.PDF

J. Rosier and Y. Grenier, Unsupervised classification techniques for multipitch estimation, 2004.

L. Martins and A. Ferreira, PCM to MIDI transposition, Proc. of Audio Engineering Society (AES), 2002.

M. Lagrange and S. Marchand, Assessing the quality of the extraction and tracking of sinusoidal components: Towards an evaluation methodology, Proceedings of the Digital Audio Effects (DAFx'06) Conference, pp.239-245, 2006.
URL : https://hal.archives-ouvertes.fr/hal-00308186

, Signal Processing Methods for Music Transcription, 2006.

A. Klapuri, Multiple fundamental frequency estimation by summing harmonic amplitudes, International Conference on Music Information Retrieval (ISMIR), 2006.

P. Boersma and D. Weenink, Praat: doing phonetics by computer (version 4.5.06), 2006.

S. Davis and P. Mermelstein, Experiments in syllable-based recognition of continuous speech, IEEE Transactions on Acoustics, Speech and Signal Processing, vol.28, pp.357-366, 1980.

I. Witten and E. Frank, Data Mining: Practical machine learning tools and techniques, 2005.

A. Jourjine, S. Richard, and O. Yilmaz, Blind separation of disjoint orthogonal signals: Demixing n sources from 2 mixtures, Proc. ICASSP, 2000.
DOI : 10.1109/icassp.2000.861162

C. Avendano, Frequency-domain source identification and manipulation in stereo mixes for enhancement, suppression, and re-panning applications, Proc. IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), 2003.
DOI : 10.1109/aspaa.2003.1285818

URL : http://www.ee.columbia.edu/~dpwe/papers/Aven03-unmix.pdf

M. Lagrange, L. G. Martins, and G. Tzanetakis, Semi-Automatic Mono to Stereo Up-mixing using Sound Source Formation, 122th Convention of the Audio Engineering Society, 2007.

. Mathieu-lagrange, He obtained a postgraduate diploma focusing on spectral sound synthesis at the University of Bordeaux 1, M'07) was born in, 1978.