.. Corpus-d-'interprétations-musicales, 163 6.7.1 Particularités de l'interprétation musicale 163 6.7.2 Support : Partition utilisée, p.165

A. Résumé-du-chapitre and .. , 170 A.2 Introduction 171 A.3 Systèmes de gestion et de création de corpus de parole, p.175

A. Chapitre and .. Annexe, 194 C.2 Marquage des signaux EGG

E. Post-processing-des-signaux, G. .. , and .. , 198 C.7.1 Corrélation entre les marqueurs PSOLA, C.7.2 Corrélation entre les signaux dEGG et, 0200.

.. Synthesis, 208 D.6.1 Phones selection, p.211

]. V. Aubergé, N. Audibert, and A. Rilliard, E-Wiz : A trapper protocol for hunting the expressive speech corpora, LREC2004, pp.179-182, 2004.

]. V. Aubergé, N. Audibert, and A. Rilliard, Auto-annotation : an alternative method to label expressive corpora, LREC2006 -Workshop on Emotional Corpora, pp.45-46, 2006.

]. J. Averill, Emotion : Theory, research and experience, chapitre A constructivist view of emotion, pp.305-339, 1980.

]. Bachorowski, M. J. Smoski, and M. J. Owren, The acoustic features of human laughter, The Journal of the Acoustical Society of America, vol.110, issue.3, pp.1581-1597, 2001.
DOI : 10.1121/1.1391244

]. G. Bailly and C. Benoit, Talking machines : Theories, models and designs, 1992.

]. G. Bailly, Characterising formant trajectories by tracking vocal tract resonances, 1995.

]. G. Bailly and . Holm, Special issue on quantitative prosody modelling for natural speech description and generation, chapitre SFC : a trainable prosodic model, pp.348-364, 2005.

]. C. Barras, E. Geoffrois, Z. Wu, and M. Liberman, Transcriber : a Free Tool for Segmenting, Labeling and Transcribing Speech, LREC1998, pp.1373-1376, 1998.

]. V. Beaucousin, A. Lacheret, M. R. Turbelin, M. Morel, and B. , Mazoyer et N. Mazoyer. FRMI Study of Emotional Speech Comprehension, Cerebral Cortex, 2006.

]. G. Beller and A. Marty, Talkapillar : outil d'analyse de corpus oraux, Rencontres Jeunes Chercheurs de L'Ecole Doctorale 268, pp.97-100, 2006.

]. G. Beller, D. Schwarz, T. Hueber, and X. Rodet, Speech Rates in French Expressive Speech, Speech Prosody, pp.672-675, 2006.

]. G. Bibliographie and . Beller, Context Dependent Transformation of Expressivity in Speech Using a Bayesian Network, ParaLing, pp.48-89, 2007.

]. G. Beller-]-g, N. Beller, X. Obin, and . Rodet, Influence de l'expressivité sur le degré d'articulation Articulation Degree as a Prosodic Dimension of Expressive Speech, RJCP, Rencontres Jeunes Chercheurs de la Parole Speech Prosody, pp.24-27, 2007.

]. G. Beller, C. Veaux, X. Rodet-]-g, C. Beller, G. Veaux et al., Transformation of Expressivity in Speech The Role of Prosody in the Expression of Emotions in English and in French IRCAM Corpus Tools : Système de Gestion de Corpus de Parole. TAL LinguaStream : An Integrated Environment for Computational Linguistics Experimentation ATLAS : A Flexible and Extensible Architecture for Linguistic Annotation A formal framework for linguistic annotation, LREC2008 -workshop on emotions 11th Conference of the European Chapter of the Association of Computational Linguistics (Companion Volume) Proceedings of the Second International Conference on Language Resources and Evaluation Blankinship et R. Beckwith. Uist '01 : Proceedings of the 14th annual acm symposium on user interface software and technology, chapitre Tools for expressive text-to-speech markup, pp.95-98, 2000.

]. P. Boersma and D. Weenink, Praat, a system for doing phonetics by computer, Glot international-9 of 10 Roebel et X. Rodet. Sound Analysis and Processing with AudioSculpt 2, pp.341-345, 2001.

]. R. Bresin, Virtual Virtuosity. Studies in Automatic Music Performance, 2000.

]. R. Bresin and S. Dahl, The sounding object, chapitre Experiments on gestures : walking, running, and hitting, pp.111-136, 2003.

]. R. Buck, The communication of emotion, 1984.

]. M. Bulut, S. Lee, and S. Narayanan, A Statistical Approach for Modeling Prosody Features using POS Tags for Emotional Speech Synthesis, 2007 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP '07, 2007.
DOI : 10.1109/ICASSP.2007.367300

]. F. Burkhardt, N. Audibert, L. Malatesta, O. Urk, and L. , Arslan et V. Auberge. Emotional Prosody -Does Culture Make A Difference ?, Speech Prosody, 2006.

]. N. Campbell and P. Mokhtari, Voice quality : the 4th prosodic dimension, XVth ICPhS, pp.2417-2420, 2003.

]. N. Campbell, H. Kashioka, and R. Ohara, No laughing matter, Interspeech, pp.465-468, 2005.

]. N. Campbell, Changes in voice quality due to social conditions, ICPhS, 2007.

]. N. Campbell, Whom we laugh with affects how we laugh, Interdisciplinary Workshop on The Phonetics of Laughter, 2007.

]. S. Cassidy and J. Harrington, Multi-level annotation in the Emu speech database management system, Speech Communication, vol.33, issue.1-2, pp.1-2, 2001.
DOI : 10.1016/S0167-6393(00)00069-8

]. A. De-cheveigné and H. Kawahara, YIN, a fundamental frequency estimator for speech and music, The Journal of the Acoustical Society of America, vol.111, issue.4, pp.1917-1930, 2002.
DOI : 10.1121/1.1458024

]. Chung, L'expression et la perception de l'´ emotion extraite de la parole spontannée : ´ evidences du coréen et de l'anglais. phonétique, 2000.

]. H. Cunningham, D. Maynard, K. Bontcheva, and V. Tablan, GATE : A framework and graphical development environment for robust NLP tools and applications, Proceedings of the 40th Anniversary Meeting of the Association for Computational Linguistics, pp.168-175, 2002.

]. C. D-'alessandro-et and P. Mertens, Automatic pitch contour stylization using a model of tonal perception, Computer Speech and Language, pp.257-288, 1995.

]. C. D-'alessandro-et and B. Doval, Spectral representation and modelling of glottal flow signals, 1997.

B. [-d-'alessandro-2003a-]-c.-d-'alessandro-et and . Doval, Voice quality modification for emotional speech synthesis, Eurospeech, 2003.

A. , B. Doval, and K. Scherer, Voice quality : functions, analysis and synthesis, ISCAtutorial and research workshop VOQUAL'03, 2003.

]. C. Darwin, Expression of emotion in man and animals, 1965.

]. G. Degottex, E. Bianco, and X. Rodet, Usual to particular phonatory situations studied with high-speed videoendoscopy, The 6th International Conference on Voice Physiology and Biomechanics, pp.19-26, 2008.
URL : https://hal.archives-ouvertes.fr/hal-01161394

]. G. Degottex and X. Rodet, Voice source and vocal tract separation, 2008.

]. P. Delattre, M. Liberman, and F. Cooper, Acoustic Loci and Transitional Cues for Consonants, The Journal of the Acoustical Society of America, vol.27, issue.4, pp.769-773, 1955.
DOI : 10.1121/1.1908024

]. R. Descartes, Les passions de l'? ame, oeuvres philosophiques et morales, 1948.

]. L. Devillers, L. Lamel, and I. Vasilescu, Emotion detection in Taskoriented spoken dialogs, IEEE ICME, 2003.

]. L. Devillers, I. Vasilescu, and C. Mathon, Prosodic cues for perceptual emotion detection in task-oriented Human-Human corpus, ICPhS, 2003.

]. L. Devillers, L. Vidrascu, and L. Lamel, Emotion detection in real-life spoken dialogs recorded in call center, Journal of Neural Networks, special issue on Emotion and Brain, vol.18, issue.4, pp.407-422, 2005.

R. Douglas-cowie, M. Cowie, E. Schroeder, R. Douglas-cowie, I. Cowie et al., A New Emotion Database : Considerations, Sources and Scope The HUMAINE Database : Addressing the Collection and Annotation of Naturalistic and, SpeechEmotion2000 Induced Emotional Data, Affective Computing and Intelligent Interaction Doval et C. D'Alessandro. Spectral Correlates of Glottal Waveform models : An analytic study. Rapport technique, LIMSI, pp.39-44, 2000.

]. B. Doval, C. D-'alessandro-et, and N. Henrich, The Spectrum of Glottal Flow Models, Acta Acustica united with Acustica, vol.92, issue.6, pp.1026-1046, 2006.
URL : https://hal.archives-ouvertes.fr/hal-00368131

]. G. Duncan, B. Yegnanarayanan, A. Hema, and . Murthy, A non Parametric Method of Formant Estimation Using Group Delay Spectra, IEEE, 1989.

]. J. Durand, B. Laks, and C. Lyche, Un corpus numérisé pour la phonologie du français, La linguistique de corpus, pp.205-217, 2005.

]. E. Eide, A. Aaron, R. Bakis, W. Hamza, M. Picheny et al., A Corpus-Based Approach to <Ahem/> Expressive Speech Synthesis, 5th ISCA Speech Synthesis Workshop, 2004.

]. P. Ekman-1999a and . Ekman, The handbook of cognition and emotion, chapitre Basic Emotions, 1999.

]. P. Ekman-]-e, R. D. Eriksson, R. C. Rodman, and . Hubal, The handbook of cognition and emotion, chapitre Facial Expressions, chapitre Emotions in Speech : Juristic Implications, pp.301-320, 1999.

]. G. Evangelista and S. Cavaliere, Real-time time-varying frequency warping via short-time Laguerre transform Audio Effects Based on Biorthogonal Time-Varying Frequency Warping, Proc. DAFx00, pp.7-12, 2000.

]. G. Bibliographie, S. Evangelista, and . Salvatore, Time-Varying Frequency Warping : Results and Experiments, 2001.

]. G. Fant, Acoustic theory of speech production. Mouton, The Hague, 1960.

]. G. Fant, J. Liljencrants, and Q. Lin, A four-parameter model of glottal flow, STL-QPSR, vol.4, pp.1-13, 1985.

]. G. Fant, Quarterly Progress and Status Report : The LF-model revisited . Transformations and frequency domain analysis, pp.2-3, 1995.

]. G. Fant, The voice source in connected speech, Speech Communication, vol.22, issue.2-3, pp.125-139, 1997.
DOI : 10.1016/S0167-6393(97)00017-4

]. S. Fitrianie and L. J. Rothkrantz, Constructing Knowledge for Automated Text-Based Emotion Expressions, 2006.

]. I. Fónagy-1972a, E. Fónagy, and . Bérard, ??Il est huit heures??: contribution ?? I???analyse s??mantique de la vive voix, Phonetica, vol.26, issue.3, pp.157-192, 1972.
DOI : 10.1159/000259408

]. C. Fougeron and S. A. Jun, Rate effects on French intonation: prosodic organization and phonetic realization, Journal of Phonetics, vol.26, issue.1, pp.45-69, 1998.
DOI : 10.1006/jpho.1997.0062

URL : https://hal.archives-ouvertes.fr/halshs-00184989

]. A. Friberg and J. Sundberg, derived from measurements of stopping runners, The Journal of the Acoustical Society of America, vol.105, issue.3, pp.1469-1484, 1999.
DOI : 10.1121/1.426687

]. N. Frijda, Emotions, 1986.
DOI : 10.4135/9781848608399.n12

]. H. Fujisaki, Dynamic characteristics of voice fundamental frequency in speech and singing. Acoustical analysis and physiological interpretations. Rapport technique, Dept. for Speech, Music and Hearing, 1981.

]. A. Gabrielson and P. N. Juslin, Emotional Expression in Music Performance: Between the Performer's Intention and the Listener's Experience, Psychology of Music, vol.34, issue.1, pp.68-91, 1996.
DOI : 10.1177/0305735696241007

]. M. Garnier, Communiquer en environnement bruyant : de l'adaptation jusqu'au forcage vocal, 2007.
URL : https://hal.archives-ouvertes.fr/tel-00177691

]. J. Gauvain and L. Lamel, Large Vocabulary Continuous Speech Recognition : from Laboratory Systems towards Real-World Applications, 1996.
DOI : 10.1109/5.880079

]. C. Gendrot and M. Adda-decker, Analyses formantiques automatiques de voyelles orales : ´ evidence de la réduction vocalique en langues franàaise et allemande, MIDL, 2004.

]. C. Gobl and A. N. Chasaide, The role of voice quality in communicating emotion, mood and attitude, Speech Communication, vol.40, issue.1-2, pp.189-212, 2003.
DOI : 10.1016/S0167-6393(02)00082-1

]. E. Gouvea and R. M. Stern, Speaker normalization through formantbased warping of the frequency scale, Eurospeech, pp.1139-1142, 1997.

]. E. Gouvea, acoustic-feature-based frequency warping for speaker normalization, 1999.

]. A. Grobet and A. C. Simon, Diffèrents critères de d` efinition des unitès prosodiques maximales, pp.143-163, 2001.

]. C. Gussenhoven and A. C. Rietveld, Fundamental Frequency Declination in Dutch : Testing Three Hypotheses, Journal of Phonetics, vol.16, pp.355-369, 1988.

]. U. Gut, J. Milde, H. Voormann, and U. Heid, Querying Annotated Speech Corpora, Proceedings of Speech Prosody, pp.569-572, 2004.

]. A. Haermae, M. Karjalainen, L. Savioja, V. Vaelimaeki, U. K. Laine et al., Frequency-warped signal processing for audio applications, Audio Engineering Society, editeur, AES 108th Convention, 2000.

]. T. Hastie, R. Tibshirani, and J. Friedman, The elements of statistical learning, 2001.

C. Henrich, B. D-'alessandro-et, and . Doval, Glottal Flow Models : Waveforms, Spectra And Physical Measurements, 2002.
URL : https://hal.archives-ouvertes.fr/hal-00371689

]. K. Bibliographie and . Hevner, Experimental studies of the elements of expression in music, American Journal of Psychology, vol.48, pp.246-268, 1936.

]. D. Hirst and R. Espesser, Automatic modelling of fundamental frequency using a quadratic spline function, pp.71-85, 1993.

]. D. Hirst, A. Di-cristo, and R. Espesser, Prosody : Theory and experiment , chapitre Levels of representation and levels of analysis for intonation, pp.51-87, 2000.

]. V. Hozjan and Z. Kacic, A rule-based emotion-dependent feature extraction method for emotion analysis from speech, The Journal of the Acoustical Society of America, vol.119, issue.5, pp.3109-3120, 2006.
DOI : 10.1121/1.2188647

]. Hsia, C. Wu, . Et-jian-qi, and . Wu, Conversion Function Clustering and Selection for Expressive Voice Conversion, 2007 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP '07, 2007.
DOI : 10.1109/ICASSP.2007.367006

]. X. Hu, J. S. Downie, and A. F. Ehmann, Exploiting Recommended Usage Metadata : Exploratory Analyses, ISMIR, 2006.

X. Hu and J. Stephen-downie, Exploring mood metadata : Relationships with genre, artist and usage metadata, ISMIR, 2007.

]. W. James, What is an emotion?, 2007.
DOI : 10.1037/10735-001

]. T. Johnstone and K. R. Scherer, The effects of emotions on voice quality, XIVth ICPhS, pp.2029-2032, 1999.

]. P. Juslin, Five Facets of Musical Expression: A Psychologist's Perspective on Music Performance, Psychology of Music, vol.31, issue.3, pp.273-302, 2003.
DOI : 10.1177/03057356030313003

]. P. Juslin and P. Laukka, Expression, Perception, and Induction of Musical Emotions: A Review and a Questionnaire Study of Everyday Listening, Journal of New Music Research, vol.33, issue.3, pp.217-238, 2004.
DOI : 10.1080/0929821042000317813

]. P. Juslin, J. Karlsson, E. Lindstrom, A. Friberg, and . Schoonderwaldt, Play it again with feeling: Computer feedback in musical communication of emotions., Journal of Experimental Psychology: Applied, vol.12, issue.2, pp.79-95, 2006.
DOI : 10.1037/1076-898X.12.2.79

]. B. Kaehler, J. Smith, and J. Wolfe, Longueur de confusion sur la plage vocalique, JEP XXIII, 2000.

]. A. Kain, Q. Miao, and J. P. Van-santen, Spectral Control in Concatenative Speech Synthesis, SSW6, 2007.

]. J. Kates and K. H. Arehart, Multichannel Dynamic-Range Compression Using Digital Frequency Warping, EURASIP Journal on Advances in Signal Processing, vol.2005, issue.18, pp.3003-3014, 2005.
DOI : 10.1155/ASP.2005.3003

URL : http://doi.org/10.1155/asp.2005.3003

]. H. Kawahara, A. De-cheveigne, H. Banno, T. Takahashi, and T. Irino, Nearly Defect-free F0 Trajectory Extraction for Expressive Speech Modifications based on STRAIGHT, Interspeech2005, pp.537-540, 2005.

]. H. Kawahara and T. Irino, Speech separation by humans and machines, volume Engineering, chapitre Underlying Principles of a Highquality Speech Manipulation System STRAIGHT and Its Application to Speech Segregation, pp.167-180, 2005.

]. M. Kendall and A. Stuart, The advanced theory of statistics, Charles Griffin, vol.1, 1969.

]. S. Kipper and D. Todt, Series of similar vocal elements as a crucial acoustic structure in human laughter, 2007.

]. D. Klatt, Software for a cascade/parallel formant synthesizer, The Journal of the Acoustical Society of America, vol.67, issue.3, pp.971-995, 1980.
DOI : 10.1121/1.383940

]. F. Klingholz and F. Martin, Quantitative Spectral Evaluation of Shimmer and Jitter, Journal of Speech Language and Hearing Research, vol.28, issue.2, pp.169-174, 1985.
DOI : 10.1044/jshr.2802.169

]. G. Kreutz, U. Ott, D. Teichmann, P. Osawa, and D. Vaitl, Psychology of music chapitre Using music to induce emotions : Influences of musical preference and absorption, pp.101-126, 2008.

]. C. Lai and S. Bird, Querying and updating treebanks : A critical survey and requirements analysis, Proceedings of the Australasian Language Technology Workshop, pp.139-146, 2004.

]. L. Lamel, J. Gauvain, and M. Eskénazi, Bref, a large vocabulary spoken corpus for French, EuroSpeech, pp.505-508, 1991.

]. P. Lanchantin, A. C. Morris, X. Rodet, and C. Veaux, Automatic Phoneme Segmentation with Relaxed Textual Constraints, LREC2008, 2008.
URL : https://hal.archives-ouvertes.fr/hal-01161385

]. R. Lazarus, Emotion and adaptation, 1991.

]. L. Lee and R. C. Rose, Speaker normalization using efficient frequency warping procedures, 1996 IEEE International Conference on Acoustics, Speech, and Signal Processing Conference Proceedings, pp.353-356, 1996.
DOI : 10.1109/ICASSP.1996.541105

]. R. Levenson, P. Ekman, and W. V. Friesen, Voluntary Facial Action Generates Emotion-Specific Autonomic Nervous System Activity, Psychophysiology, vol.228, issue.4, pp.363-384, 1990.
DOI : 10.1016/0022-1031(84)90047-7

B. Li and . Ogihara, Detecting emotion in music, ISMIR, 2003.

]. B. Lindblom, Economy of speech gestures, volume The Production of Speech, 1983.

]. Lolive, N. Barbot, and O. Boeffard, Modélisation B-spline de contours mélodiques avec estimation du nombre de paramètres libres par un critère MDL, JEP, 2006.

]. H. Lu and J. O. Smith, Joint Estimation of Vocal Tract Filter and Glottal Source Waveform via Convex Optimization, 1999.

]. Lu, Toward a High-quality Singing Synthesizer with Vocal Texture Control, 2002.

]. L. Lu and S. Zhang, Automatic mood detection and tracking of music audio signals, IEEE Transactions on Audio, Speech and Language Processing, vol.14, issue.1, 2006.
DOI : 10.1109/TSA.2005.860344

]. G. Luck, P. Toiviainen, J. Erkkila, O. Lartillot, K. Riikkila et al., Psychology of music chapitre Modelling the relationships between emotional responses to, and musical content of, music therapy improvisations, pp.25-45, 2008.

]. B. Macwhinney, The CHILDES Project: Tools for Analyzing Talk (third edition): Volume I: Transcription format and programs, Volume II: The database, Computational Linguistics, vol.26, issue.4, 2000.
DOI : 10.1162/coli.2000.26.4.657

]. P. Mandel and D. Ellis, Support vector machine active learning for music retrieval, Multimedia Systems, vol.10, issue.5, 2006.
DOI : 10.1007/s00530-006-0032-2

]. L. Meyer, Emotion and meaning in music, 1956.
DOI : 10.7208/chicago/9780226521374.001.0001

. Mixdorff, A novel approach to the fully automatic extraction of Fujisaki model parameters, 2000 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings (Cat. No.00CH37100), 1999.
DOI : 10.1109/ICASSP.2000.861811

]. A. Morris, Automatic segmentation. Rapport technique, IRCAM, 2006.
URL : https://hal.archives-ouvertes.fr/hal-01161385

]. K. Murphy, The Bayes Net Toolbox for Matlab, Computing Science and Statistics, 2001.

]. I. Murray and J. L. Arnott, Applying an analysis of acted vocal emotions to improve the simulation of synthetic speech, Computer Speech & Language, vol.22, issue.2, pp.107-129, 2008.
DOI : 10.1016/j.csl.2007.06.001

]. H. Murthy, Formant Extraction from Phase Weigthed Group Delay Function, 1989.
DOI : 10.1049/el:19891080

]. H. Murthy, non parametric method using group delay function, 1989.

]. H. Murthy, K. V. Madhu-murthy, and B. Yegnanarayana, Formant extraction from phase using weighted group delay function, Electronics Letters, pp.1609-1611, 1989.
DOI : 10.1049/el:19891080

H. A. Murthy and B. Yegnanarayana, Speech processing using group delay functions, Signal Processing, vol.22, issue.3, pp.259-267, 1991.
DOI : 10.1016/0165-1684(91)90014-A

]. H. Murthy and B. Yegnanarayana, Formant extraction from group delay function, Speech Communication, vol.10, issue.3, pp.209-221, 1991.
DOI : 10.1016/0167-6393(91)90011-H

URL : http://www.sciencedirect.com/science?_ob=ShoppingCartURL&_method=add&_eid=1-s2.0-016763939190011H&originContentFamily=serial&_origin=article&_ts=1487195584&md5=37085bd73fb5c9d1739be19605cb713e

]. H. Murthy and V. R. Rao-gadde, The modified group delay function and its application to phoneme recognition, 2003 IEEE International Conference on Acoustics, Speech, and Signal Processing, 2003. Proceedings. (ICASSP '03)., 2003.
DOI : 10.1109/ICASSP.2003.1198718

]. P. Nakov, A. Schwartz, B. Wolf, and M. Hearst, Supporting annotation layers for natural language processing, Proceedings of the ACL 2005 on Interactive poster and demonstration sessions , ACL '05, pp.65-68, 2005.
DOI : 10.3115/1225753.1225770

]. N. Obin, J. P. Goldman, M. Avanzi, and A. Lacheret-dujour, Comparaison de 3 outils de détection automatique de proéminence en français parlé, XXVIIèmes Journées d' ´ Etudes de la Parole, pp.153-157, 2008.

]. N. Oostdijk, The Spoken Dutch Corpus : Overview and first evaluation, LREC2000, pp.887-893, 2000.

]. C. Palmer, Mapping musical thought to musical performance., Journal of Experimental Psychology: Human Perception and Performance, vol.15, issue.2, pp.331-346, 1989.
DOI : 10.1037/0096-1523.15.2.331

]. A. Patel, Music, language, and the brain, 2008.
DOI : 10.1093/acprof:oso/9780195123753.001.0001

]. G. Peeters, Modeles et modification du signal sonore adaptés a ses caracteristiques locales, 2001.

]. G. Peeters, A large set of audio features for sound description (similarity and classication) in the CUIDADO project, 2004.

H. R. Pfitzinger, Five Dimensions of Prosody : Intensity, Intonation , Timing, Voice Quality, and Degree of Reduction, H Hoffmann R. ; Mixdorff, editeur, Speech Prosody, numéro 40 de Abstract Book, pp.6-9, 2006.

]. R. Bibliographie and . Picard, Affective computing, 1997.

]. M. Piu and R. Bove, Annotation des disfluences dans les corpus oraux, RECITAL, 2007.

]. R. Plutchik, What is an Emotion?, The Journal of Psychology, vol.50, issue.2, pp.3-33, 1980.
DOI : 10.1080/00223980.1965.10543417

. Pohle, . Pampalk, and . Widmer, Evaluation of Frequently Used Audio Features for Classification of Music into Perceptual Categories, CBMI, 2005.

]. A. Potamianos and R. C. Rose, On combining frequency warping and spectral shaping in HMM based speech recognition, 1997 IEEE International Conference on Acoustics, Speech, and Signal Processing, pp.1275-1278, 1997.
DOI : 10.1109/ICASSP.1997.596178

]. N. Rasamimanana, F. Kaiser, and F. Bevilacqua, Transients control of violin players : relationships between bow acceleration and string irregular vibrations, 2007.

]. N. Rasamimanana, Geste instrumental du violoniste en situation de jeu : analyse et modélisation, 2008.

]. A. Roebel and X. Rodet, Real time signal transposition with envelope preservation in the phase vocoder, ICMC, 2005.
URL : https://hal.archives-ouvertes.fr/hal-01161347

]. A. Roebel and X. Rodet, Real Time Signal Transposition With Envelope Preservation In The Phase Vocoder, ICMC, 2005.
URL : https://hal.archives-ouvertes.fr/hal-01161347

]. M. Rossi, A. Di-cristo, D. Hirst, P. Martin, and Y. Nishinuma, L'intonation . de l'acoustiquè a la sémantique, Klincksieck, 1981.

]. K. Scherer, Emotion as a multicomponent process : A model and some cross-cultural data, Review of Personality and Social Psychology, vol.5, pp.37-63, 1984.

]. K. Scherer, Toward a Dynamic Theory of Emotion : The Component Process Model of Affective States, Geneva Studies in Emotion and Communication, vol.1, pp.1-98, 1987.

]. K. Scherer, What are emotions? And how can they be measured?, Social Science Information, vol.42, issue.1, pp.695-729, 2005.
DOI : 10.1177/0539018405058216

]. K. Scherer, Chinese spoken language processing, The Affective and Pragmatic Coding of Prosody, pp.13-14, 2006.

]. J. Schoentgen and R. De-guchtenneere, Time series analysis of jitter, Journal of Phonetics, vol.23, issue.1-2, pp.189-201, 1995.
DOI : 10.1016/S0095-4470(95)80042-5

M. Schroeder and J. Trouvain, How (Not) to Add Laughter to Synthetic Speech, Workshop on Affective Dialogue Systems Kloster Irsee, 2004.

]. D. Schwarz, G. Beller, B. Verbrugghe, and S. Britton, Real-Time Corpus-Based Concatenative Synthesis with CataRT, DAFx, 2006.
DOI : 10.1109/msp.2007.323274

URL : https://hal.archives-ouvertes.fr/hal-01161358

]. P. Shove and B. Repp, Musical motion and performance: theoretical and empirical perspectives, J. Rink, pp.55-83, 1995.
DOI : 10.1017/CBO9780511552366.004

. Shuang, . Zhi-wei, . Bakis, . Raimo, . Shechtman et al., Frequency warping based on mapping formant parameters, Interspeech, numéro 1768, 2006.

]. K. Sjölander and J. Beskow, WaveSurfer -An Open Source Speech Tool, International Conference on Spoken Language Processing, pp.464-467, 2000.

]. S. Sundaram and S. Narayanan, Automatic acoustic synthesis of human-like laughter, The Journal of the Acoustical Society of America, vol.121, issue.1, pp.527-535, 2007.
DOI : 10.1121/1.2390679

]. J. Sundberg and J. Skoog, Jaw opening, vowel and pitch, pp.43-050, 1995.
DOI : 10.1016/s0892-1997(97)80008-2

]. D. Szameitat, C. J. Darwin, A. J. Szameitat, D. Wildgruber, A. Sterr et al., Formant Characteristics of Human Laughter, Journal of Voice, vol.25, issue.1, 2007.
DOI : 10.1016/j.jvoice.2009.06.010

]. J. Tao, Y. Kang, and A. Li, Prosody conversion from neutral speech to emotional speech, IEEE Transactions on Audio, Speech, and Language Processing, vol.14, issue.4, pp.1145-1154, 2006.

]. P. Taylor, Analysis and Synthesis of Intonation using the Tilt Model. Rapport technique, Centre for Speech Technology Research, 2000.

]. P. Taylor, A. W. Black, and R. Caley, Heterogeneous Relation Graphs as a Mechanism for Representing Linguistic Information, Speech Communication, vol.3, pp.153-174, 2001.

]. T. Toda, H. Saruwatari, and K. Shikano, high quality voice conversion based on gaussian mixture model with dynamic frequency warping, Eurospeech, 2001.

]. S. Umesh, L. Cohen, N. Marinovic, and D. Nelson, Frequency-warping in speech, Proceeding of Fourth International Conference on Spoken Language Processing. ICSLP '96, pp.414-417, 1996.
DOI : 10.1109/ICSLP.1996.607142

]. J. Van-santen, L. Black, G. Cohen, A. Kain, E. Klabbers et al., Applications of computer generated expressive speech for communication disorders, Eurospeech, 2003.

]. C. Veaux, G. Beller, and X. Rodet, IrcamCorpusTools : an Extensible Platform for Spoken Corpora Exploitation, European Language Resources Association (ELRA), 2008.

]. L. Vidrascu and L. Devillers, Detection of Real-Life Emotions in Call Centers, Interspeech, 2005.

]. S. Vieillard, I. Peretz, N. Gosselin, S. Khalfa, L. Gagnon et al., Happy, sad, scary and peaceful musical excerpts for research on emotions, Cognition & Emotion, vol.58, issue.4, 2007.
DOI : 10.1007/BF02294516

]. F. Villavicencio, A. Roebel, and X. Rodet, Improving Lpc Spectral Envelope Extraction Of Voiced Speech By True-Envelope Estimation, 2006 IEEE International Conference on Acoustics Speed and Signal Processing Proceedings, 2006.
DOI : 10.1109/ICASSP.2006.1660159

URL : https://hal.archives-ouvertes.fr/hal-01161354

]. D. Vincent, O. Rosec, and T. Chonavel, Estimation du signal glottique basée sur un modèle ARX, GRETSI, 2005.

]. D. Vincent, O. Rosec, and T. Chonavel, Estimation of LF glottal source parameters based on an ARX model, 9th European Conference on Speech Communication and Technology, pp.333-336, 2005.

]. D. Vincent, O. Rosec, and T. Chonavel, Glottal Closure Instant Estimation using an Appropriateness Measure of the Source and Continuity Constraints, 2006 IEEE International Conference on Acoustics Speed and Signal Processing Proceedings, pp.14-19, 2006.
DOI : 10.1109/ICASSP.2006.1660037

]. D. Vincent, O. Rosec, and T. Chonavel, A New Method for Speech Synthesis and Transformation Based on an ARX-LF Source-Filter Decomposition and HNM Modeling, 2007 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP '07, pp.525-528, 2007.
DOI : 10.1109/ICASSP.2007.366965

]. A. Vines, C. L. Krumhansl, M. M. Wanderley, I. M. Dalca, and D. J. Levitin, Dimensions of Emotion in Expressive Musical Performance, The Neurosciences and Music II : From Perception to Performance, pp.462-466, 2005.
DOI : 10.1207/s15327957pspr0101_2

]. S. Wabnik, G. Schuller, U. Kraemer, and J. Hirschfeld, Frequency Warping in Low Delay Audio Coding, Proceedings. (ICASSP '05). IEEE International Conference on Acoustics, Speech, and Signal Processing, 2005., 2005.
DOI : 10.1109/ICASSP.2005.1415676

]. C. Wightman and N. Campbell, Improved labeling of prosodic structure, IEEE Trans. on Speech and Audio Processing, 1995.

]. J. Wouters and M. Macon, Control of spectral dynamics in concatenative speech synthesis, IEEE Transactions on Speech and Audio Processing, pp.30-38, 2001.
DOI : 10.1109/89.890069

]. J. Yamagishi, K. Onishi, T. Masuko, and T. Kobayashi, Acoustic Modeling of Speaking Styles and Emotional Expressions in HMM-Based Speech Synthesis, IEICE Transactions on Information and Systems, vol.88, issue.3, pp.503-509, 2005.
DOI : 10.1093/ietisy/e88-d.3.502

L. Yang and C. , Music emotion classification, Proceedings of the 14th annual ACM international conference on Multimedia , MULTIMEDIA '06, 2006.
DOI : 10.1145/1180639.1180665

]. R. Zajonc, Feeling and thinking: Preferences need no inferences., American Psychologist, vol.35, issue.2, pp.151-175, 1980.
DOI : 10.1037/0003-066X.35.2.151

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.651.9447

]. J. Zarader, B. Gas, J. C. Didiot, and P. Sellem, Neural Predictive Coding : application to phoneme recognition, ICONIP, 1997.

]. B. Zellner, Caractérisation du débit de parole en Français, Journées d'Etude sur la Parole, 1998.

]. P. Zhan and M. Westphal, Speaker Normalization Based on Frequency Warping, Proc. ICASSP '97, pp.1039-1042, 1997.

]. Zhu, K. Kuldip, and . Paliwal, Product of Power Spectrum and Group Delay Function for Speech Recognition, ICASSP, 2004.

. Stylisation-hiérarchique, Explications données par l'algorithme 1, p.53