C. Références-bibliographiques, L. Abry, and . Boë, laws " for lips, Speech Communication, vol.5, pp.97-104, 1986.

C. Abry and M. T. Lallouache, Audibility and stability of artiulatory movements. deciphering two experiments on anticipatory rounding in french, XIIth International Congress of Phonetic Sciences, pp.220-225, 1991.

A. Adjoudani, reconnaissance automatique de la parole audiovisuelle, Thèse de doctorat, 1998.

A. Adjoudani and C. Beno??tbeno??t, Speechreading by Humans and Machines, chapitre On the integration of auditory and visual parameters in an HMM-based ASR, pp.461-471, 1996.

A. Adjoudani, T. Guiard-marigny, B. L. Goff, and C. Beno??tbeno??t, Un modèle 3d de l` evres parlantes, Actes des XX?Journées d' ´ Etude sur la Parole (JEP), pp.143-146, 1994.

D. W. Aha, K. Dennis, and A. Marc, Instance-based learning algorithms, Machine Learning, vol.57, issue.1, pp.37-66, 1991.
DOI : 10.1007/BF00153759

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.138.635

J. Alegria, B. Charlier, and S. Mattys, The Role of Lip-reading and Cued Speech in the Processing of Phonological Information in French-educated Deaf Children, European Journal of Cognitive Psychology, vol.36, issue.4, pp.451-472, 1999.
DOI : 10.1121/1.1907309

J. Alegria and J. Lechat, Phonological Processing in Deaf Children: When Lipreading and Cues Are Incongruent, Journal of Deaf Studies and Deaf Education, vol.10, issue.2, pp.122-133, 2005.
DOI : 10.1093/deafed/eni013

J. Alegria and J. Leybaert, L'acquisition du langage par l'enfant sourd : les signes, l'oral et l'´ ecrit, chapitre Le langage par les yeux chez l'enfant sourd : lecture, lecture labiale et langage parlé complété, Marseille : Editions SOLAL, collection Troubles du Développement psychologique et les Apprentissages, pp.213-251, 2005.

J. Alegria, J. Leybaert, B. Charlier, and C. Hage, Analytic approaches to human cognition, chapitre On the origin of phonological representations in the deaf : hearing lips and hands, pp.107-132, 1992.

M. Alissali, P. Deleglise, and A. Rogozan, Asynchronous integration of visual information in an automatic speech recognition system, Proceeding of Fourth International Conference on Spoken Language Processing. ICSLP '96, pp.34-37, 1996.
DOI : 10.1109/ICSLP.1996.607018

URL : https://hal.archives-ouvertes.fr/hal-01437232

R. André-obrecht, B. Jacob, and N. Parlangeau, Audio-visual speech recognition and segmental master-slave hmm, International Conference on Auditory-Visual Speech Processing Septembre, 1997.

O. Aran, T. Burger, A. Caplier, and L. Akarun, Sequential belief-based fusion of manual and non-manual signs, 7th International Workshop on Gesture in Human-Computer Interaction and Simulation, 2007.
URL : https://hal.archives-ouvertes.fr/hal-00327833

V. Attina, La Langue française Parlée Complétée : production et perception, Thèse de doctorat en sciences cognitives, 2005.

V. Attina, D. Beautemps, and M. Cathiard, Coordination of hand and orofacial movements for cv sequences in french cued speech, International Conference on Spoken Langage Processing, pp.1945-1948, 2002.

V. Attina, D. Beautemps, M. A. Cathiard, and M. Odisio, A pilot study of temporal organization in Cued Speech production of French syllables: rules for a Cued Speech synthesizer, Speech Communication, vol.44, issue.1-4, pp.197-214, 2004.
DOI : 10.1016/j.specom.2004.10.013

E. T. Jr, L. E. Auer, and . Bernstein, Enhanced visual speech perception in individuals with early onset hearing impairment, Journal of Speech, Hearing, and Language Research

D. Beautemps, M. Cathiard, and Y. L. Borgne, Benefit of audiovisual presentation in close shadowing task, 15th International Congress of Phonetic Sciences, pp.841-844, 2003.

B. and B. Mosbah, Utilisation de la mémoire de parole pour la reconnaissance : Application pour des personnes handicapées, Thèse de doctorat, 2005.

A. Benguerel and M. K. Pichora-fuller, Coarticulation Effects in Lipreading, Journal of Speech Language and Hearing Research, vol.25, issue.4, pp.600-607, 1982.
DOI : 10.1044/jshr.2504.600

C. Benoit, T. Lallouache, T. Mohamadi, and C. Abry, Talking Machines : Theories, Models and Designs, chapitre A set of French visemes for visual French speech synthesis, pp.485-504, 1992.

C. Benoit, T. Mohamadi, and S. Kandel, Effects of Phonetic Context on Audio-Visual Intelligibility of French, Journal of Speech Language and Hearing Research, vol.37, issue.5, pp.1195-1203, 1994.
DOI : 10.1044/jshr.3705.1195

URL : https://hal.archives-ouvertes.fr/hal-00828874

C. Beno??tbeno??t, T. Guiard-marigny, B. L. Goff, and A. Adjoudani, Speechreading by Humans and Machines, chapitre Which components of the face do humans and machines best speechread ?, pp.315-328, 1996.

K. W. Berger, M. Garner, and J. Sudman, The effect of degree of facial exposure and the vertical angle of vision on speechreading performance. Teacher of the Deaf, pp.322-326, 1971.

L. E. Bernstein, E. T. Jr, P. E. Auer, and . Tucker, Enhanced Speechreading in Deaf Adults, Journal of Speech Language and Hearing Research, vol.44, issue.1, pp.5-18, 2001.
DOI : 10.1044/1092-4388(2001/001)

L. E. Bernstein, M. E. Demorest, and P. E. Tucker, Speech perception without hearing. Perception and Psychophysics, pp.233-252, 2000.
DOI : 10.3758/bf03205546

C. Binnie, P. Jackson, and A. Montgomery, Visual Intelligibility of Consonants: A Lipreading Screening Test with Implications for Aural Rehabilitation, Journal of Speech and Hearing Disorders, vol.41, issue.4, pp.530-539, 1976.
DOI : 10.1044/jshd.4104.530

C. A. Binnie, A. Montgomery, and P. Jackson, Auditory and Visual Contributions to the Perception of Consonants, Journal of Speech Language and Hearing Research, vol.17, issue.4, pp.619-630, 1974.
DOI : 10.1044/jshr.1704.619

I. Bloch, Information combination operators for data fusion: a comparative review with classification, IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans, vol.26, issue.1, pp.52-67, 1996.
DOI : 10.1109/3468.477860

L. D. Braida, Crossmodal integration in the identification of consonant segments, The Quarterly Journal of Experimental Psychology Section A, vol.17, issue.3, pp.647-677, 1991.
DOI : 10.1121/1.1907309

L. D. Braida, M. A. Picheny, J. R. Cohen, W. M. Rabinowitz, and J. S. , Use of articulatory signals in automatic speech recognition, The Journal of the Acoustical Society of America, vol.80, issue.S1, p.18, 1986.
DOI : 10.1121/1.2023685

M. Brand and E. Irfan, Causal analysis for visual gesture understanding, 1995.

M. S. Bratakos, P. Duchnowski, and L. D. Braida, Toward the automatic generation of cued speech, Cued Speech Journal, vol.18, pp.299-320, 1998.

C. Bregler, H. Hild, S. Manke, and A. Waibel, Improving connected letter recognition by lipreading, IEEE International Conference on Acoustics Speech and Signal Processing, pp.557-560, 1993.
DOI : 10.1109/ICASSP.1993.319179

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.455.6690

C. Bregler and Y. Konig, "Eigenlips" for robust speech recognition, Proceedings of ICASSP '94. IEEE International Conference on Acoustics, Speech and Signal Processing, pp.669-672, 1994.
DOI : 10.1109/ICASSP.1994.389567

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.2.5233

R. Brunelli and T. Poggio, Face recognition: features versus templates, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol.15, issue.10, pp.1042-1052, 1993.
DOI : 10.1109/34.254061

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.90.4868

T. Burger, O. Aran, and A. Caplier, Modeling Hesitation and Conflict: A Belief-Based Approach for Multi-class Problems, 2006 5th International Conference on Machine Learning and Applications (ICMLA'06), 2006.
DOI : 10.1109/ICMLA.2006.35

URL : https://hal.archives-ouvertes.fr/hal-00133121

T. Burger, A. Benoit, and A. Caplier, Intercepting static hand gestures in dynamic context, International Conference on Image Processing (ICIP), 2006.
URL : https://hal.archives-ouvertes.fr/hal-00133126

O. Cappé, h2m : A set of matlab functions for the em estimation of hidden markov models with gaussian stateconditional distributions, 2001.

F. Carton, IntroductionàIntroduction`Introductionà la phonétique du Français, 1974.

M. Cathiard, Identification visuelle des voyelles et des consonnes dans le jeu de la protrusion-rétraction des l` evres en français, 1988.

M. Cathiard, La perception visuelle de l'anticipation des gestes vocaliques : cohérence desévénements desévénements audibles et visibles dans le flux de la parole, Thèse de doctorat de psychologie cognitive, 1994.

M. A. Cathiard, V. Attina, C. Abry, and D. Beautemps, La langue française parlée complétée (lpc) : sa coproduction avec la parole et l'organisation temporelle de sa perception. Revue Parole. n?spécial sur " Handicap langagier et recherches cognitives : apports mutuels, pp.31-3231255, 2004.

M. T. Chan, Hmm based audio-visual speech recognition integrating geometric and apparence based visual features, Workshop on multimedia signal processing, pp.9-14, 2001.
DOI : 10.1109/mmsp.2001.962703

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.23.8707

D. Chandramohan and P. L. Silsbee, A multiple deformable template approach for visual speech recognition, Proceeding of Fourth International Conference on Spoken Language Processing. ICSLP '96, 1996.
DOI : 10.1109/ICSLP.1996.607022

B. L. Charlier, C. Hage, J. Alegria, and O. Périer, Evaluation d'une pratique prolongée du la pc sur la comprehension de la parole par l'enfant atteint de déficience auditive, Glossa, vol.22, pp.28-39, 1990.

B. L. Charlier and J. Leybaert, The rhyming skills of deaf children educated with phonetically augmented speechreading, The Quarterly Journal of Experimental Psychology Section A, vol.28, issue.2, pp.349-375, 2000.
DOI : 10.1080/713755898

T. Chen, Audiovisual speech processing, IEEE Signal Processing Magazine, vol.18, issue.1, pp.9-21, 2001.
DOI : 10.1109/79.911195

G. I. Chiou and J. Hwang, Lipreading from color video, IEEE Transactions on Image Processing, vol.6, issue.8, pp.1192-1195, 1997.
DOI : 10.1109/83.605417

B. R. Clarke and D. Ling, The effects of using cued speech : A follow-up study. The Volta Review, pp.23-34, 1976.

T. F. Cootes, G. J. Edwards, and C. J. Taylor, Active appearance models, Lecture Notes in Computer Science, vol.1407, pp.484-498, 1998.
DOI : 10.1007/BFb0054760

T. F. Cootes, C. J. Taylor, D. H. Cooper, and J. Graham, Active Shape Models-Their Training and Application, Computer Vision and Image Understanding, vol.61, issue.1, pp.38-59, 1995.
DOI : 10.1006/cviu.1995.1004

URL : https://www.escholar.manchester.ac.uk/api/datastream?publicationPid=uk-ac-man-scw:1d1862&datastreamId=POST-PEER-REVIEW-PUBLISHERS.PDF

R. O. Cornett, Cued speech, American Annals of the Deaf, vol.112, pp.3-13, 1967.

R. O. Cornett, Annotated bibliography of research in cued speech, Cued Speech Journal, vol.4, pp.1-23, 1990.

S. Cox, I. Matthews, and J. Bangham, Combining noise compensation with visual information in speech recognition, Proceedings of the ESCA Workshop on Audio-Visual Speech Processing, pp.53-56, 1997.

B. Dalton, R. Kaucic, and A. Blake, Speechreading by Man and Machine : Models, Systems and Applications, chapitre Automatic speechreading using dynamic contours, pp.373-382

P. Daubias and P. Deleglise, Statistical Lip-Appearance Models Trained Automatically Using Audio Information, EURASIP Journal on Advances in Signal Processing, vol.2002, issue.11, pp.1202-1212, 2002.
DOI : 10.1155/S1110865702206186

URL : https://hal.archives-ouvertes.fr/hal-01434609

J. Davis and M. Shah, Gesture recognition. Rapport technique CSTR-93-11, 1993.

P. B. Denes, On the Statistics of Spoken English, The Journal of the Acoustical Society of America, vol.35, issue.6, pp.892-904, 1963.
DOI : 10.1121/1.1918622

F. Destombes, Aides manuellesàmanuelles`manuellesà la lecture labiale et perspectives d'aides automatiques, chapitre Le projet VIDVOX, pp.35-36, 1982.

B. Dodd, The Phonological Systems of Deaf Children, Journal of Speech and Hearing Disorders, vol.41, issue.2, pp.185-198, 1976.
DOI : 10.1044/jshd.4102.185

B. Dodd, The Role of Vision in the Perception of Speech, Perception, vol.3, issue.1, pp.31-40, 1977.
DOI : 10.1037/10037-000

B. Dodd, Hearing by Eye : The Psychology of Lipreading, chapitre The acquisition of lipreading skills by normally-hearing children, 1987.

E. R. Dougherty and C. R. Giardina, Image processing -Continous to Discrete

P. Duchnowski, D. Lum, J. Krause, M. Sexton, M. Bratakos et al., Development of speechreading supplements based on automatic speech recognition, IEEE Transactions on Biomedical Engineering, vol.47, issue.4, pp.487-496, 2000.
DOI : 10.1109/10.828148

P. Duchnowski, U. Meuer, and A. Waibel, See me, hear me : integrating automatic speech recognition and lip-reading, International Conference on Spoken Language Processing, pp.547-550, 1994.

S. Dupont and J. Luettin, Audio-visual speech modeling for continuous speech recognition, IEEE Transactions on Multimedia, vol.2, issue.3, pp.141-151, 2000.
DOI : 10.1109/6046.865479

N. P. Erber, Interaction of Audition and Vision in the Recognition of Oral Speech Stimuli, Journal of Speech Language and Hearing Research, vol.12, issue.2, pp.423-425, 1969.
DOI : 10.1044/jshr.1202.423

N. P. Erber, Effects of Angle, Distance, and Illumination on Visual Reception of Speech by Profoundly Deaf Children, Journal of Speech Language and Hearing Research, vol.17, issue.1, pp.99-112, 1974.
DOI : 10.1044/jshr.1701.99

N. Eveno, A. Caplier, and P. Y. Coulon, Jumping snakes and parametric model for lip segmentation, Proceedings 2003 International Conference on Image Processing (Cat. No.03CH37429), pp.867-870, 2003.
DOI : 10.1109/ICIP.2003.1246818

H. W. Ewertsen and H. Birk-nielsen, A Comparative Analysis of the Audiovisual, Auditive and Visual Perception of Speech, Acta Oto-Laryngologica, vol.26, issue.1-6, pp.201-205, 1971.
DOI : 10.1044/jshd.1904.429

C. G. Fisher, Confusions Among Visually Perceived Consonants, Journal of Speech Language and Hearing Research, vol.11, issue.4, pp.796-804, 1968.
DOI : 10.1044/jshr.1104.796

J. Gagné, Visual and audiovisual speech-perception training, Journal of the Academy of Rehabilitative Audiology (Monograph Supplement), vol.27, pp.133-159, 1994.

J. Gagné, V. Materson, K. G. Munhall, N. Bilida, and . Querengesser, Across talker variability in auditory, visual, and audiovisual speech intelligibility for conversational and clear speech, Journal of the Academy of Rehabilitative Audiology, 1995.

C. Garcia and M. Delakis, Convolutional face finder: a neural architecture for fast and robust face detection, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol.26, issue.11, pp.1408-1423, 2004.
DOI : 10.1109/TPAMI.2004.97

G. Gibert, Conception etévaluationetévaluation d'un système de synthèse 3D de Langue françaiseParì ee Complétée (LPC) ` a partir du texte, Thèse de doctorat, 2006.

G. Gibert, G. Bailly, D. Beautemps, F. Elisei, R. Et et al., Analysis and synthesis of the three-dimensional movements of the head, face, and hand of a speaker using cued speech, The Journal of the Acoustical Society of America, vol.118, issue.2, pp.1144-1153, 2005.
DOI : 10.1121/1.1944587

G. Gibert, G. Bailly, and F. Elisei, Evaluating a virtual speech cuer, International Conference on Spoken Langage Processing, 2006.
URL : https://hal.archives-ouvertes.fr/hal-00366491

L. Golipour and D. O. Shaughnessy, A new approach for phoneme segmentation of speech signals, Interspeech'07, 2007.

M. S. Gray, J. R. Movellan, and T. J. Sejnowski, Advances in Neural Information Processing Systems 9, chapitre Dynamic features for visual speech-reading : A systematic comparison, pp.751-757, 1997.

K. Grobel and M. Assan, Isolated sign language recognition using hidden Markov models, 1997 IEEE International Conference on Systems, Man, and Cybernetics. Computational Cybernetics and Simulation, pp.162-167, 1996.
DOI : 10.1109/ICSMC.1997.625742

C. Hage, J. Alegria, and O. Périer, Cued speech and language acquisition : with specifics related to grammatical gender, Cued Speech Journal, vol.4, pp.36-46, 1990.

C. Hage, J. Alegria, and O. Périer, Advances in cognition, education and deafness, chapitre Cued speech and language acquisition : The case of grammatical gender morpho-phonology, pp.395-399, 1991.

A. J. Heap and F. Samaria, Real-time hand tracking and gesture recognition using smart snakes, Proceedings of interface to real and virtual worlds, 1995.

M. Heckmann, F. Bertommier, and K. Kroschel, A hybrid ann/hmm audio-visual speech recognition system, International Conference on Auditory-Visual Speech Processing, pp.190-195, 2001.

F. Heider and G. Heider, An experimental investigation of lip-reading, Psychological Monographs, vol.52, pp.124-153, 1940.

M. Hennecke, K. Prasad, and D. Stork, Using deformable templates to infer visual speech dynamics, Proceedings of 1994 28th Asilomar Conference on Signals, Systems and Computers, 1994.
DOI : 10.1109/ACSSC.1994.471518

S. E. Ohman, Coarticulation in VCV Utterances: Spectrographic Measurements, The Journal of the Acoustical Society of America, vol.39, issue.1, pp.151-168, 1966.
DOI : 10.1121/1.1909864

T. S. Huang, C. P. Hess, H. Pan, and Z. Liang, A neuronet approach to information fusion, Proceedings of First Signal Processing Society Workshop on Multimedia Signal Processing, pp.45-50, 1997.
DOI : 10.1109/MMSP.1997.602611

K. Imagawa, S. Lu, and S. Igi, Color-based hands tracking system for sign language recognition, Proceedings Third IEEE International Conference on Automatic Face and Gesture Recognition, 1998.
DOI : 10.1109/AFGR.1998.670991

P. L. Jackson, A. A. Montgomery, and C. A. Binnie, Perceptual Dimensions Underlying Vowel Lipreading Performance, Journal of Speech Language and Hearing Research, vol.19, issue.4, pp.796-812, 1976.
DOI : 10.1044/jshr.1904.796

P. L. Jackson, The theoretical minimal unit for visual speech perception : Visemes and coarticulation . The Volta Review, pp.11-599, 1988.

B. Jacob and C. Sénac, Un modèle ma??trema??tre-esclave pour la fusion de données acoustiques et articulatoires en reconnaissance, Actes des Journées d' ´ Etude sur la Parole (JEP), pp.363-366, 1996.

P. Jourlin, Handling disynchronization phenomena with hmm in connected speech, Proceedings of European Signal Processing Conference, pp.133-136, 1996.

P. Jourlin, Word dependent acoustic-labial weights in hmm-based speech recognition, Proc. European Tutorial Workshop on Audio-Visual Speech Processing, pp.69-72, 1997.

P. Jourlin, Approche Bimodale du Traitement Automatique de la Parole : applicationàapplication`applicationà la Reconnaissance du Message et du Locuteur, Thèse de doctorat, 1998.

B. Jutras, J. Gagné, M. Picard, and J. Roy, Identification visuelle et catégorisation de consonnes en français québécois. Revues d'orthophonie et d'audiologie, pp.8-87, 1998.

W. Kadous, Machine recognition of auslan signs using powerglove : towards large-lexicon recognition of sign language, The workshop on the Integration of Gesture in Language and Speech, pp.165-174, 1996.

M. Kass, A. Witkins, and . Terzopoulos, Snakes: Active contour models, International Journal of Computer Vision, vol.5, issue.6035, pp.321-331, 1988.
DOI : 10.1007/BF00133570

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.124.5318

E. Kipila, Analysis of an oral language sample from a prelingually child's cued speech : a case study, Cued Speech annual, vol.1, pp.46-59, 1985.

D. H. Klatt, Speech perception : A model of acoustic-phonetic analysis and lexical access, Journal Phonetique, vol.7, pp.279-312, 1979.

J. Kramer and L. Liefer, The talking glove, ACM SIGCAPH Computers and the Physically Handicapped, issue.39, 1989.
DOI : 10.1145/47937.47938

P. B. Kricos and S. A. Lesner, Differences in Visual Intelligibility Across Talkers, Volta Review, vol.84, pp.219-225, 1982.
DOI : 10.1007/978-3-662-13015-5_4

P. B. Kricos and S. A. Lesner, Effect of talker differences on the speechreading of hearingimpaired teenagers. The Volta Review, pp.5-16, 1985.

G. Krone, B. Talle, A. Wichert, and G. Palm, Neural architecture for sensor fusion in speech recognition, ESCA/ESCOP Workshop Audio-Visual and Speech Processing, pp.57-60, 1997.

M. T. Lallouache, Un poste visage-parole. acquisition et traitement des contours labiaux, Journées d'Etude sur la Parole (JEP), Montréa, 1990.

M. Lallouache, Un poste Visage-Parole couleur. Acquisition et traitement automatique des contours des l` evres, Thèse de doctorat, 1991.

R. Lamy, D. Moraru, B. Bigi, and L. Besacier, Premiers pas du clips sur les données d'´ evaluation ester, Proceedings Journées d'Etude, 2004.

A. L. Larr, Speechreading through closed-circuit television, Volta Review, vol.61, pp.19-21, 1959.

J. Laviola, A survey of hand posture and gesture recognition techniques and technology. Rapport technique, 1999.

B. , L. Goff, T. Guiard-marigny, and C. Beno??tbeno??t, Read my lips ... and my jaw ! how intelligible are the components of a speaker's face ?, Eurospeech'95, 1995.

L. Goff, T. Guiard-marigny, and C. Beno??tbeno??t, Progress in Speech Synthesis, chapitre Analysis-synthesis and intelligibility of a talking face, pp.235-246, 1996.

S. A. Lesner, The talker, Volta Review, pp.90-589, 1988.

J. Leybaert, Phonological representations in deaf children: the importance of early linguistic experience, Scandinavian Journal of Psychology, vol.39, issue.3, pp.169-173, 1998.
DOI : 10.1111/1467-9450.393074

J. Leybaert, Phonology Acquired through the Eyes and Spelling in Deaf Children, Journal of Experimental Child Psychology, vol.75, issue.4, pp.291-318, 2000.
DOI : 10.1006/jecp.1999.2539

J. Leybaert and J. Alegria, Spelling development in deaf and hearing children: Evidence for use of morpho-phonological regularities in French, Reading and Writing, vol.45, issue.83, pp.89-109, 1995.
DOI : 10.1007/BF01026949

J. Leybaert and B. L. Charlier, Visual Speech in the Head: The Effect of Cued-Speech on Rhyming, Remembering, and Spelling, Journal of Deaf Studies and Deaf Education, vol.1, issue.4, pp.234-248, 1996.
DOI : 10.1093/oxfordjournals.deafed.a014299

J. Leybaert and J. Lechat, Phonological Similarity Effects in Memory for Serial Order of Cued Speech, Journal of Speech Language and Hearing Research, vol.44, issue.5, pp.949-963, 2001.
DOI : 10.1044/1092-4388(2001/074)

A. M. Liberman and I. G. Mattingly, The motor theory of speech perception revised, Cognition, vol.21, issue.1, pp.1-36, 1985.
DOI : 10.1016/0010-0277(85)90021-6

D. Ling and B. R. Clarke, Cued speech : An evaluative study, American Annals of the deaf, vol.120, pp.480-488, 1975.

N. Liu and B. C. Lovell, Hand Gesture Extraction by Active Shape Models, DICTA ' 05. Proceedings Digital Image Computing: Technqiues and Applications, 2005., pp.6-8, 2005.
DOI : 10.1109/DICTA.2005.1578108

J. Luettin and S. Dupont, Continuous audio-visual speech recognition, 5th European Conf. on Computer Vision, 1998.
DOI : 10.1007/BFb0054771

J. Luettin, N. A. Thacker, and S. W. Beet, Visual speech recognition using active shape models and hidden Markov models, 1996 IEEE International Conference on Acoustics, Speech, and Signal Processing Conference Proceedings, 1996.
DOI : 10.1109/ICASSP.1996.543246

J. Macdonald and H. Mcgurk, Visual influences on speech perception processes, Perception & Psychophysics, vol.17, issue.3, pp.253-257, 1978.
DOI : 10.3758/BF03206096

A. Macleod and Q. Summerfield, Quantifying the contribution of vision to speech perception in noise, British Journal of Audiology, vol.4, issue.2, pp.131-141, 1987.
DOI : 10.1044/jshr.1702.270

A. Macleod and Q. Summerfield, A procedure for measuring auditory and audiovisual speech-reception thresholds for sentences in noise: Rationale, evaluation, and recommendations for use, British Journal of Audiology, vol.2, issue.1, pp.29-43, 1990.
DOI : 10.1037/11774-000

D. J. Mapes and M. J. Moshell, A Two-Handed Interface for Object Manipulation in Virtual Environments, Presence: Teleoperators and Virtual Environments, vol.4, issue.4, pp.403-416, 1995.
DOI : 10.1162/pres.1995.4.4.403

K. Mase and A. P. Pentland, Automatic lipreading by optical-flow analysis, Systems and Computers in Japan, vol.17, issue.4, pp.67-76, 1991.
DOI : 10.1002/scj.4690220607

D. W. Massaro, Categorical Perception : The Groundwork of Cognition, chapitre Categorical partition : a fuzzy logical model of categorization behavior, 1987.

D. W. Massaro, Speechreading by Humans and Machines : Models, Systems, and Applications, chapitre Bimodal speech perception : a progress report, pp.79-101, 1996.

D. W. Massaro, Perceiving talking faces : From speech perception to a behavioral principle, 1998.

D. W. Massaro, M. M. Cohen, and A. T. Gesi, Long-term training, transfer, and retention in learning to lipread, Perception & Psychophysics, vol.8, issue.5, pp.549-562, 1993.
DOI : 10.3758/BF03205203

I. Matthews, J. A. Bangham, R. Harvey, and S. Cox, A comparison of active shape model and scale decomposition based features for visual speech recognition, Workshop on Audio Visual Speech Processing, volume 1407 de Lecture Notes in Computer Science, p.514, 1998.
DOI : 10.1007/BFb0054762

I. Matthews, J. A. Bangham, and S. Cox, Audio-visual speech recognition using multiscale nonlinear image decomposition, International Conference on Spoken Language Processing (ICSLP)'96, 1996.
DOI : 10.1109/icslp.1996.607019

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.16.4278

I. Matthews, G. Potamianos, C. Neti, and J. Luettin, A comparaison of model and transform-based visual features for audio-visual lvcsr, International Conference on Multimedia and Expo, 2001.

H. Mcgurk and J. Macdonald, Hearing lips and seeing voices, Nature, vol.65, issue.5588, pp.746-748, 1976.
DOI : 10.1038/264746a0

S. Mehrotra, H. F. Korth, and A. Silberschatz, Concurrency control in hierarchical multidatabase systems, The VLDB Journal The International Journal on Very Large Data Bases, vol.6, issue.2, pp.152-172, 1997.
DOI : 10.1007/s007780050038

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.36.3355

M. Metzger, First language acquisition in deaf children of hearing parents : Cued English input, Thèse de doctorat, 1994.

T. Mohammed, R. Campbell, M. Macsweeney, E. Milne, P. Hansen et al., Speechreading Skill and Visual Movement Sensitivity are Related in Deaf Speechreaders, Perception, vol.26, issue.2, pp.205-216, 2005.
DOI : 10.1068/p5211

A. A. Montgomery and P. L. Jackson, Physical characteristics of the lips underlying vowel lipreading performance, The Journal of the Acoustical Society of America, vol.73, issue.6, pp.2134-2144, 1983.
DOI : 10.1121/1.389537

A. A. Montgomery, B. E. Walden, and R. A. Prosek, Effects of Consonantal Context on Vowel Lipreading, Journal of Speech Language and Hearing Research, vol.30, issue.1, pp.50-59, 1987.
DOI : 10.1044/jshr.3001.50

L. Mourand-dornier, Le rôle de la lecture labiale dans la reconnaissance de la parole, Thèse de médecine, 1980.

J. R. Movellan and G. Chadderdon, Speechreading by Man and Machine : Models, Systems and Applications, chapitre Channel separability in the audiovisual integration of speech : A Bayesian approach, pp.473-488, 1996.

K. Murakami and H. Taguchi, Gesture recognition using recurrent neural networks, Proceedings of the SIGCHI conference on Human factors in computing systems Reaching through technology, CHI '91, pp.237-242, 1991.
DOI : 10.1145/108844.108900

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.461.3399

S. Nakamura, R. Nagai, and K. Shikano, Adaptative determination of audio and visual weights for automatic speech recognition, Eurospeech, pp.1623-1626, 1997.

Y. Nakano, A study on the factors which influence lipreading of deaf children Language research in countries other than the United States, Volta Review, vol.68, pp.68-83, 1961.

K. K. Neely, Effect of Visual Factors on the Intelligibility of Speech, The Journal of the Acoustical Society of America, vol.28, issue.6, pp.1275-1277, 1956.
DOI : 10.1121/1.1908620

C. C. Neocleous and C. N. Schizas, Neural network, review and critic : Methods and applications of artificial intelligence, 2nd Pan-Hellenic Conference on Artificial Intelligence, pp.300-313, 2002.

C. Neti, G. Iyengar, G. Potamianos, A. Senior, and B. Maison, Perceptual interfaces for information interaction : joint processing of audio and visual information for human-computer interaction, International Conference on Spoken Langage Processing, pp.11-14, 2000.

G. Nicholls and D. Ling, Cued Speech and the Reception of Spoken Language, Journal of Speech Language and Hearing Research, vol.25, issue.2, pp.262-269, 1982.
DOI : 10.1044/jshr.2502.262

S. Ong and S. Ranganath, Automatic Sign Language Analysis: A Survey and the Future beyond Lexical Meaning, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol.27, issue.6, 2005.
DOI : 10.1109/TPAMI.2005.112

E. Owens and B. Blazek, Visemes Observed by Hearing-Impaired and Normal-Hearing Adult Viewers, Journal of Speech Language and Hearing Research, vol.28, issue.3, pp.381-393, 1985.
DOI : 10.1044/jshr.2803.381

E. D. Petajan, Automatic lipreading to enhance speech recognition, Proceedings Global Telecommunications Conference (GLOBCOM)'84, pp.265-272, 1984.
DOI : 10.1145/57167.57170

E. D. Petajan, Automatic lipreading to enhance speech recognition, IEEE Computer, pp.19-23, 1985.
DOI : 10.1145/57167.57170

E. D. Petajan and H. P. Graf, Robust face feature analysis for automatic speechreading and character animation, Proceedings of the Second International Conference on Automatic Face and Gesture Recognition, pp.357-362, 1996.
DOI : 10.1109/AFGR.1996.557291

G. Potamianos, H. P. Graf, and E. Cosatto, An image transform approach for HMM based automatic lipreading, Proceedings 1998 International Conference on Image Processing. ICIP98 (Cat. No.98CB36269), pp.173-177, 1998.
DOI : 10.1109/ICIP.1998.999008

G. Potamianos, C. Neti, G. Gravier, A. Garg, and A. W. Senior, Recent advances in the automatic recognition of audio-visual speech, Proceedings of the IEEE, pp.1306-1326, 2003.

G. Potamianos, C. Neti, G. Iyengar, A. W. Senior, and A. Verma, A cascade visual front end for speaker independent automatic speechreading, International Journal of Speech Technology, vol.4, issue.3/4, pp.193-208, 2001.
DOI : 10.1023/A:1011352422845

G. Potamianos, C. Neti, J. Luettin, and I. Matthews, Audio-Visual Speech Processing, chapitre Audio-Visual Automatic Speech Recognition : An Overview, 2006.

K. Prasad, D. Stork, and G. Wolff, Preprocessing video images for neural learning of lipreading, 1997.

O. Périer, L'enfantàenfantà audition déficiente. Acta oto-rhino-laryng, pp.125-420, 1987.

M. Qutaishat, H. Moussa, A. A. Bayan, and T. Hiba, American sign language (asl) recognition based on hough transform and neural networks, Expert Syst. Appl. Expert systems with Applications, vol.32, issue.321, pp.24-3724, 2007.

B. H. Rabiner and R. Juang, An introduction to hidden Markov models, IEEE ASSP Magazine, vol.3, issue.1, pp.4-16, 1986.
DOI : 10.1109/MASSP.1986.1165342

D. Reisberg, J. Mclean, and A. Goldfield, Hearing by Eye : The Psychology of Lipreading, chapitre Easy to hear but hard to understand : a lipreading advantage with intact auditory stimuli, pp.97-113, 1987.

G. Rizzolatti, M. Luppino, and . Matelli, Supplementary sensorimotor area, chapitre The classic supplementary motor area is formed by two independent areas, pp.45-56, 1996.

J. Robert-ribes, M. Piquemal, J. L. Schwartz, and P. Escudier, Speechreading by Man and Machine : Models, Systems and Applications, chapitre Exploiting sensor fusion architectures and stimuli complementarity in AV speech recognition, pp.193-210, 1996.
DOI : 10.1007/978-3-662-13015-5_14

J. Robert-ribès, Modèles d'intégration audiovisuelle de signaux linguistigues : de la perception humainè a la reconnaissance automatique des voyelles, Thèse de doctorat, 1995.

J. Robert-ribès, J. Schwartz, M. T. Lallouache, and P. Escudier, Complementarity and synergy in bimodal speech: Auditory, visual, and audio-visual identification of French oral vowels in noise, The Journal of the Acoustical Society of America, vol.103, issue.6, pp.3677-3689, 1998.
DOI : 10.1121/1.423069

A. Rogozan and P. Deléglise, Adaptive fusion of acoustic and visual sources for automatic speech recognition, Speech Communication, vol.26, issue.1-2, pp.149-161, 1998.
DOI : 10.1016/S0167-6393(98)00056-9

URL : https://hal.archives-ouvertes.fr/hal-01437163

A. Rogozan, P. Deléglise, and M. Alissali, Adaptive determination of audio and visual weights for automatic speech recognition, Proceedings European Tutorial Research Workshop on Audio-Visual Speech Processing (AVSP)'97, pp.61-64, 1997.
URL : https://hal.archives-ouvertes.fr/hal-01437207

D. Rubine, Specifying gestures by example, Proceedings of SIGGRAPH'91, pp.329-337, 1991.
DOI : 10.1145/122718.122753

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.620.3598

J. Schwartz, Traitement automatique du langage parlé 2 : reconnaissance de la parole, chapitre La parole multimodale :deux ou trois sens valent mieux qu'un, pp.141-178, 2002.

J. Schwartz, La parole multisensorielle : Plaidoyer,probì emes et perscpectives, Actes des XXVèmes Journées d'Etude sur la Parole (JEP), pp.11-17, 2004.

J. Schwartz, J. Robert-ribès, and P. Escudier, Hearing by Eye II : Advances in the Psychology of Speechreading and Auditory-Visual Speech, chapitre Ten years after Summerfield : A taxonomy of models for audio-visual fusion in speech perception, pp.85-108, 1998.

P. L. Silsbee, Motion in deformable templates, Proceedings of 1st International Conference on Image Processing, pp.323-327, 1994.
DOI : 10.1109/ICIP.1994.413328

P. L. Silsbee and Q. Su, NATO ASI : Speechreading by Humans and Machines, chapitre Audiovisual sensory integration using hidden Markovmodels, pp.489-495, 1996.
DOI : 10.1007/978-3-662-13015-5_37

T. Starner, Visual recognition of american sign language using hidden markov models

T. Starner and A. Pentland, Real-time american sign language recgnition from video using hidden markov models. Perceptual Computing Section N?375, 1996.
DOI : 10.1007/978-94-015-8935-2_10

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.153.6485

D. G. Stork, G. Wolff, and E. Levine, Neural network lipreading system for improved speech recognition, [Proceedings 1992] IJCNN International Joint Conference on Neural Networks, pp.285-295, 1992.
DOI : 10.1109/IJCNN.1992.226994

D. Sturman, Whole-hand Input, Thèse de doctorat, Massachusetts Institute of Technology, 1992.
DOI : 10.1117/12.164909

D. Sturman and D. Zeltzer, A survey of glove-based input, IEEE Computer Graphics and Applications, vol.14, issue.1, pp.30-39, 1994.
DOI : 10.1109/38.250916

Q. Su and P. L. Silsbee, Robust audiovisual integration using semicontinuous hidden markov models, International Conference on Spoken Language Pocessing, pp.42-45, 1996.

W. H. Sumby and I. Pollack, Visual Contribution to Speech Intelligibility in Noise, The Journal of the Acoustical Society of America, vol.26, issue.2, pp.212-215, 1954.
DOI : 10.1121/1.1907309

Q. Summerfield, Use of Visual Information for Phonetic Perception, Phonetica, vol.36, issue.4-5, pp.314-331, 1979.
DOI : 10.1159/000259969

Q. Summerfield, Audio-visual speech perception, lipreading and artificial stimulation. Hearing Science and Hearing Disorders, pp.131-182, 1983.
DOI : 10.1016/b978-0-12-460440-7.50010-7

Q. Summerfield, Hearing by Eye : The Psychology of Lipreading, chapitre Some preliminaries to a comprehensive account of audio-visual speech perception, pp.3-51, 1987.

Q. Summerfield, A. Macleod, M. Mcgrath, and M. Brooke, Handbook of Research on Face Processing, chapitre Lips, teeth, and the benefits of lipreading, pp.223-233, 1989.

P. Teissier, J. Robert-ribès, and J. Schwartz, Comparing models for audiovisual fusion in a noisy-vowel recognition task, IEEE Transactions on Speech and Audio Processing, vol.7, issue.6, pp.629-642, 1999.
DOI : 10.1109/89.799688

A. Tseva, L'arrondissement dans l'identification visuelle des voyelles du français. premiers acquis, pp.149-186, 1989.

A. Tseva and M. Cathiard, PAROLES VUES : LA DIMENSION D'ARRONDISSEMENT DANS L'IDENTIFICATION VISUELLE DES VOYELLES DU FRAN??AIS, Actes du 1er Congré Français d'Acoustique, Colloque de physique, Colloque C2, pp.507-510, 1990.
DOI : 10.1051/jphyscol:19902119

R. M. Uchanski, L. A. Delhorme, A. K. Dix, L. D. Braida, C. M. Reed et al., Automatic speech recognition to aid the hearing impaired : Prospects for the automatic generation of cued speech, Journal of Rehabilitation Research and Development, issue.1, pp.3120-3161, 1994.

D. Vaufreydaz, J. Bergamini, J. F. Serignat, L. Besacier, and M. Akbar, A new methodology for speech corpora definition from internet documents, Proceedings 2nd International Conference on Language Resources and Evaluation (LREC2000), pp.423-426, 2000.
URL : https://hal.archives-ouvertes.fr/inria-00326150

B. Walden, R. Prosek, A. Mongomery, C. Scherr, and C. Jones, Effects of Training on the Visual Recognition of Consonants, Journal of Speech Language and Hearing Research, vol.20, issue.1, pp.135-145, 1977.
DOI : 10.1044/jshr.2001.130

B. E. Walden, S. A. Erdman, A. A. Montgomery, D. M. Schwartz, and R. A. Prosek, Some Effects of Training on Speech Recognition by Hearing-Impaired Adults, Journal of Speech Language and Hearing Research, vol.24, issue.2, pp.207-216, 1981.
DOI : 10.1044/jshr.2402.207

T. Watanabe and M. Kohda, Lip-reading of japanese vowels using neural networks, International Conference Spoken Language Procesing, pp.1373-1376, 1990.

R. Watson, A survey of gesture recognition techniques, 1993.

A. Wexelblat, An approach to natural gesture in virtual environments, ACM Transactions on Computer-Human Interaction, vol.2, issue.3, pp.179-200, 1995.
DOI : 10.1145/210079.210080

M. F. Woodward and C. G. Barber, Phoneme Perception in Lipreading, Journal of Speech Language and Hearing Research, vol.3, issue.3, pp.212-222, 1960.
DOI : 10.1044/jshr.0303.212

W. Wouts, Aides manuellesàmanuelles`manuellesà la lecture labiale et perspectives d'aides automatiques, chapitre L'AKA, pp.16-29, 1982.

M. H. Yang, N. Ahuja, and M. Tabb, Extraction of 2D motion trajectories and its application to hand gesture recognition, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol.24, issue.8, pp.1061-1074, 2002.
DOI : 10.1109/TPAMI.2002.1023803

H. Yehia, P. Rubin, and E. Vatikiotis-bateson, Quantitative association of orofacial and vocal-tract shapes, International Conference on Audio-Visual Speech Processing, pp.41-44, 1997.

B. P. Yuhas, M. H. Goldstein, and T. J. Sejnowski, Integration of acoustic and visual speech signals using neural networks, IEEE Communications Magazine, vol.27, issue.11, pp.65-71, 1989.
DOI : 10.1109/35.41402

B. P. Yuhas, M. H. Goldstein-jr, T. J. Sejnowski, and R. E. Jenkins, Neural network models of sensory integration for improved vowel recognition, Proc. IEEE, pp.781658-1667, 1990.
DOI : 10.1109/5.58349

A. L. Yuille, P. W. Hallinan, and D. S. Cohen, Feature extraction from faces using deformable templates, International Journal of Computer Vision, vol.26, issue.6, pp.99-111, 1992.
DOI : 10.1007/BF00127169

D. Liste-des-publications-@bullet-noureddine-aboutabit, &. Beautemps, and . Laurent-besacier, Hand and Lips desynchronization analysis in French Cued Speech : Automatic segmentation of Hand flow, Proceedings of ICASSP'06, 2006.

D. @bullet-noureddine-aboutabit, &. Beautemps, and . Laurent-besacier, Characterization of Cued Speech vowels from the inner lip contour, Proceedings of ICSLP'06, 2006.

D. @bullet-noureddine-aboutabit, &. Beautemps, and . Laurent-besacier, Vowel classification from lips : the Cued Speech production case, proceeding of International Seminar on Speech Production (ISSP), pp.127-134, 2006.

D. @bullet-noureddine-aboutabit, &. Beautemps, and . Laurent-besacier, Automatic identification of vowels in the Cued Speech context, International Conference on Auditory-Visual Speech Processing (AVSP), 2007.

D. @bullet-noureddine-aboutabit, J. Beautemps, L. Clarke, and . Besacier, A HMM recognition of consonant-vowel syllabes from lip contours : the Cued Speech case, Interspeech, 2007.

D. @bullet-noureddine-aboutabit, &. Beautemps, and . Laurent-besacier, Lips and Hand Modeling for Recognition of the Cued Speech Gestures : The French Vowel Case, Speech Communication

. Pourcentages-moyens-obtenus-par-alegria, dans son expérience en perception de mots et de pseudo-mots en condition de lecture labiale seule ou avec les clés, p.35, 1999.

. Attina, Image du codeur LPC avec les axes x et y en superposition, p.38, 2004.

. De-haut-en-basattina, ) trajectoires x (cm) et (2) y (cm) de la main pour une séquence [pupøpu], (3) décours temporel de l'aire intérolabiale S (cm 2 ) ; (4) signal acoustique correspondant, p.39, 2004.

. Attina, Schéma général de coordination de la main et des l` evres en relation avec le son de parole pour le code LPC, p.39, 2004.

/. Image-du-gant, 5DT Data Glove 16 MRI Ce gant ne contient aucune partie métallique ou magnétique et il est connecté au bo??tierbo??tier d'interface par un ruban long de 5-7m via une fibre optique (image copiée du site internet de Fifth Dimension Technologies (5DT), p.58

. Structure-d-'un-neurone, Le neurone calcule la somme pondérée de ses entrées puis cette valeur passè a travers la fonction de seuillage pour produire sa sortie, p.66

L. and D. Schwartz, un processus d'intégration audio-visuelle dans la perception de la parole (d'après, p.76, 1998.

A. , A. ?. , B. , B. ?-en-cm-ainsi-que, S. et al., cm 2 calculéscalculésà l'aide du logiciel TACLEàTACLEà partir de la séquence : " ma chemise est roussie " . Les erreurs d'extraction des paramètres sont par ailleurs marquées, p.115

A. ?. , B. , B. ?-en-cm-ainsi-que, S. , and S. , en cm 2 calculéscalculésà l'aide de la solution que nous proposons appliquéeappliquéè a la séquence : " ma chemise est roussie ", p.117

. De-haut-en-bas, les paramètres du contour interne des l` evres (A, B et S), les coordonnées x et y de la pastille haute, la réalisation acoustique, p.119

. Sur-cette-figure-le-terme-"............, clé " signifie " configuration, p.138

. Effets-de-la-coarticulation-sur-la, une syllabe CV caractérisée par des paramètres labiaux différents d'un contextè a l'autre. En haut : la consonne [t] de la syllabe [ta] est modifiée par le contexte avant ( ` a gauche [? O] etàetà droite [a]), ce qui donne des allures différentes des paramètres labiaux (ici du contour interne) de la syllabe entre les instants A1 et A3. En bas : la syllabe [pa] est moins influencée par le contexte avant, Notons dans ces figures la variabilité des instants acoustiques par rapport au début et la fin de la transition labiale d'une syllabe. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 156

.. Identification-de-la-voyelle, la main en premier, ensuite les l` evres " pour le schéma de fusion des informations labiale et manuelle s'appuyant sur un modèle ma??trema??treesclave, p.179

L. Cued-speech-et-historique and .. , Définition, p.28