[. I. Publication, M. Sehili, F. Yang, V. Leynaert, and L. Devillers, A corpus of social interaction between nao and elderly people, 5th International Workshop on Emotion, Social Signals, 2014.

[. Ii, M. Sehili, F. Yang, and L. Devillers, Attention detection in elderly people-robot spoken interaction, Proceedings of the 2014 Workshop on Multimodal, Multi-Party, Real-World Human-Robot Interaction, pp.7-12, 2014.

[. Iii, F. Yang, M. A. Sehili, C. Barras, and L. Devillers, Smile and Laughter Detection for Elderly People-Robot Interaction, Social Robotics, pp.694-703, 2015.

[. Iv and L. Devillers, Multimodal data collection of humanrobot humorous interactions in the joker project, 6th International Conference on Affective Computing and Intelligent Interaction (ACII), 2015.

A. Ahonen, M. P. Hadid, R. Timur, M. Almaev, R. Francis et al., Face description with local binary patterns: Application to face recognition. Pattern Analysis and Machine Intelligence Recognition of blurred faces using local phase quantization Local gabor binary patterns from three orthogonal planes for automatic facial expression recognition Multiple kernel learning, conic duality, and the smo algorithm Emotion analysis in manmachine interaction systems, A HMM recognition of consonant-vowel syllables from lip contours: the Cued Speech case Esa Rahtu, Ville Ojansivu, and Janne Heikkilä Pattern Recognition 19th International Conference on Affective Computing and Intelligent Interaction (ACII), 2013 Humaine Association Conference on Keith Anderson and Peter W McOwan. A real-time automated system for the recognition of human facial expressions. Systems, Man, and Cybernetics, Part B: Cybernetics Proceedings of the twenty-first international conference on Machine learning Machine learning for multimodal interaction Tanja Bänziger and Klaus R Scherer. Introducing the geneva multimodal emotion portrayal (gemep) corpus. Blueprint for affective computing: A sourcebook, pp.2037-2041, 1998.

M. Stewart-bartlett, G. Littlewort, M. Frank, C. Lainscsek, I. Fasel et al., Recognizing Facial Expression: Machine Learning and Application to Spontaneous Behavior, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05), pp.568-573, 2005.
DOI : 10.1109/CVPR.2005.297

M. Stewart-bartlett, G. Littlewort, M. Frank, C. Lainscsek, I. Fasel et al., Fully Automatic Facial Action Recognition in Spontaneous Behavior, 7th International Conference on Automatic Face and Gesture Recognition (FGR06), pp.223-230, 2006.
DOI : 10.1109/FGR.2006.55

S. Basu, T. Choudhury, B. Clarkson, and A. Pentland, Towards measuring human interactions in conversational settings, Proc. IEEE CVPR Workshop on Cues in Communication, 2001.

A. Batliner, C. Hacker, S. Steidl, E. Nöth, S. D. Arcy et al., you stupid tin box"-children interacting with the aibo robot: A cross-linguistic emotional speech corpus, LREC, 2004.

A. Battocchi, F. Pianesi, and D. Goren-bar, DaFEx: Database of Facial Expressions, Intelligent Technologies for Interactive Entertainment, pp.303-306, 2005.
DOI : 10.1007/11590323_39

Y. Bengio, R. Ducharme, P. Vincent, and C. Janvin, Neural Probabilistic Language Models, The Journal of Machine Learning Research, vol.3, pp.1137-1155, 2003.
DOI : 10.1007/3-540-33486-6_6

URL : https://hal.archives-ouvertes.fr/hal-01434258

D. Bernhardt and P. Robinson, Detecting Emotions from Connected Action Sequences, Visual Informatics: Bridging Research and Practice, pp.1-11
DOI : 10.1007/978-3-642-05036-7_1

C. Breazeal, Robot in society: friend or appliance, Proceedings of the 1999 Autonomous Agents Workshop on Emotion-Based Agent Architectures, pp.18-26, 1999.

H. Bredin and G. Chollet, Making talking-face authentication robust to deliberate imposture, 2008 IEEE International Conference on Acoustics, Speech and Signal Processing, pp.1693-1696, 2008.
DOI : 10.1109/ICASSP.2008.4517954

A. Buendia and L. Devillers, From Informative Cooperative Dialogues to Long-Term Social Relation with a Robot, Natural Interaction with Robots, Knowbots and Smartphones, pp.135-151, 2014.
DOI : 10.1007/978-1-4614-8280-2_13

URL : https://hal.archives-ouvertes.fr/hal-01126424

C. Busso, Z. Deng, S. Yildirim, M. Bulut, M. Lee et al., Analysis of emotion recognition using facial expressions, speech and multimodal information, Proceedings of the 6th international conference on Multimodal interfaces , ICMI '04, pp.205-211, 2004.
DOI : 10.1145/1027933.1027968

N. Campbell, Accounting for voice-quality variation, Speech Prosody 2004, International Conference, 2004.

G. Caridakis, L. Malatesta, L. Kessous, N. Amir, A. Raouzaiou et al., Modeling naturalistic affective states via facial and vocal expressions recognition, Proceedings of the 8th international conference on Multimodal interfaces , ICMI '06, pp.146-154, 2006.
DOI : 10.1145/1180995.1181029

G. Castellano, I. Leite, A. Pereira, C. Martinho, A. Paiva et al., Affect recognition for interactive companions: challenges and??design in real world scenarios, Journal on Multimodal User Interfaces, vol.39, issue.4, pp.89-98, 2010.
DOI : 10.1007/s12193-009-0033-5

C. Chang and C. Lin, LIBSVM, ACM Transactions on Intelligent Systems and Technology, vol.2, issue.3, p.27, 2011.
DOI : 10.1145/1961189.1961199

Y. Chang, C. Hu, R. Feris, and M. Turk, Manifold based analysis of facial expression, Image and Vision Computing, vol.24, issue.6, pp.605-614, 2006.
DOI : 10.1016/j.imavis.2005.08.006

Y. Chang, C. Hu, and M. Turk, Probabilistic expression analysis on manifolds, Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004., p.520, 2004.
DOI : 10.1109/CVPR.2004.1315208

Y. Chang, M. Vieira, M. Turk, and L. Velho, Automatic 3D Facial Expression Analysis in Videos, Analysis and Modelling of Faces and Gestures, pp.293-307, 2005.
DOI : 10.1007/11564386_23

C. Clavel, I. Vasilescu, L. Devillers, G. Richard, and T. Ehrette, Fear-type emotion recognition for future audio-based surveillance systems, Speech Communication, vol.50, issue.6, pp.487-503, 2008.
DOI : 10.1016/j.specom.2008.03.012

URL : https://hal.archives-ouvertes.fr/hal-00499211

I. Cohen, N. Sebe, A. Garg, S. Lawrence, . Chen et al., Facial expression recognition from video sequences: temporal and static modeling, Computer Vision and Image Understanding, vol.91, issue.1-2, pp.160-187, 2003.
DOI : 10.1016/S1077-3142(03)00081-X

F. Jeffrey and . Cohn, Foundations of human computing: facial expression and emotion, Proceedings of the 8th international conference on Multimodal interfaces, pp.233-238, 2006.

F. Jeffrey, L. I. Cohn, Z. Reed, J. Ambadar, T. Xiao et al., Automatic analysis and recognition of brow actions and head motion in spontaneous facial behavior, Systems, Man and Cybernetics IEEE International Conference on, pp.610-616, 2004.

F. Jeffrey, . Cohn, L. Karen, and . Schmidt, The timing of facial motion in posed and spontaneous smiles, Multiresolution and Information Processing, pp.121-132, 2004.

F. Timothy, . Cootes, J. Christopher, and . Taylor, Active shape models ???? mart snakes ?? In BMVC92, pp.266-275, 1992.

F. Timothy, . Cootes, J. Christopher, . Taylor, H. David et al., Active shape models-their training and application. Computer vision and image understanding, pp.38-59, 1995.

C. Cortes and V. Vapnik, Support-vector networks, Machine Learning, vol.1, issue.3, pp.273-297, 1995.
DOI : 10.1007/BF00994018

R. Cowie, E. Douglas-cowie, N. Tsapatsoulis, G. Votsis, S. Kollias et al., Emotion recognition in human-computer interaction, IEEE Signal Processing Magazine, vol.18, issue.1, pp.32-80, 2001.
DOI : 10.1109/79.911197

A. Delaborde and L. Devillers, Use of nonverbal speech cues in social interaction between human and robot, Proceedings of the 3rd international workshop on Affective interaction in natural environments, AFFINE '10, pp.75-80, 2010.
DOI : 10.1145/1877826.1877846

O. Déniz, . Castrillon, . Lorenzo, G. Anton, and . Bueno, Smile Detection for User Interfaces, Advances in Visual Computing, pp.602-611
DOI : 10.1109/34.895976

L. Devillers, L. Vidrascu, and L. Lamel, Challenges in real-life emotion annotation and machine learning based detection, Neural Networks, vol.18, issue.4, pp.407-422, 2005.
DOI : 10.1016/j.neunet.2005.03.007

L. Devillers, L. Vidrascu, and O. Layachi, Automatic detection of emotion from vocal expression. A Blueprint for an Affectively Competent Agent, Cross-Fertilization Between Emotion Psychology, Affective Neuroscience, and Affective Computing, pp.232-244, 2010.

A. Dhall, A. Asthana, R. Goecke, and T. Gedeon, Emotion Université Paris-Saclay Espace Technologique / Immeuble Discovery recognition using phog and lpq features, Automatic Face & Gesture Recognition and Workshops 2011 IEEE International Conference on, pp.878-883, 2011.
DOI : 10.1109/fg.2011.5771366

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.654.1887

A. Dhall, Collecting Large, Richly Annotated Facial-Expression Databases from Movies, IEEE MultiMedia, vol.19, issue.3, 2012.
DOI : 10.1109/MMUL.2012.26

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.407.4632

A. Dhall, R. Goecke, J. Joshi, K. Sikka, and T. Gedeon, Emotion Recognition In The Wild Challenge 2014, Proceedings of the 16th International Conference on Multimodal Interaction, ICMI '14, pp.461-466, 2014.
DOI : 10.1145/2663204.2666275

J. Gareth, . Edwards, J. Christopher, T. F. Taylor, and . Cootes, Interpreting face images using active appearance models, Automatic Face and Gesture Recognition Proceedings. Third IEEE International Conference on, pp.300-305, 1998.

P. Ekman, J. Friesen, and . Hager, Facs manual. A Human Face, 2002.

P. Ekman, V. Wallace, M. O. Friesen, A. Sullivan, I. Chan et al., Universals and cultural differences in the judgments of facial expressions of emotion., Journal of Personality and Social Psychology, vol.53, issue.4, p.712, 1987.
DOI : 10.1037/0022-3514.53.4.712

G. Fanelli, T. Weise, J. Gall, and L. Van-gool, Real Time Head Pose Estimation from Consumer Depth Cameras, Pattern Recognition, pp.101-110, 2011.
DOI : 10.1007/978-3-642-23123-0_11

M. Farrús, J. Hernando, and P. Ejarque, Jitter and shimmer measurements for speaker recognition, INTERSPEECH, pp.778-781, 2007.

B. Fasel and J. Luettin, Automatic facial expression analysis: a survey, Pattern Recognition, vol.36, issue.1, pp.259-275, 2003.
DOI : 10.1016/S0031-3203(02)00052-3

X. Feng, M. Pietikäinen, and A. Hadid, Facial expression recognition based on local binary patterns, Pattern Recognition and Image Analysis, vol.17, issue.4, pp.592-598, 2007.
DOI : 10.1134/S1054661807040190

H. Go, K. Kwak, D. Lee, and M. Chun, Emotion recognition from the facial image and speech signal, SICE 2003 Annual Conference, pp.2890-2895, 2003.

D. Samuel, . Gosling, J. Peter, . Rentfrow, B. William et al., A very brief measure of the big-five personality domains, Journal of Research in personality, vol.37, issue.6, pp.504-528, 2003.

N. Gourier, D. Hall, L. James, and . Crowley, Estimating face orientation from robust detection of salient facial structures, FG Net Workshop on Visual Observation of Deictic Gestures, pp.1-9

H. Gunes and M. Piccardi, Affect Recognition from Face and Body: Early Fusion vs. Late Fusion, 2005 IEEE International Conference on Systems, Man and Cybernetics, pp.3437-3443, 2005.
DOI : 10.1109/ICSMC.2005.1571679

URL : https://opus.lib.uts.edu.au/bitstream/10453/2745/3/2005003128.pdf

G. Guo, R. Charles, and . Dyer, Learning From Examples in the Small Sample Case: Face Expression Recognition, IEEE Transactions on Systems, Man and Cybernetics, Part B (Cybernetics), vol.35, issue.3, pp.477-488, 2005.
DOI : 10.1109/TSMCB.2005.846658

S. Hoch, F. Althoff, G. Mcglaun, and G. Rigoll, Bimodal Fusion of Emotional Data in an Automotive Environment, Proceedings. (ICASSP '05). IEEE International Conference on Acoustics, Speech, and Signal Processing, 2005., p.1085, 2005.
DOI : 10.1109/ICASSP.2005.1415597

S. Kohsia, . Huang, M. Mohan, T. Trivedi, and . Gandhi, Driver's view and vehicle surround estimation using omnidirectional video stream, Intelligent Vehicles Symposium, pp.444-449, 2003.

V. Spiros, . Ioannou, T. Amaryllis, . Raouzaiou, A. Vasilis et al., Emotion recognition through facial expression analysis based on a neurofuzzy network, Neural Networks, vol.18, issue.4, pp.423-435, 2005.

A. Ito, X. Wang, M. Suzuki, and S. Makino, Smile and laughter recognition using speech processing and face recognition from conversation video, 2005 International Conference on Cyberworlds (CW'05), p.8, 2005.
DOI : 10.1109/CW.2005.82

V. Jain, L. James, and . Crowley, Head Pose Estimation Using Multi-scale Gaussian Derivatives, Image Analysis, pp.319-328, 2013.
DOI : 10.1007/978-3-642-38886-6_31

URL : https://hal.archives-ouvertes.fr/hal-00839527

Q. Ji, P. Lan, and C. Looney, A probabilistic framework for modeling and real-time monitoring human fatigue. Systems, Man and Cybernetics, Part A: Systems and Humans, IEEE Transactions on, vol.36, issue.5, pp.862-875, 2006.

B. Jiang, F. Michel, M. Valstar, and . Pantic, Action unit detection using sparse appearance descriptors in space-time video volumes, Face and Gesture 2011, pp.314-321, 2011.
DOI : 10.1109/FG.2011.5771416

M. Johansson, G. Skantze, and J. Gustafson, Head Pose Patterns in Multiparty Human-Robot Team-Building Interactions, Social Robotics, pp.351-360, 2013.
DOI : 10.1007/978-3-319-02675-6_35

I. Jolliffe, Principal component analysis, 2002.
DOI : 10.1007/978-1-4757-1904-8

T. Kanade, F. Jeffrey, Y. Cohn, and . Tian, Comprehensive database for facial expression analysis, Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580), pp.46-53, 2000.
DOI : 10.1109/AFGR.2000.840611

A. Kapoor, W. Burleson, and R. W. Picard, Automatic prediction of frustration, International Journal of Human-Computer Studies, vol.65, issue.8, pp.724-736, 2007.
DOI : 10.1016/j.ijhcs.2007.02.003

K. Karpouzis, G. Caridakis, L. Kessous, N. Amir, A. Raouzaiou et al., Modeling Naturalistic Affective States Via Facial, Vocal, and Bodily Expressions Recognition, Artifical intelligence for human computing, pp.91-112, 2007.
DOI : 10.1007/978-3-540-72348-6_5

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.148.4428

M. Katzenmaier, R. Stiefelhagen, and T. Schultz, Identifying the addressee in human-human-robot interactions based on head pose and speech, Proceedings of the 6th international conference on Multimodal interfaces , ICMI '04, pp.144-151, 2004.
DOI : 10.1145/1027933.1027959

S. Lyndon, . Kennedy, P. Daniel, and . Ellis, Laughter detection in meetings, NIST ICASSP 2004 Meeting Recognition Workshop National Institute of Standards and Technology, pp.118-121, 2004.

M. Kipp, Anvil: The video annotation research tool. Handbook of Corpus Phonology, 2010.
DOI : 10.1093/oxfordhb/9780199571932.013.024

L. Chris, . Kleinke, A. Armando, . Bustos, B. Frederick et al., Effects of self-attributed and other-attributed gaze on interpersonal evaluations between males and females, Journal of experimental social Psychology, vol.9, issue.2, pp.154-163, 1973.

U. Kowalik, T. Aoki, and H. Yasuda, Broaference?a next generation multimedia terminal providing direct feedback on audience ? ? satisfaction level, Human-Computer Interaction-INTERACT 2005, pp.974-977

R. Stephen, H. Langton, E. Honeyman, and . Tessler, The influence of head contour and nose angle on the perception of eye-gaze direction, Perception & psychophysics, vol.66, issue.5, pp.752-771, 2004.

G. Littlewort, J. Whitehill, T. Wu, I. Fasel, M. Frank et al., The computer expression recognition toolbox (CERT), Face and Gesture 2011
DOI : 10.1109/FG.2011.5771414

C. Gwen, M. S. Littlewort, K. Bartlett, and . Lee, Faces of pain: automated measurement of spontaneousallfacial expressions of genuine and posed pain, Proceedings of the 9th international conference on Multimodal interfaces, pp.15-21, 2007.

S. Lucey, A. B. Ashraf, F. Jeffrey, and . Cohn, Investigating Spontaneous Facial Action Recognition through AAM Representations of the Face, 2007.
DOI : 10.5772/4841

M. Lyons, S. Akamatsu, M. Kamachi, and J. Gyoba, Coding facial expressions with Gabor wavelets, Proceedings Third IEEE International Conference on Automatic Face and Gesture Recognition, pp.200-205, 1998.
DOI : 10.1109/AFGR.1998.670949

M. Maazaoui, K. Abed-meraim, and Y. Grenier, Blind source separation for robot audition using fixed HRTF beamforming, EURASIP Journal on Advances in Signal Processing, vol.2012, issue.1, pp.1-18, 2012.
DOI : 10.1162/089976698300017746

URL : https://hal.archives-ouvertes.fr/hal-00683198

A. Macleod and Q. Summerfield, Quantifying the contribution of vision to speech perception in noise, British Journal of Audiology, vol.4, issue.2, pp.131-141, 1987.
DOI : 10.1044/jshr.1702.270

P. Paul, T. Maglio, . Matlock, S. Christopher, S. Campbell et al., Gaze and speech in attentive user interfaces, Advances in Multimodal Interfaces ?? CMI 2000, pp.1-7, 2000.

M. Mancini, L. Ach, E. Bantegnie, T. Baur, N. Berthouze et al., Laugh when you ?? e winning, Innovative and Creative Developments in Multimodal Interaction Systems, pp.50-79, 2014.

I. Mccowan, D. Gatica-perez, S. Bengio, G. Lathoud, M. Barnard et al., Automatic analysis of multimodal group actions in meetings. Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol.27, issue.3, pp.305-317, 2005.

G. Mckeown, M. Valstar, R. Cowie, M. Pantic, and M. Schröder, The SEMAINE Database: Annotated Multimodal Records of Emotionally Colored Conversations between a Person and a Limited Agent, IEEE Transactions on Affective Computing, vol.3, issue.1, pp.5-17, 2012.
DOI : 10.1109/T-AFFC.2011.20

A. Metallinou, C. Busso, S. Lee, and S. Narayanan, Visual emotion recognition using compact facial representations and viseme information, 2010 IEEE International Conference on Acoustics, Speech and Signal Processing, pp.2474-2477, 2010.
DOI : 10.1109/ICASSP.2010.5494893

S. Moore and R. Bowden, Local binary patterns for multi-view facial expression recognition, Computer Vision and Image Understanding, vol.115, issue.4, pp.541-558, 2011.
DOI : 10.1016/j.cviu.2010.12.001

E. Murphy-chutorian and M. Trivedi, Head pose estimation in computer vision: A survey. Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol.31, issue.4, pp.607-626, 2009.

S. Dupont, Laugh-aware virtual agent and its impact on user amusement, Proceedings of the 2013 international conference on Autonomous agents and multi-agent systems International Foundation for Autonomous Agents and Multiagent Systems, pp.619-626, 2013.
URL : https://hal.archives-ouvertes.fr/hal-00869751

J. Odobez and S. Ba, A Cognitive and Unsupervised Map Adaptation Approach to the Recognition of the Focus of Attention from Head Pose, Multimedia and Expo, 2007 IEEE International Conference on, pp.1379-1382, 2007.
DOI : 10.1109/ICME.2007.4284916

T. Ojala, M. Pietikäinen, and D. Harwood, A comparative study of texture measures with classification based on featured distributions, Pattern Recognition, vol.29, issue.1, pp.51-59, 1996.
DOI : 10.1016/0031-3203(95)00067-4

T. Ojala, M. Pietikäinen, and T. Mäenpää, Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol.24, issue.7, pp.971-987, 2002.

J. Panksepp, Affective neuroscience: The foundations of human and animal emotions, 1998.

M. Pantic and I. Patras, Dynamics of facial expression: recognition of facial actions and their temporal segments from face profile image sequences, IEEE Transactions on Systems, Man and Cybernetics, Part B (Cybernetics), vol.36, issue.2, pp.433-449, 2006.
DOI : 10.1109/TSMCB.2005.859075

M. Pantic and L. Rothkrantz, Case-based reasoning for user-profiled recognition of emotions from face images, 2004 IEEE International Conference on Multimedia and Expo (ICME) (IEEE Cat. No.04TH8763), pp.391-394, 2004.
DOI : 10.1109/ICME.2004.1394211

M. Pantic, M. Valstar, R. Rademaker, and L. Maat, Web-Based Database for Facial Expression Analysis, 2005 IEEE International Conference on Multimedia and Expo, p.5, 2005.
DOI : 10.1109/ICME.2005.1521424

P. Constantine, M. Papageorgiou, T. Oren, and . Poggio, A general framework for object detection, Computer vision, 1998. sixth international conference on, pp.555-562, 1998.

A. Pentland and T. Choudhury, Face recognition for smart environments, Computer, vol.33, issue.2, pp.50-55, 2000.
DOI : 10.1109/2.820039

S. Petridis and M. Pantic, Audiovisual discrimination between laughter and speech, 2008 IEEE International Conference on Acoustics, Speech and Signal Processing, pp.5117-5120, 2008.
DOI : 10.1109/ICASSP.2008.4518810

S. Petridis and M. Pantic, Audiovisual discrimination between laughter and speech, 2008 IEEE International Conference on Acoustics, Speech and Signal Processing, pp.216-234, 2011.
DOI : 10.1109/ICASSP.2008.4518810

W. Rosalind, R. Picard, and . Picard, Affective computing, 1997.

R. Plutchik, What is an Emotion?, The Journal of Psychology, vol.50, issue.2, 1980.
DOI : 10.1080/00223980.1965.10543417

D. Robert and R. John, Store atmosphere: an environmental psychology approach, Journal of retailing, vol.58, pp.34-57, 1982.

A. James, A. Russell, and . Mehrabian, Evidence for a three-factor theory of emotions, Journal of research in Personality, vol.11, issue.3, pp.273-294, 1977.

J. Ruusuvuori, Looking means listening: coordinating displays of engagement in doctor???patient interaction, Social Science & Medicine, vol.52, issue.7, pp.1093-1108, 2001.
DOI : 10.1016/S0277-9536(00)00227-6

H. Salamin, A. Polychroniou, and A. Vinciarelli, Automatic Detection of Laughter and Fillers in Spontaneous Mobile Phone Conversations, 2013 IEEE International Conference on Systems, Man, and Cybernetics, pp.4282-4287, 2013.
DOI : 10.1109/SMC.2013.730

R. Klaus and . Scherer, Affect bursts. Emotions: Essays on emotion theory, pp.161-196, 1994.

M. Schröder, Experimental study of affect bursts, Speech Communication, vol.40, issue.1-2, pp.99-116, 2003.
DOI : 10.1016/S0167-6393(02)00078-X

B. Schuller, F. Eyben, and G. Rigoll, Static and Dynamic Modelling for the Recognition of Non-verbal Vocalisations in Conversational Speech, Perception in multimodal dialogue systems, pp.99-110, 2008.
DOI : 10.1007/978-3-540-69369-7_12

B. Schuller, S. Steidl, and A. Batliner, The interspeech 2009 emotion challenge, pp.312-315, 2009.

J. Schwartz, P. Escudier, and P. Teissier, Multimodal speech: Two or three senses are better than one. Spoken Language Processing, pp.377-415

N. Sebe, I. Cohen, T. Gevers, S. Thomas, and . Huang, Emotion Recognition Based on Joint Visual and Audio Cues, 18th International Conference on Pattern Recognition (ICPR'06), pp.1136-1139, 2006.
DOI : 10.1109/ICPR.2006.489

N. Sebe, S. Michael, Y. Lew, I. Sun, T. Cohen et al., Authentic facial expression analysis, Image and Vision Computing, vol.25, issue.12, pp.1856-1863, 2007.
DOI : 10.1016/j.imavis.2005.12.021

M. Sehili, F. Yang, L. Leynaert, and . Devillers, A corpus of social interaction between nao and elderly people, 5th International Workshop on Emotion, Social Signals, Sentiment & Linked Open Data (ES3LOD2014). LREC, 2014.

M. El-amine, S. , F. Yang, and L. Devillers, Attention detection in elderly people-robot spoken interaction, Proceedings of the 2014

. Workshop, Multi-Party, Real-World Human-Robot Interaction, pp.7-12, 2014.

T. Senechal, V. Rapp, H. Salam, R. Seguier, K. Bailly et al., Combining aam coefficients with lgbp histograms in the multikernel svm framework to detect facial action units, Automatic Face & Gesture Recognition and Workshops 2011 IEEE International Conference on, pp.860-865, 2011.
URL : https://hal.archives-ouvertes.fr/hal-00657734

T. Sénéchal, V. Rapp, H. Salam, R. Seguier, K. Bailly et al., Facial Action Recognition Combining Heterogeneous Features via Multikernel Learning, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), vol.42, issue.4, pp.993-1005, 2012.
DOI : 10.1109/TSMCB.2012.2193567

Y. Shinohara and N. Otsu, Facial expression recognition using fisher weight maps, Sixth IEEE International Conference on Automatic Face and Gesture Recognition, 2004. Proceedings., pp.499-504, 2004.
DOI : 10.1109/AFGR.2004.1301582

T. Sim, S. Baker, and M. Bsat, The CMU Pose, Illumination, and Expression (PIE) database, Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition, pp.46-51, 2002.
DOI : 10.1109/AFGR.2002.1004130

G. Skantze and J. Gustafson, Attention and interaction control in a human-human-computer dialogue setting, Proceedings of the SIGDIAL 2009 Conference on The 10th Annual Meeting of the Special Interest Group on Discourse and Dialogue, SIGDIAL '09, 2009.
DOI : 10.3115/1708376.1708421

M. Song, J. Bu, C. Chen, and N. Li, Audio-visual based emotion recognition-a new approach, Computer Vision and Pattern Recognition Proceedings of the 2004 IEEE Computer Society Conference on, p.1020, 2004.

Y. Song, Y. Zhang, Z. Wang, and P. Xie, The head-trace mouse for elderly: A human-computer interaction system based on detection of poses of head and mouth, International Journal of Information Technology, vol.19, issue.2, p.2013

R. Stiefelhagen, Tracking focus of attention in meetings, Proceedings. Fourth IEEE International Conference on Multimodal Interfaces, p.273, 2002.
DOI : 10.1109/ICMI.2002.1167006

M. Tahon, A. Delaborde, C. Barras, and L. Devillers, A corpus for identification of speakers and their emotions, LREC, 2010.

M. Tahon and L. Devillers, Acoustic measures characterizing anger across corpora collected in artificial or natural context, Proceedings of the Fifth International Conference on Speech Prosody, 2010.

M. Tahon, A. Mohamed, L. Sehili, and . Devillers, Cross-Corpus Experiments on Laughter and Emotion Detection in HRI with Elderly People, Social Robotics, pp.633-642, 2015.
DOI : 10.1007/978-3-319-25554-5_63

W. James and . Jr, Effects of eye position on person perception. Perceptual and motor skills, pp.883-893, 1970.

H. Tao, S. Thomas, and . Huang, Explanation-based facial motion tracking using a piecewise bezier volume deformation model, Computer Vision and Pattern Recognition, 1999.

Y. Tian, T. Kanade, F. Jeffrey, and . Cohn, Recognizing action units for facial expression analysis. Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol.23, issue.2, pp.97-115, 2001.

E. Michael and . Tipping, Sparse bayesian learning and the relevance vector machine. The journal of machine learning research, pp.211-244, 2001.

M. Mohan and . Trivedi, Human movement capture and analysis in intelligent environments. Machine Vision and Applications, pp.215-217, 2003.

M. Trivedi, K. S. Huang, and I. Mikic, Dynamic context capture and distributed video arrays for intelligent spaces. Systems, Man and Cybernetics, Part A: Systems and Humans, IEEE Transactions on, vol.35, issue.1, pp.145-163, 2005.

K. Truong and D. Van-leeuwen, Evaluating automatic laughter segmentation in meetings using acoustic and acoustics-phonetic features, Proc Proc Workshop on the Phonetics of Laughter at the 16th International Congress of Phonetic Sciences (ICPhS), pp.49-53, 2007.

P. Khiet, D. A. Truong, and . Van-leeuwen, Automatic discrimination between laughter and speech, Speech Communication, vol.49, issue.2, pp.144-158, 2007.

J. Tu, T. Huang, and H. Tao, Accurate head pose tracking in low resolution video, Automatic Face and Gesture Recognition 7th International Conference on, pp.573-578, 2006.

M. Valstar and M. Pantic, Induced disgust, happiness and surprise: an addition to the mmi facial expression database, Proc. 3rd Intern. Workshop on EMOTION (satellite of LREC): Corpora for Research on Emotion and Affect, p.65, 2010.

M. Valstar, M. Pantic, and I. Patras, Motion history for facial action detection in video, 2004 IEEE International Conference on Systems, Man and Cybernetics (IEEE Cat. No.04CH37583), pp.635-640, 2004.
DOI : 10.1109/ICSMC.2004.1398371

F. Michel, M. Valstar, Z. Pantic, . Ambadar, F. Jeffrey et al., Spontaneous vs. posed facial behavior: automatic analysis of brow actions, Proceedings of the 8th international conference on Multimodal interfaces, pp.162-170, 2006.

H. Stephanie, . Van-goozen, E. Nanne, . Van-de-poll, A. Joseph et al., Emotions: Essays on emotion theory, 2014.

R. Vertegaal, R. Slagter, G. Van-der, A. Veer, and . Nijholt, Eye gaze patterns in conversations, Proceedings of the SIGCHI conference on Human factors in computing systems , CHI '01, pp.301-308, 2001.
DOI : 10.1145/365024.365119

P. Viola and M. Jones, Rapid object detection using a boosted cascade of simple features, Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. CVPR 2001, p.511, 2001.
DOI : 10.1109/CVPR.2001.990517

M. Voit, K. Nickel, and R. Stiefelhagen, A Bayesian Approach for Multi-view Head Pose Estimation, 2006 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems, pp.31-34, 2006.
DOI : 10.1109/MFI.2006.265627

A. Waibel, T. Schultz, M. Bett, M. Denecke, R. Malkin et al., SMaRT: the Smart Meeting Room Task at ISL, 2003 IEEE International Conference on Acoustics, Speech, and Signal Processing, 2003. Proceedings. (ICASSP '03)., p.752, 2003.
DOI : 10.1109/ICASSP.2003.1202752

Y. Wang and L. Guan, Recognizing human emotion from audiovisual information, Acoustics, Speech, and Signal Processing, 2005. Proceedings.(ICASSP'05). IEEE International Conference on, p.1125, 2005.

J. Weizenbaum, ELIZA---a computer program for the study of natural language communication between man and machine, Communications of the ACM, vol.9, issue.1, pp.36-45, 1966.
DOI : 10.1145/365153.365168

T. Wu, S. Marian, J. R. Bartlett, and . Movellan, Facial expression recognition using Gabor motion energy filters, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Workshops, pp.42-47, 2010.
DOI : 10.1109/CVPRW.2010.5543267

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.226.4332

T. Wu, J. Nicholas, P. Butko, J. Ruvolo, . Whitehill et al., Multilayer architectures for facial action unit recognition. Systems, Man, and Cybernetics, Part B: Cybernetics IEEE Transactions on, vol.42, issue.4, pp.1027-1038, 2012.

J. Xiao, T. Moriyama, T. Kanade, F. Jeffrey, and . Cohn, Robust full-motion recovery of head by dynamic templates and re-registration techniques, International Journal of Imaging Systems and Technology, vol.23, issue.1, pp.85-94, 2003.
DOI : 10.1002/ima.10048

F. Yang, A. Mohamed, C. Sehili, L. Barras, and . Devillers, Smile and Laughter Detection for Elderly People-Robot Interaction, Social Robotics, pp.694-703, 2015.
DOI : 10.1007/978-3-319-25554-5_69

L. Yin, X. Wei, Y. S. Wang, J. Matthew, and . Rosato, A 3d facial expression database for facial behavior research, Automatic face and gesture recognition 7th international conference on, pp.211-216, 2006.

Z. Zeng, Y. Fu, I. Glenn, Z. Roisman, Y. Wen et al., Spontaneous Emotional Facial Expression Detection, Journal of Multimedia, vol.1, issue.5, pp.1-8, 2006.
DOI : 10.4304/jmm.1.5.1-8

Z. Wen, Audio-visual emotion recognition in adult attachment interview, Proceedings of the 8th international conference on Multimodal interfaces, pp.139-145, 2006.

Z. Zeng, Y. Hu, M. Liu, Y. Fu, S. Thomas et al., Training combination strategy of multi-stream fused hidden Markov model for audio-visual affect recognition, Proceedings of the 14th annual ACM international conference on Multimedia , MULTIMEDIA '06, pp.65-68, 2006.
DOI : 10.1145/1180639.1180661

Z. Zeng, J. Tu, M. Liu, S. Thomas, B. Huang et al., Audio-visual affect recognition. Multimedia, IEEE Transactions on, vol.9, issue.2, pp.424-428, 2007.

Z. Zeng, J. Tu, B. Pianfetti, M. Liu, T. Zhang et al., Audio-Visual Affect Recognition through Multi-Stream Fused HMM for HCI, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05), pp.967-972, 2005.
DOI : 10.1109/CVPR.2005.77

Y. Zhang and Q. Ji, Active and dynamic information fusion for facial expression understanding from image sequences. Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol.27, issue.5, pp.699-714, 2005.

Z. Zhang, Microsoft Kinect Sensor and Its Effect, IEEE Multimedia, vol.19, issue.2, pp.4-10, 2012.
DOI : 10.1109/MMUL.2012.24

G. Zhao and M. Pietikainen, Dynamic texture recognition using local binary patterns with an application to facial expressions. Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol.29, issue.6, pp.915-928, 2007.