skip to main content
10.1145/1290128.1290134acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
Article

Affective multimodal mirror: sensing and eliciting laughter

Published: 28 September 2007 Publication History

Abstract

In this paper, we present a multimodal affective mirror that senses and elicits laughter. Currently, the mirror contains a vocal and a facial affect-sensing module, a component that fuses the output of these two modules to achieve a user-state assessment, a user state transition model, and a component to present audiovisual affective feedback that should keep or bring the user in the intended state. Interaction with this intelligent interface involves a full cyclic process of sensing, interpreting, reacting, sensing (of the reaction effects), interpreting & The intention of the mirror is to evoke positive emotions, to make people laugh and to increase the laughter. The first user experiences tests showed that users show cooperative behavior, resulting in mutual user-mirror action-reaction cycles. Most users enjoyed the interaction with the mirror and immersed in an excellent user experience.

References

[1]
Aarts, E. Ambient intelligence: a multimedia perspective. IEEE Multimedia 11 (1), 2004, 12--19.
[2]
Andrés del Valle, A.C. and Opalach, A. The Persuasive Mirror. In Proceedings of Persuasive, 2006.
[3]
Bachorowski, J.A., amd Owren, M.J. (2001). Not all laugs are alike: Voiced but not unvoiced laughter readily elicits positive affect. Psychological Science, Vol. 12(3), 252--257.
[4]
Bradley, M.M. and Lang, P.J. Affective norms for English words (ANEW). Gainesville, FL. The NIMH Center for the Study of Emotion and Attention, University of Florida.
[5]
Brøndsted, T., Nielsen, T.D., and Ortega, S. Affective multimodal interaction with a 3d agent. In Proceedings of the 8th international workshop on the cognitive science of natural language processing, 1999, 102--109.
[6]
Clavel, C., Ehrette, T., and Richard, G. Events detection for an audio-based surveillance system. In Proceedings of IEEE International Conference on Multimedia and Expo (ICME 2005), 2005, 1306--1309.
[7]
Collobert, R., Bengio, S., and Mariéthoz, J. Torch: a modular machine learning software library. Technical Report IDIAP-RR 02-46, IDIAP, 2002.
[8]
Darrell, T., Gordon, G., Woodfill, J., and Harville, M. A virtual mirror interface using real-time robust face tracking. In Proceedings of the 3th International Conference on Face and Gesture Recognition, 1998.
[9]
Den Uyl, M.J. and Van Kuilenburg, H. The FaceReader: online facial expression recognition. In Proceedings of Measuring Behavior, 2005.
[10]
Duric, Z., Gray, W.D., Heishman, R., Li, F., Rosenfeld, A., Schoelles, M.J., Schunn, C., and Wechsler, H. Integrating perceptual and cognitive modeling for adaptive and intelligent human-computer interaction. In Proceedings of the IEEE, Vol.90 (7), 2002, 1272--1289.
[11]
Ex'ovision EyeCatcher: http://www.exovision.nl/download/EyeCatcher-leaflet.pdf
[12]
Grootjen, M., Neerincx, M.A., Weert, J.C.M., and Truong, K.P. Measuring cognitive task load on a naval ship: implications of a real world environment. In Proceedings of International Conference on Human-Computer Interaction (HCII'07), 2007.
[13]
Hauptmann, A.G. Speech and gestures for graphic image manipulation. In Proceedings of ACM conference on Human Factors in Computing Systems (CHI'89), 1989, 241--245.
[14]
Janin, A., Baron, D., Edwards, J., Ellis, D., Gelbart, D., Morgan, N., Peskin, B., Pfau, T., Shriberg, E., Stolcke, A., and Wooters, C. The ICSI meeting corpus. In Proceedings of IEEE International conference on acoustics, speech and signal processing (ICASSP'03), 2003, 364--367.
[15]
Kapoor, A. and Picard, R.W. Multimodal affect recognition in learning environments. In Proceedings of the 13th annual ACM International Conference on Multimedia, 2005, 677--682.
[16]
Kim, J., André, E., Rehm, M., Vogt, T., and Wagner, J.allIntegrating information from speech and physiological signals to achieve emotional sensitivity. In Proceedings of Interspeech, 2005.
[17]
Leatherby, J.H., and Pausch, R. Voice input as a replacement for keyboard accelerators in a mouse-based graphical editor: an empirical study. In Journal of the American Voice Input/Output Society 11 (2), 1992.
[18]
Lisetti, C.L. and Nasoz, F. MAUI: a multimodal affective user interface. In Proceedings of ACM Multimedia Interaction Conference 2002, 161--170.
[19]
Litman, D.J. and Forbes-Riley, K. Recognizing student emotions and attitudes on the basis of utterances in spoken tutoring dialogues with both human and computer tutors. Speech Communication 48 (5), 2006, 559--590.
[20]
Lundqvist, D., Flykt, A., and Öhman, A. The Karolinska Directed Emotional Faces-KDEF, Department of Clinical Neuroscience, Psychology section, Karolinska Institute.
[21]
Maat, L. and Pantic, M. Gaze-X: adaptive affective multimodal interface for single-user office scenarios. In Artificial Intelligence for Human Computing, Vol.4451, 2007, 251--271.
[22]
Merkx, P.A.B., Truong, K.P. and Neerincx, M.A. (2007). Inducing and measuring emotion through a multiplayer first-person shooter computer game. In: H.J. van den Herik, J.W.H.M. Uiterwijk, M.H.M. Winands, and M.P.D. Schadd (Eds.). Proceedings of the Computer Games Workshop 2007, Amsterdam, The Netherlands.
[23]
Neerincx, M.A. and Lindenberg, J. Situated cognitive engineering for complex task environments. In: Schraagen, J.M.(Ed.), Natural Decision Making & Macro cognition. Ashley.
[24]
Olivier, M., Benoit, M., Irène, K., Arman, S., and Jordi, A. Multimodal caricatural mirror. In Proceedings of ENTERFACE'05, 1st Summer Workshop on Multimodal Interfaces, 2005, 13--20.
[25]
Oviatt, S. User-centered modeling and evaluation of multimodal interfaces. In Proceedings of the IEEE, Vol.91 (9), 2003, 1457--1468.
[26]
Pantic, M. and Rothkrantz, L.J.M. Towards an affect-sensitive multimodal human-computer interaction. In Proceedings of the IEEE, Vol.91, 2003.
[27]
Pantic, M., Sebe, N., Cohn, J.F. and Huang, T. Affective multimodal human-computer interaction, In Proceedings of the 13th annual ACM international conference on multimedia, 2006.
[28]
Pellom, B., SONIC: The university of Colorado Continuous Speech Recognizer, University of Colorado, tech report#TR-CSLR-2001-01, Boulder, Colorado, March, 2001.
[29]
Picard, R.W. Affective computing. MIT Press, Cambridge, MA, 1997.
[30]
Prendinger, H., Mayer, S., Mori, J., and Ishizuka, M. Persona effect revisited: using bio-signal to measure and reflect the impact of character-based interfaces. In Proceedings of the Intelligent Virtual Agents 4th International Workshop (IVA'03), 2003.
[31]
Sebe, N., Cohen, I., Gevers, T., and Huang, T. Emotion recognition based on joint visual and audio cues. In Proceedings of the International Conference on Pattern Recognition (ICPR 2006), 2006, 1136--1139.
[32]
SPRACHcore. http://www.icsi.berkeley.edu/~dpwe/projects/sprach/
[33]
Sung, K.K. and Poggio, T. Example-based learning for view-based human face detection. IEEE Transactions on Pattern Analysis and Machine Intelligence 20 (1), 1998, 39--51.
[34]
Truong, K.P. and Van Leeuwen, D.A. Automatic discrimination between laughter and speech, Speech Communication 49 (2), 2007, 144--158.
[35]
Truong, K.P, van Leeuwen, D.A., and Neerincx, M.A. (in press). Unobtrusive Multimodal Emotion Detection in Adaptive Interfaces: Speech and Facial Expressions. In: D.D. Schmorrow & L.M. Reeves (Eds), Foundations of Augmented Cognition, 3rd Edition, LNAI 4565 proceedings (10 pages), ISBN 978-3-540-73215-0.
[36]
V2_ institute for the unstable media, MUSH device v3: http://multimedian.v2.nl/mush/mushv3_documentation.pdf
[37]
Van Kuilenburg, H., Wiering, Marco, Den Uyl, M. A model based method for automatic facial expression recognition. In Proceedings of the European Conference on Machine Learning (ECML' 05), 2005, 194--205.
[38]
Vidrascu, L., and Devillers, L. Detection of real-life emotions in dialogs recorded in a call center. In Proceedings of Interspeech, 2005, 1841--1844.
[39]
Vogt, T. and André, E. Comparing feature sets for acted and spontaneous speech in view of automatic emotion recognition. In Proceedings of IEEE International Conference on Multimedia & Expo (ICME'05), 2005.
[40]
Waag Society, KeyWorx: http://www.keyworx.org, Realizer: http://kwlive.dev.waag.org/realizer/doc/doxygen/html/
[41]
Wilting, J., Krahmer, E., and Swerts, M. Real vs. acted emotional speech. In Proceedings of Interspeech, 2006.
[42]
Yacoub, S., Simske, S.,Lin, X., and Burns, J. Recognition of emotions in interactive voice response systems. In Proceedings of Eurospeech, 2003, 729--732.
[43]
Yang, G., Lin, Y., and Bhattacharya, P. A driver fatigue recognition model using fusion of multiple features. In IEEE SMC, Vol.2, 2005, 1777--1784.
[44]
Zeng, Z., Hu, Y., Fu, Y., Huang, T.S., Roisman, Z.W., and Wen, Z. Audio-visual emotion recognition in adult attachment interview. In Poceedings of the 8th International Conference on Multimodal Interfaces (ICMI '06), 2006, 139--145.
[45]
Zhai, J. and Barreto, A. Stress recognition using non-invasive technology. In Proceedings of FLAIRS Conference, 2006, 395--401.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
HCM '07: Proceedings of the international workshop on Human-centered multimedia
September 2007
112 pages
ISBN:9781595937810
DOI:10.1145/1290128
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 28 September 2007

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. affective mirror
  2. face voice emotion expression
  3. multi modal laughter recognition

Qualifiers

  • Article

Conference

MM07
MM07: The 15th ACM International Conference on Multimedia 2007
September 28, 2007
Bavaria, Augsburg, Germany

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)13
  • Downloads (Last 6 weeks)0
Reflects downloads up to 24 Dec 2024

Other Metrics

Citations

Cited By

View all

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media