Evaluation of a Self-report System for Assessing Mood Using Facial Expressions

  • Hristo ValevEmail author
  • Tim Leufkens
  • Corina Sas
  • Joyce Westerink
  • Ron Dotsch
Conference paper
Part of the Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering book series (LNICST, volume 288)


Effective and frequent sampling of mood through self-reports could enable a better understanding of the interplay between mood and events influencing it. To accomplish this, we built a mobile application featuring a sadness-happiness visual analogue scale and a facial expression-based scale. The goal is to evaluate, whether a facial expression based scale could adequately capture mood. The method and mobile application were evaluated with 11 participants. They rated the mood of characters presented in a series of vignettes, using both scales. Participants also completed a user experience survey rating the two assessment methods and the mobile interface. Findings reveal a Pearson’s correlation coefficient of 0.97 between the two assessment scales and a stronger preference for the face scale. We conclude with a discussion of the implications of our findings for mood self-assessment and an outline future research.


Mood assessment Self-report system User interface 



This work has been supported by AffecTech: Personal Technologies for Affective Health, Innovative Training Network funded by the H2020 People Programme under Marie Skłodowska-Curie grant agreement No. 722022.


  1. 1.
    Barrows, P.D., Thomas, S.A.: Assessment of mood in aphasia following stroke: validation of the Dynamic Visual Analogue Mood Scales (D-VAMS). Clin. Rehabil. 32(1), 94–102 (2018). Scholar
  2. 2.
    Broekens, J., Brinkman, W.P.: AffectButton: a method for reliable and valid affective self-report. Int. J. Hum. Comput. Stud. 71(6), 641–667 (2013). Scholar
  3. 3.
    Price, D.D., McGrath, P.A., Rafii, A., Buckingham, B.: The validation of visual analogue scales as ratio scale measures for chronic and experimental pain. Pain 17(1), 45–56 (1983)CrossRefGoogle Scholar
  4. 4.
    Duque, A., Vázquez, C.: Mental Health; researchers at Complutense University have reported new data on depression (Double attention bias for positive and negative emotional faces in clinical depression: evidence from an eye-tracking study). Mental Health Weekly Digest 46, 124 (2015). Scholar
  5. 5.
    Ekman, P.: Universal-Facial-Expressions-of-Emotion (1970)Google Scholar
  6. 6.
    Ekman, P.: Facial expression and emotion. Am. Psychol. 48(4), 384–392 (1993). Scholar
  7. 7.
    Huang, S.T.Y., Kwan, C.M.Y., Sano, A.: The moment. In: Proceedings of the 2014 ACM International Joint Conference on Pervasive and Ubiquitous Computing Adjunct Publication - UbiComp 2014 Adjunct, pp. 235–238 (2014).
  8. 8.
    Joormann, J., Gotlib, I.H.: Is this happiness i see? Biases in the identification of emotional facial expressions in depression and social phobia. J. Abnorm. Psychol. 115(4), 705–714 (2006). Scholar
  9. 9.
    Kamashita, Y., Sonoda, T., Kamada, Y., Nishi, Y., Nagaoka, E.: Reliability, validity, and preference of an original faces scale for assessing the mood of patients with dentures. Prosthodontic Res. Pract. 6, 93–98 (2007)CrossRefGoogle Scholar
  10. 10.
    Lagotte, A.: Eliciting discrete positive emotions with vignettes and films: a validation study (2014).
  11. 11.
    Lorish, C.D., Maisiak, R.: The face scale: a brief, nonverbal method for assessing patient mood. Arthritis Rheum. 29(7), 906–909 (1986). Scholar
  12. 12.
    McKinley, S., Coote, K., Stein-Parbury, J.: Development and testing of a faces scale for the assessment of anxiety in critical ill patients. J. Adv. Nurs. 41(1), 73–79 (2003)CrossRefGoogle Scholar
  13. 13.
    Pollak, J., Adams, P., Gay, G.: PAM: a photographic affect meter for frequent, in situ measurement of affect. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 725–734 (2011).
  14. 14.
    Rodriguez, I., Herskovic, V., Fuentes, C., Campos, M.: B-ePain: a wearable interface to self-report pain and emotions. In: UbiComp Adjunct, pp. 1120–1125 (2016).
  15. 15.
    Sanches, P., Hook, K., Sas, C., Stahl, A.: Ambiguity as a resource to inform proto-practices: the case of skin conductance. TOCHI 26(1) (2019). Scholar
  16. 16.
    Sanches, P., et al.: HCI and affective health taking stock of a decade of studies and charting future research directions, pp. 123–4567 (2019)Google Scholar
  17. 17.
    Sas, C., Rees, M.: AffectCam: arousal – augmented SenseCam for RicherRecall of episodic memories. In: CHI 2013, pp. 1041–1046 (2013).
  18. 18.
    Stone, A.A., Shiffman, S.: Ecological momentary assessment (EMA) in behavioral medicine. Ann. Behav. Med. 16(3), 199–202 (2018). Scholar
  19. 19.
    Umair, M., Latif, M.H., Sas, C.: Dynamic displays at wrist for real time visualization of affective data, pp. 201–205 (2018).
  20. 20.
    Watson, D., Clark, L.A.: PANAS-X manual, pp. 1–27 (1999). Scholar
  21. 21.
    Watson, D., Clark, L.A.: JPSP Watson Clark Tellegen 1988 54(6), 1–8 (2004). papers3://publication/uuid/CA1FC7DA-0BA7-4563-8673-5E3942A0983FCrossRefGoogle Scholar
  22. 22.
    Westie, F.R.: A technique for the measurement of race attitudes. Am. Sociol. Rev. 18(1), 73 (2006). Scholar

Copyright information

© ICST Institute for Computer Sciences, Social Informatics and Telecommunications Engineering 2019

Authors and Affiliations

  • Hristo Valev
    • 1
    • 2
    Email author
  • Tim Leufkens
    • 1
    • 3
  • Corina Sas
    • 2
  • Joyce Westerink
    • 1
    • 3
  • Ron Dotsch
    • 1
  1. 1.Philips ResearchEindhovenThe Netherlands
  2. 2.Lancaster UniversityBailrigg, LancasterUK
  3. 3.Eindhoven University of Technology Tu/eEindhovenThe Netherlands

Personalised recommendations