Estimating Age-Dependent Degradation Using Nonverbal Feature Analysis of Daily Conversation

  • Natsumi Kana
  • Yumi WakitaEmail author
  • Yoshihisa Nakatoh
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 11582)


In this paper, we study a system that estimates the degree of decline in the driving ability of elderly people using non-verbal information from daily conversations. It is necessary for us to find the factors that would affect the calculation of the degree of decline that has reached a problematic level for functioning daily life. We focus on the cases where elderly people cannot understand their partner’s speech as their hearing and concentration abilities decrease with age. We analyze the relationship between the degree of understanding of the partner’s speech and the non-verbal characteristic of the response scene. Based on the results of the acoustic analysis of each utterance, the fundamental frequency (F0) and acoustic power levels of when a person can understand their partner’s speech tend to be higher than those when they cannot. The analysis of the synchronization of the head motions shows that brightness value of difference image when a person can understand their partner’s speech is also higher than when they cannot. These results indicate that these non-verbal factors are effective in estimating the decline in the hearing and concentration abilities of the elderly.


Degree of decline Fundamental frequency Synchronism of motion Understanding level 


  1. 1.
    Wakita, Y., Matsumoto, S.: Communication smoothness estimation using F0 information. In: 2016 Proceedings of the 4th IIAE International Conference on Intelligent Systems and Image Processing, September 2016Google Scholar
  2. 2.
    Nueller, P.B., Sweeney, R.J., Barbeau, L.J.: Acoustic and morphologic study of the senescent voice. Ear Noise Throat J. 63, 71–75 (1985)Google Scholar
  3. 3.
    Sato, K., Sakaguchi, S., Hirano, M.: Histologic investigation of bowing of the aged vocal folds. Throat J. 8, 11–14 (1996)Google Scholar
  4. 4.
    Nishio, M., Tanaka, Y., Niimi, S.: Analysis of age-related changes in the acoustic characteristics of the voice. Jpn. Soc. Logop. Phoniatr. 50, 6–13 (2009)CrossRefGoogle Scholar
  5. 5.
    Tanaka, Y., Igaue, H., Mizumachi, M., Nakatoh, Y.: Study of improvement of intelligibility for the elderly speech based on formant frequency shift. Int. J. Comput. Consum. Control (IJ3C) 3(3), 57–65 (2014)Google Scholar
  6. 6.
    Talkin, D.: A robust algorithm for pitch tracking (RAPT). In: Kleijn, W.B., Pailwal, K.K. (eds.) Speech Coding & Synthesis, pp. 495–518. Elsevier, Amsterdam (1995)Google Scholar
  7. 7.
    Natsumi, K., Wakita, Y., Nakatoh, Y.: Changes in fundamental frequency and gesture of response corresponding to the understanding level of partner’s talk. In: 2018 IEEE International Conference on Artificial Intelligence in Engineering and Technology, November 2018Google Scholar

Copyright information

© Springer Nature Switzerland AG 2019

Authors and Affiliations

  1. 1.Osaka Institute of TechnologyOsakaJapan
  2. 2.Kyushu Institute of TechnologyFukuokaJapan

Personalised recommendations