Skip to main content

Experimental Analysis of the Changes in Speech while Normal Speaking, Walking, Running, and Eating

  • Conference paper
  • First Online:
  • 741 Accesses

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 1090))

Abstract

Speech communication is a mundane activity. People talk while ambling, running, biking, eating, and so on. This paper analyzes the variability in speech parameters while a person speaking when walking, eating, or running with the normal speaking. In this paper, we get information about the analysis of speech signals while normally speaking and speaking along with walking, eating, or running. The analysis of speech signal, according to minimum pitch, maximum pitch, mean energy intensity and mean F1 (first formant), has been carried out and also experimented in Praat, and then the problem is implemented in MATLAB. This paper meets the variation in speech parameters in different scenarios. The clustering is performed, and the paper presents an experimental comparison of two different features of speech.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Lamberg, E.M., and L.M. Muratori. 2012. Cell phones change the way we walk. Gait and Posture 35: 688–690.

    Article  Google Scholar 

  2. Al-Yahyaa, E., H. Dawesa, L. Smith, A. Dennisa, K. Howells, and J. Cockburn. 2011. Cognitive Motor Interference While Walking: A Systematic Review and Meta-Analysis. Neuroscience and Biobehavioral Reviews 35: 715–728.

    Article  Google Scholar 

  3. Reichel, U.D. 2012. PermA and Balloon: Tools for String Alignment and Text Processing. In Proceedings of Interspeech, Portland, Oregon, paper no. 346.

    Google Scholar 

  4. Schuller, B., S. Steidl, A. Batliner, F. Burkhardt, L. Devillers, C. MuLler, and S. Narayanan. 2013. Paralinguistics in Speech and Language State-of-the-Art and the Challenge. Computer Speech Language 27 (1): 4–39.

    Google Scholar 

  5. Schuller, B., S. Steidl, A. Batliner, S. Hantke, F. Honig, J.R. Orozco Arroyave, E. Noth, Y. Zhang, and F. Weninger. 2015. The Interspeech 2015 Computational Paralinguistics Challenge: Nativeness, Parkinson’s Eating Condition. In Proceedings INTERSPEECH.

    Google Scholar 

  6. Prasad, A. and P.K. Ghosh. 2015. Automatic Classification of Eating Conditions from Speech Using Acoustic Feature Selection and a Set of Hierarchical Support Vector Machine Classifiers. In Sixteenth Annual Conference of the International Speech Communication Association.

    Google Scholar 

  7. Schuller, B., S. Steidl, A. Batliner, S. Hantke, F. Honig, J.R. Orozco-Arroyave, E. Noth, Y. Zhang, and F. Weninger. 2015. The Interspeech 2015 Computational Paralinguistics Challenge: Nativeness, Parkinson’s and Eating Condition. In Proceedings INTERSPEECH.

    Google Scholar 

  8. Hantke, S., F. Weninger, R. Kurle, A. Batliner, and B. Schuller. 2015. I Hear You Eat and Speak: Automatic Recognition of Eating Condition and Food Type, ms. to appear.

    Google Scholar 

  9. Fuchs, S., U.D. Reichel, and A. Rochet-Capellan. 2015. Changes in Speech and Breathing Rate While Speaking and Biking.

    Google Scholar 

  10. Lee, Sangkil, Jieun Kim, and Insung Lee. Speech/Audio Signal Classification Using Spectral Flux Pattern Recognition. 978-0-7695-4856-2/12 26.00 © 2012 IEEE https://doi.org/10.1109/sips.2012.36. Bayram, S., I. Avcibas, B. Sankur, and N. Memon. 2006. Image Manipulation Detection. Journal of Electronic Imaging 15 (4): 041102-1–041102-17.

  11. Lu, Li, Qingwei Zhao, Yonghong Yan and Kun Liu. 2010. A Study on Sports Video Classification Based on Audio Analysis and Speech Recognition. 978-1-4244-5858-5/10/26.00 ©2010 IEEE.

    Google Scholar 

  12. https://www.stanford.edu/.

  13. http://practicalcryptography.com/miscellaneous/machine-learning/guide-mel-frequency-cepstral-coefficients-mfccs/#implementation-steps.

  14. Wagstaff, Kiri, Seth Rogers. 2001. Constrained K-means Clustering with Background Knowledge. In Proceedings of the Eighteenth International Conference on Machine Learning, 577–584.

    Google Scholar 

  15. Xiong, Z., R. Radhakrishnan, A. Divakaran, and T.S. Huang. 2005. Highlights Extraction from Sports Video Based on an Audio-Visual Marker Detection Framework. In IEEE International Conference on Multimedia and Expo, ICME 2005, 4.

    Google Scholar 

Download references

Acknowledgement

The authors would like to express their gratitude for all participants for their precious time that they have spent for data recording.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Sakil Ansari .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ansari, S., Mittal, S.K., Kamakshi Prasad, V. (2020). Experimental Analysis of the Changes in Speech while Normal Speaking, Walking, Running, and Eating. In: Raju, K., Govardhan, A., Rani, B., Sridevi, R., Murty, M. (eds) Proceedings of the Third International Conference on Computational Intelligence and Informatics . Advances in Intelligent Systems and Computing, vol 1090. Springer, Singapore. https://doi.org/10.1007/978-981-15-1480-7_7

Download citation

Publish with us

Policies and ethics