Abstract
This brief paper is an editorial for the special issue on “Speech communication integrated with other modalities”. This special issue contains extended versions of selected topical papers from the 19th International Conference on Speech and Computer SPECOM-2017, organized on 12–16 September 2017 in Hatfield, UK. Five extended articles were selected for this special issue, all of which deal with speech-based human–computer communication jointly with visual, textual and/or other interaction modalities. The first three papers study various aspects of multimodal human–computer interaction, and in the remaining two papers, the authors study the video components of audio-visual speech recognition systems. In this editorial, we present an overview of the accepted articles and the selection process.
Similar content being viewed by others
References
Karpov A, Potapova R, Mporas I (eds) Proceedings of the 19th international conference on speech and computer SPECOM 2017, Hatfield, UK, 2017, vol 10458. Springer LNCS. https://doi.org/10.1007/978-3-319-66429-3
Schuller B, Zhang Y, Weninger F (2018) Three recent trends in paralinguistics on the way to omniscient machine intelligence. J Multimodal User Interfaces. https://doi.org/10.1007/s12193-018-0270-6
Schuller BW (2017) Big data, deep learning—at the edge of X-ray speaker analysis. In: Speech and computer. SPECOM 2017. Lecture notes in computer science, vol 10458. Springer, Cham. https://doi.org/10.1007/978-3-319-66429-3_2
Salim FA, Haider F, Conlan O et al (2018) An approach for exploring a video via multimodal feature extraction and user interactions. J Multimodal User Interfaces. https://doi.org/10.1007/s12193-018-0268-0
Gilmartin E, Cowan B, Vogel C, Campbell N (2018) Explorations in multiparty casual social talk and its relevance for social human machine dialogue. J Multimodal User Interfaces. https://doi.org/10.1007/s12193-018-0274-2
Paleček K (2018) Experimenting with lipreading for large vocabulary continuous speech recognition. J Multimodal User Interfaces. https://doi.org/10.1007/s12193-018-0266-2
Ivanko D, Karpov A, Fedotov D et al (2018) Multimodal speech recognition: increasing accuracy using high speed video data. J Multimodal User Interfaces. https://doi.org/10.1007/s12193-018-0267-1
Acknowledgements
The guest editors are grateful to the Editor-in-Chief Prof. Jean-Claude Martin for the cooperation and support of this Special Issue, as well as to all outstanding reviewers, who provided detailed and insightful reviews of the extended papers submitted for this Special Issue (in the alphabetical order): Gerard Bailly, Marie-Luce Bourguet, Nick Campbell, Nikos Fakotakis, Kristiina Jokinen, Oliver Jokisch, Irina Kipyatkova, Wolfgang Minker, and Milos Zelezny.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Karpov, A., Mporas, I. Speech communication integrated with other modalities. J Multimodal User Interfaces 12, 271–272 (2018). https://doi.org/10.1007/s12193-018-0275-1
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12193-018-0275-1