Computer Assisted Transcription of Speech Signals
Automatic Speech Recognition has been widely employed in the last years. However, when a perfect transcription of the input is required, it is still necessary to rely on a human operator that supervises and corrects the mistakes that recognition systems usually make. Although the use of automatic systems can speed up the transcription process significantly, the intervention of these human supervisors can slow down this job considerably. Owing to this fact, the application of the Interactive Pattern Recognition approach to this task turns out to be a good opportunity to improve the cooperation between the computer and the human when an error-free transcribed document is needed.
In this chapter, an interactive multimodal approach for efficient transcriptions of speech signal is presented. This approach, rather than full automation, aims at assisting the expert in the proper transcription process. In this sense, an interactive scenario is proposed and it is based on a cooperative process between an automatic recognition system and a human transcriber to generate the final transcription of the speech signal. It will be shown how user’s feedback directly allows one to improve the system accuracy, while multimodality increases system ergonomics and user acceptability.
KeywordsDiscrete Cosine Transform Speech Recognition Speech Signal Automatic Speech Recognition User Feedback
- 3.Chen, S. F., & Goodman, J. (1996). An empirical study of smoothing techniques for language (Technical Report). Google Scholar
- 4.Civera, J., Vilar, J. M., Cubel, E., Lagarda, A. L., Casacuberta, F., Vidal, E., Picó, D., & González, J. (2004). A syntactic pattern recognition approach to computer assisted translation. In A. Fred, T. Caelli, A. Campilho, R. P. Duin & D. de Ridder (Eds.), Lecture notes in computer science. Advances in statistical, structural and syntactical pattern recognition—joint IAPR international workshops on syntactical and structural pattern recognition (SSPR 2004) and statistical pattern recognition (SPR 2004). Berlin: Springer. Google Scholar
- 5.Cubel, E., Civera, J., Vilar, J. M., Lagarda, A. L., Barrachina, S., Vidal, E., Casacuberta, F., Picó, D., González, J., & Rodríguez, L. (2004). Finite-state models for computer assisted translation. In Proceedings of the 16th European conference on artificial intelligence (ECAI04) (pp. 586–590), Valencia, Spain. Google Scholar
- 6.Llorens, D., Casacuberta, F., Segarra, E., Sánchez, J. A., & Aibar, P. (1999). Acoustical and syntactical modeling in ATROS system. In Proceedings of international conference on acoustic, speech and signal processing (ICASSP99) (pp. 641–644), Phoenix, Arizona, USA. Google Scholar
- 7.Ney, H., & Ortmanns, S. (1997). Extensions to the word graph method for large vocabulary continuous speech recognition. In IEEE international conference on acoustics, speech, and signal processing (Vol. 3, pp. 1791–1794), Munich, Germany. Google Scholar
- 11.Sankoff, D., & Kruskal, J. B. (1983). Time warps, string edits, and macromolecules: The theory and practice of sequence comparison. Reading: Addison-Wesley. Google Scholar
- 12.Stolcke, A. (2002). SRILM—an extensible language modeling toolkit. In Proceedings of the international conference on spoken language processing (ICSLP02) (pp. 901–904), Denver, Colorado, USA. Google Scholar
- 14.Young, S. J. (1994). The htk hidden Markov model toolkit: Design and philosophy. Entropic Cambridge Research Laboratory, Ltd, 2, 2–44. Google Scholar