Computer Assisted Transcription of Speech Signals

  • Alejandro Héctor Toselli
  • Enrique Vidal
  • Francisco Casacuberta


Automatic Speech Recognition has been widely employed in the last years. However, when a perfect transcription of the input is required, it is still necessary to rely on a human operator that supervises and corrects the mistakes that recognition systems usually make. Although the use of automatic systems can speed up the transcription process significantly, the intervention of these human supervisors can slow down this job considerably. Owing to this fact, the application of the Interactive Pattern Recognition approach to this task turns out to be a good opportunity to improve the cooperation between the computer and the human when an error-free transcribed document is needed.

In this chapter, an interactive multimodal approach for efficient transcriptions of speech signal is presented. This approach, rather than full automation, aims at assisting the expert in the proper transcription process. In this sense, an interactive scenario is proposed and it is based on a cooperative process between an automatic recognition system and a human transcriber to generate the final transcription of the speech signal. It will be shown how user’s feedback directly allows one to improve the system accuracy, while multimodality increases system ergonomics and user acceptability.


Discrete Cosine Transform Speech Recognition Speech Signal Automatic Speech Recognition User Feedback 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


  1. 1.
    Amengual, J. C., & Vidal, E. (1998). Efficient error-correcting Viterbi parsing. IEEE Transactions on Pattern Analysis and Machine Intelligence, PAMI-20(10), 1109–1116. CrossRefGoogle Scholar
  2. 2.
    Amengual, J. C., Benedí, J. M., Casacuberta, F., Castaño, A., Castellanos, A., Jiménez, V., Llorens, D., Marzal, A., Pastor, M., Prat, F., Vidal, E., & Vilar, J. M. (2000). The EuTrans-I speech translation system. Machine Translation, 15, 75–103. MATHCrossRefGoogle Scholar
  3. 3.
    Chen, S. F., & Goodman, J. (1996). An empirical study of smoothing techniques for language (Technical Report). Google Scholar
  4. 4.
    Civera, J., Vilar, J. M., Cubel, E., Lagarda, A. L., Casacuberta, F., Vidal, E., Picó, D., & González, J. (2004). A syntactic pattern recognition approach to computer assisted translation. In A. Fred, T. Caelli, A. Campilho, R. P. Duin & D. de Ridder (Eds.), Lecture notes in computer science. Advances in statistical, structural and syntactical pattern recognition—joint IAPR international workshops on syntactical and structural pattern recognition (SSPR 2004) and statistical pattern recognition (SPR 2004). Berlin: Springer. Google Scholar
  5. 5.
    Cubel, E., Civera, J., Vilar, J. M., Lagarda, A. L., Barrachina, S., Vidal, E., Casacuberta, F., Picó, D., González, J., & Rodríguez, L. (2004). Finite-state models for computer assisted translation. In Proceedings of the 16th European conference on artificial intelligence (ECAI04) (pp. 586–590), Valencia, Spain. Google Scholar
  6. 6.
    Llorens, D., Casacuberta, F., Segarra, E., Sánchez, J. A., & Aibar, P. (1999). Acoustical and syntactical modeling in ATROS system. In Proceedings of international conference on acoustic, speech and signal processing (ICASSP99) (pp. 641–644), Phoenix, Arizona, USA. Google Scholar
  7. 7.
    Ney, H., & Ortmanns, S. (1997). Extensions to the word graph method for large vocabulary continuous speech recognition. In IEEE international conference on acoustics, speech, and signal processing (Vol. 3, pp. 1791–1794), Munich, Germany. Google Scholar
  8. 8.
    Nyquist, H. (2002). Certain topics in telegraph transmission theory. Proceedings of the IEEE, 90(2), 280–305. CrossRefGoogle Scholar
  9. 9.
    Ortmanns, S., Ney, H., & Aubert, X. (1997). A word graph algorithm for large vocabulary continuous speech recognition. Computer Speech and Language, 11(1), 43–72. CrossRefGoogle Scholar
  10. 10.
    Paul, D. B., & Baker, J. M. (1992). The design for the wall street journal-based csr corpus. In HLT’91: Proceedings of the workshop on speech and natural language (pp. 357–362), Morristown, NJ, USA. Menlo Park: Association for Computational Linguistics. CrossRefGoogle Scholar
  11. 11.
    Sankoff, D., & Kruskal, J. B. (1983). Time warps, string edits, and macromolecules: The theory and practice of sequence comparison. Reading: Addison-Wesley. Google Scholar
  12. 12.
    Stolcke, A. (2002). SRILM—an extensible language modeling toolkit. In Proceedings of the international conference on spoken language processing (ICSLP02) (pp. 901–904), Denver, Colorado, USA. Google Scholar
  13. 13.
    Vidal, E., Casacuberta, F., Rodríguez, L., Civera, J., & Martínez, C. (2006). Computer-assisted translation using speech recognition. IEEE Transactions on Speech and Audio Processing, 14(3), 941–951. CrossRefGoogle Scholar
  14. 14.
    Young, S. J. (1994). The htk hidden Markov model toolkit: Design and philosophy. Entropic Cambridge Research Laboratory, Ltd, 2, 2–44. Google Scholar

Copyright information

© Springer-Verlag London Limited 2011

Authors and Affiliations

  • Alejandro Héctor Toselli
    • Enrique Vidal
      • Francisco Casacuberta

        There are no affiliations available

        Personalised recommendations