Skip to main content

Assessing PRESEMT

  • Chapter
  • First Online:
  • 504 Accesses

Part of the book series: SpringerBriefs in Statistics ((BRIEFSSTATIST))

Abstract

The topic of the current chapter is the evaluation of the performance of PRESEMT both per se as well as in comparison with other MT systems, the performance relating to the translation quality being achieved. While it is possible to employ humans for this task (subjective evaluation), who assess an MT system in terms of fluency (i.e. grammaticality) and adequacy (i.e. fidelity to the original text) (van Slype 1979), this being a laborious and time-consuming process, evaluation normally relies on automatic metrics (objective evaluation) that calculate the similarity between what an MT system produces (system output) and what it should have produced (reference translation).

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Notes

  1. 1.

    Human evaluation has only been carried out once within the PRESEMT project. For the corresponding results, the interested reader is referred to the project deliverable D9.2 (http://www.presemt.eu/files/Dels/PRESEMT_D9.2_supplement.pdf).

  2. 2.

    For our experiments, we have used the online version of SYSTRAN (www.systranet.com/translate) and WorldLingo (www.worldlingo.com/en/products_services/worldlingo_translator.html).

References

  • Banerjee S, Lavie A (2005) METEOR: An Automatic Metric for MT Evaluation with Improved Correlation with Human Judgments. In: Proceedings of Workshop on Intrinsic and Extrinsic Evaluation Measures for MT and/or Summarization at the 43rd Annual Meeting of the Association of Computational Linguistics (ACL-2005), Ann Arbor, Michigan, pp 65–72

    Google Scholar 

  • Biberauer T, Holmberg A, Roberts I, Sheehan M (2010) Parametric variation: null subjects in minimalist theory. Cambridge University Press

    Google Scholar 

  • Denkowski M, Lavie A (2011) Meteor 1.3: automatic metric for reliable optimization and evaluation of machine translation systems. In: Proceedings of the EMNLP 2011 Workshop on Statistical Machine Translation, Edinburgh, Scotland, pp 85–91

    Google Scholar 

  • Levenshtein VI (1966) Binary codes capable of correcting deletions, insertions, and reversals. Sov Phys Dokl 10:707–710

    MathSciNet  MATH  Google Scholar 

  • NIST (2002) Automatic evaluation of machine translation quality using n-gram co-occurrences statistics (available at: http://www.itl.nist.gov/iad/mig/tests/mt/doc/ngram-study.pdf)

  • Papineni K, Roukos S, Ward T, Zhu WJ (2002) BLEU: a method for automatic evaluation of machine translation. In: Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, Philadelphia, U.S.A., pp 311–318

    Google Scholar 

  • Snover M, Dorr B, Schwartz R, Micciulla L, Makhoul J (2006) A study of translation edit rate with targeted human annotation. In: Proceedings of the 7th AMTA Conference, Cambridge, MA, USA, pp 223–231

    Google Scholar 

  • Sofianopoulos S, Vassiliou M, Tambouratzis G (2012) Implementing a language-independent MT methodology. In: Proceedings of the 1st Workshop on Multilingual Modeling (held within ACL-2012), Jeju, Republic of Korea, pp 1–10

    Google Scholar 

  • Tambouratzis G, Vassiliou M, Sofianopoulos S (2016) Language-independent hybrid MT: comparative evaluation of translation quality. Chapter. In: Hybrid Approaches to Machine Translation, Costa-jussà, M.R., Rapp, R., Lambert, P., Eberle, K., Banchs, R.E., Babych, B. (Eds.). Springer-Verlag, pp 131–157. ISBN 978-3-319-21311-8.

    Google Scholar 

  • van Slype G (1979) Critical study of methods for evaluating the quality of machine translation. Technical Report BR19142, Bureau Marcel van Dijk/European Commission (DG XIII), Brussels (available at: http://issco-www.unige.ch/projects/isle/van-slype.pdf)

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to George Tambouratzis .

Rights and permissions

Reprints and permissions

Copyright information

© 2017 The Author(s)

About this chapter

Cite this chapter

Tambouratzis, G., Vassiliou, M., Sofianopoulos, S. (2017). Assessing PRESEMT. In: Machine Translation with Minimal Reliance on Parallel Resources. SpringerBriefs in Statistics. Springer, Cham. https://doi.org/10.1007/978-3-319-63107-3_4

Download citation

Publish with us

Policies and ethics