Skip to main content

Showing 1–8 of 8 results for author: Läubli, S

Searching in archive cs. Search in all archives.
.
  1. arXiv:2305.11140  [pdf, other

    cs.CL

    Exploiting Biased Models to De-bias Text: A Gender-Fair Rewriting Model

    Authors: Chantal Amrhein, Florian Schottmann, Rico Sennrich, Samuel Läubli

    Abstract: Natural language generation models reproduce and often amplify the biases present in their training data. Previous research explored using sequence-to-sequence rewriting models to transform biased model outputs (or original texts) into more gender-fair language by creating pseudo training data through linguistic rules. However, this approach is not practical for languages with more complex morphol… ▽ More

    Submitted 18 May, 2023; originally announced May 2023.

    Comments: accepted to ACL 2023

    ACM Class: I.2.7

  2. arXiv:2011.05978  [pdf, other

    cs.CL cs.HC

    The Impact of Text Presentation on Translator Performance

    Authors: Samuel Läubli, Patrick Simianer, Joern Wuebker, Geza Kovacs, Rico Sennrich, Spence Green

    Abstract: Widely used computer-aided translation (CAT) tools divide documents into segments such as sentences and arrange them in a side-by-side, spreadsheet-like view. We present the first controlled evaluation of these design choices on translator performance, measuring speed and accuracy in three experimental text processing tasks. We find significant evidence that sentence-by-sentence presentation enabl… ▽ More

    Submitted 11 November, 2020; originally announced November 2020.

    Comments: Accepted for publication in Target

  3. arXiv:2006.04781  [pdf, other

    cs.CL

    What's the Difference Between Professional Human and Machine Translation? A Blind Multi-language Study on Domain-specific MT

    Authors: Lukas Fischer, Samuel Läubli

    Abstract: Machine translation (MT) has been shown to produce a number of errors that require human post-editing, but the extent to which professional human translation (HT) contains such errors has not yet been compared to MT. We compile pre-translated documents in which MT and HT are interleaved, and ask professional translators to flag errors and post-edit these documents in a blind evaluation. We find th… ▽ More

    Submitted 8 June, 2020; originally announced June 2020.

    Comments: EAMT 2020 (Research Track)

  4. A Set of Recommendations for Assessing Human-Machine Parity in Language Translation

    Authors: Samuel Läubli, Sheila Castilho, Graham Neubig, Rico Sennrich, Qinlan Shen, Antonio Toral

    Abstract: The quality of machine translation has increased remarkably over the past years, to the degree that it was found to be indistinguishable from professional human translation in a number of empirical investigations. We reassess Hassan et al.'s 2018 investigation into Chinese to English news translation, showing that the finding of human-machine parity was owed to weaknesses in the evaluation design… ▽ More

    Submitted 3 April, 2020; originally announced April 2020.

    Journal ref: Journal of Artificial Intelligence Research 67 (2020) 653-672

  5. arXiv:1906.01685  [pdf, ps, other

    cs.CL

    Post-editing Productivity with Neural Machine Translation: An Empirical Assessment of Speed and Quality in the Banking and Finance Domain

    Authors: Samuel Läubli, Chantal Amrhein, Patrick Düggelin, Beatriz Gonzalez, Alena Zwahlen, Martin Volk

    Abstract: Neural machine translation (NMT) has set new quality standards in automatic translation, yet its effect on post-editing productivity is still pending thorough investigation. We empirically test how the inclusion of NMT, in addition to domain-specific translation memories and termbases, impacts speed and quality in professional translation of financial texts. We find that even with language pairs t… ▽ More

    Submitted 4 June, 2019; originally announced June 2019.

    Comments: MT Summit 2019 (Research Track)

  6. arXiv:1808.07048  [pdf, ps, other

    cs.CL

    Has Machine Translation Achieved Human Parity? A Case for Document-level Evaluation

    Authors: Samuel Läubli, Rico Sennrich, Martin Volk

    Abstract: Recent research suggests that neural machine translation achieves parity with professional human translation on the WMT Chinese--English news translation task. We empirically test this claim with alternative evaluation protocols, contrasting the evaluation of single sentences and entire documents. In a pairwise ranking experiment, human raters assessing adequacy and fluency show a stronger prefere… ▽ More

    Submitted 21 August, 2018; originally announced August 2018.

    Comments: EMNLP 2018

  7. arXiv:1703.04357  [pdf, other

    cs.CL

    Nematus: a Toolkit for Neural Machine Translation

    Authors: Rico Sennrich, Orhan Firat, Kyunghyun Cho, Alexandra Birch, Barry Haddow, Julian Hitschler, Marcin Junczys-Dowmunt, Samuel Läubli, Antonio Valerio Miceli Barone, Jozef Mokry, Maria Nădejde

    Abstract: We present Nematus, a toolkit for Neural Machine Translation. The toolkit prioritizes high translation accuracy, usability, and extensibility. Nematus has been used to build top-performing submissions to shared translation tasks at WMT and IWSLT, and has been used to train systems for production environments.

    Submitted 13 March, 2017; originally announced March 2017.

    Comments: EACL 2017 demo track

  8. arXiv:1605.05906  [pdf, other

    cs.CL

    Automatic TM Cleaning through MT and POS Tagging: Autodesk's Submission to the NLP4TM 2016 Shared Task

    Authors: Alena Zwahlen, Olivier Carnal, Samuel Läubli

    Abstract: We describe a machine learning based method to identify incorrect entries in translation memories. It extends previous work by Barbu (2015) through incorporating recall-based machine translation and part-of-speech-tagging features. Our system ranked first in the Binary Classification (II) task for two out of three language pairs: English-Italian and English-Spanish.

    Submitted 19 May, 2016; originally announced May 2016.

    Comments: Presented at the 2nd Workshop on Natural Language Processing for Translation Memories (NLP4TM 2016)