LORENZO PROIETTI

PhD Graduate

PhD program:: XXXVIII


supervisor: Roberto Navigli

Thesis title: Towards Better Measurement of Progress in Machine Translation: Evaluation and Meta-Evaluation

Reliable evaluation is central to Machine Learning. Without it, progress cannot be measured, claims of improvement are hard to verify, and research is misdirected. In Machine Translation, evaluation is particularly difficult, as many perfectly valid translations exist for the same source. Both the automatic metrics used to score translations and the procedures used to judge the evaluative ability of those metrics must account for this. This thesis examines weaknesses in current practices for evaluating translations and for evaluating the evaluators, and proposes solutions that keep measurements of progress fair, robust, and informative. First, we show that standard Machine Translation meta-evaluation practices can distort metric rankings and introduce a revised protocol that corrects these effects. Second, we anchor metric rankings to human performance: by estimating human baselines and placing metrics and humans on the same scale, we make performance headroom explicit and show that claims of human parity are fragile. Third, we show that easy test sets blunt progress measurement; we introduce a methodology for identifying hard-to-translate source texts and use it to construct more difficult benchmarks that resist saturation and preserve headroom. Finally, we reframe automatic evaluation as a pairwise decision between two candidate translations, improving accuracy on high-quality outputs while remaining efficient. Taken together, these contributions keep progress in Machine Translation measurable and trustworthy: meta-evaluation that reflects genuine evaluative ability, benchmarks that remain challenging as systems improve, and automatic evaluation with resolution on high-quality outputs.

Research products

11573/1743271 - 2024 - Beyond Correlation: Interpretable Evaluation of Machine Translation Metrics
Perrella, Stefano; Proietti, Lorenzo; Huguet Cabot, Pere-Lluis; Barba, Edoardo; Navigli, Roberto - 04b Atto di convegno in volume
conference: Conference on Empirical Methods in Natural Language Processing (Miami; Florida)
book: Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing - (9798891761643)

11573/1720213 - 2024 - Guardians of the Machine Translation Meta-Evaluation: Sentinel Metrics Fall In!
Perrella, Stefano; Proietti, Lorenzo; Scire', Alessandro; Barba, Edoardo; Navigli, Roberto - 04b Atto di convegno in volume
conference: Association for Computational Linguistics (Bangkok; Thailand)
book: Guardians of the Machine Translation Meta-Evaluation: Sentinel Metrics Fall In! - (9798891760943)

11573/1711963 - 2024 - Analyzing Homonymy Disambiguation Capabilities of Pretrained Language Models
Proietti, Lorenzo; Perrella, Stefano; Tedeschi, Simone; Vulpis, Giulia; Lavalle, Leonardo; Sanchietti, Andrea; Ferrari, Andrea; Navigli, Roberto - 04b Atto di convegno in volume
conference: 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024) (Torino; Italy)
book: Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024) - (978-2-493814-10-4)

11573/1670755 - 2022 - MaTESe: Machine Translation Evaluation as a Sequence Tagging Problem
Perrella, Stefano; Proietti, Lorenzo; Scirã, Alessandro; Campolungo, Niccolò; Navigli, Roberto - 04b Atto di convegno in volume
conference: Conference on Machine Translation (Abu Dhabi, United Arab Emirates)
book: Proceedings of the Seventh Conference on Machine Translation (WMT) - (9781959429296)

© Università degli Studi di Roma "La Sapienza" - Piazzale Aldo Moro 5, 00185 Roma