Creating and sharing knowledge for telecommunications

TOWER-V2: Unbabel-IST 2024 Submission for the General MT Shared Task

Pombal, J. ; Rei , R. ; Guerreiro, N. ; Alves, J. ; Martins, P. H. ; Fernandes, P. ; Wu, HW ; Vaz, T. ; Alves, D. A. ; Farajian, M. ; Agrawal, S. A. ; Farinhas, A. F. ; de Souza, J. G. ; Martins, A.

TOWER-V2: Unbabel-IST 2024 Submission for the General MT Shared Task, Proc Empirical Methods in Language Processing - EMNLP, Miami, United States, Vol. , pp. - , November, 2024.

Digital Object Identifier:

 

Abstract
In this work, we present TOWER-V2, an improved iteration of the state-of-the-art open-weight TOWER models, and the backbone of our submission to the WMT24 General Translation shared task. TOWER-V2 introduces key improvements including expanded language coverage, enhanced data quality, and increased model capacity up to 70B parameters. Our final submission combines these advancements with quality-aware decoding strategies, selecting translations based on multiple translation quality signals. The resulting system demonstrates significant improvement over previous versions, outperforming closed commercial systems like GPT-4O, CLAUDE-SONNET-3.5, and DEEPL even at a smaller 7B scale.