LLMs Are Zero-Shot Context-Aware Simultaneous Translators

Type: Preprint

Publication Date: 2024-06-19

Citations: 0

DOI: https://doi.org/10.48550/arxiv.2406.13476

Abstract

The advent of transformers has fueled progress in machine translation. More recently large language models (LLMs) have come to the spotlight thanks to their generality and strong performance in a wide range of language tasks, including translation. Here we show that open-source LLMs perform on par with or better than some state-of-the-art baselines in simultaneous machine translation (SiMT) tasks, zero-shot. We also demonstrate that injection of minimal background information, which is easy with an LLM, brings further performance gains, especially on challenging technical subject-matter. This highlights LLMs' potential for building next generation of massively multilingual, context-aware and terminologically accurate SiMT systems that require no resource-intensive training or fine-tuning.

Locations

  • arXiv (Cornell University) - View - PDF

Similar Works

Action Title Year Authors
+ SMaLL-100: Introducing Shallow Multilingual Machine Translation Model for Low-Resource Languages 2022 Alireza Mohammadshahi
Vassilina Nikoulina
Alexandre BĂ©rard
Caroline Brun
James Henderson
Laurent Besacier
+ PDF Chat What do Large Language Models Need for Machine Translation Evaluation? 2024 Shenbin Qian
Archchana Sindhujan
Minnie Kabra
Diptesh Kanojia
Constantin Orăsan
Tharindu Ranasinghe
Frédéric Blain
+ Simultaneous Machine Translation with Large Language Models 2023 Minghan Wang
Jinming Zhao
Thuy-Trang Vu
Fatemeh Shiri
Ehsan Shareghi
Gholamreza Haffari
+ Searching for Needles in a Haystack: On the Role of Incidental Bilingualism in PaLM’s Translation Capability 2023 Eleftheria Briakou
Colin Cherry
George Foster
+ Simul-LLM: A Framework for Exploring High-Quality Simultaneous Translation with Large Language Models 2023 Victor Agostinelli
Max Wild
Matthew Raffel
K.M Nafiur Rahman Fuad
Lizhong Chen
+ Towards Making the Most of Multilingual Pretraining for Zero-Shot Neural Machine Translation 2021 Guanhua Chen
Shuming Ma
Yun Chen
Dongdong Zhang
Jia Pan
Wenping Wang
Furu Wei
+ Searching for Needles in a Haystack: On the Role of Incidental Bilingualism in PaLM's Translation Capability 2023 Eleftheria Briakou
Colin Cherry
George Foster
+ Improving Massively Multilingual Neural Machine Translation and Zero-Shot Translation 2020 Biao Zhang
Philip Williams
Ivan Titov
Rico Sennrich
+ Improving Massively Multilingual Neural Machine Translation and Zero-Shot Translation 2020 Biao Zhang
Philip Williams
Ivan Titov
Rico Sennrich
+ PDF Chat When LLMs Struggle: Reference-less Translation Evaluation for Low-resource Languages 2025 Archchana Sindhujan
Diptesh Kanojia
Constantin Orăsan
Shenbin Qian
+ SCALE: Synergized Collaboration of Asymmetric Language Translation Engines 2023 Xin Cheng
Xun Wang
Tao Ge
Siqing Chen
Furu Wei
Dongyan Zhao
Rui Yan
+ Language Models are Good Translators 2021 Shuo Wang
Zhaopeng Tu
Zhixing Tan
Wenxuan Wang
Maosong Sun
Yang Liu
+ MALM: Mixing Augmented Language Modeling for Zero-Shot Machine Translation 2022 Kshitij Gupta
+ PDF Chat TransLLaMa: LLM-based Simultaneous Translation System 2024 Roman Koshkin
Katsuhito Sudoh
Satoshi Nakamura
+ PDF Chat Mufu: Multilingual Fused Learning for Low-Resource Translation with LLM 2024 Zheng Wei Lim
Nitish Gupta
Hong-Lin Yu
Trevor Cohn
+ PDF Chat Simultaneous Interpretation Corpus Construction by Large Language Models in Distant Language Pair 2024 Yusuke Sakai
Mana Makinae
Hidetaka Kamigaito
Taro Watanabe
+ PDF Chat Analyzing Context Contributions in LLM-based Machine Translation 2024 Emmanouil Zaranis
Nuno M. Guerreiro
André Martins
+ PDF Chat Building Accurate Translation-Tailored LLMs with Language Aware Instruction Tuning 2024 Changtong Zan
Liang Ding
Li Shen
Yibing Zhen
Weifeng Liu
Dacheng Tao
+ Improving Zero-Shot Translation of Low-Resource Languages 2018 Surafel M. Lakew
Quintino F. Lotito
Matteo Negri
Marco Turchi
Marcello Federico
+ PDF Chat Investigating the translation capabilities of Large Language Models trained on parallel data only 2024 Javier GarcĂ­a Gilabert
Carlos Escolano
Aleix Sant Savall
Francesca De Luca Fornaciari
Audrey Mash
Xixian Liao
Maite Melero

Works That Cite This (0)

Action Title Year Authors

Works Cited by This (0)

Action Title Year Authors