Multiligual machine translation aims at learning a single tanslation model for multiple languages. However, high resource language often suffers from performance degradation. In this blog, we present a method LaSS proposed in a recent ACL paper on multilingual neural machine translation. The LaSS is an approach to jointly train a single unified multilingual MT model and learns language-specific subnetwork for each language pair. Authors conducted experiments on IWSLT and WMT datasets with various Transformer architectures. The experimental results demonstrates average 1.2 BLEU improvements on 36 language pairs. LaSS shows strong generalization capabilty and demonstrates strong performance in zero-shot translation. Specifically, LaSS achieves 8.3 BLEU on 30 language pairs.
- Multilingual MT5
- Pre-training5
- BERT4
- Random Aligned Substitution2
- Zero-shot Translation2
- mRASP2
- Speech Translation2
- MT Evaluation2
- Transformer2
- Knowledge Assessment1
- Risk Ratio1
- Vocabulary Learning1
- Optimal Transport1
- Catastrophic Forgetting1
- Model Protection1
- Relation Extraction1
- Fact Verafication1
- Reasoning1
- Logic-regularized neural network1
- Embedding1
- Shared Semantic Memory1
- Chimera1
- Language Modelling1
- Imagination1
- Visual Machine Translation1
- ImagiT1
- Model Capacity1
- Language-specific Sub-network1
- GPU Acceleration1
- CUDA1
- Variational Inference1
- Latent Variable Model1
- Semi-supervised Learning1
- Contrastive Learning1
- BERTScore1
- COMET1
- Translation Memory1
- Recurrent Attention1
- Self-training1
- Unsupervised Machine Translation1
About 4 min