Reading List for Multilingual NLP

bib file for this list.

Bilingual Translation

  1. Jiang, Qingnan, Mingxuan Wang, Jun Cao, Shanbo Cheng, Shujian Huang, and Lei Li. 2021. “Learning Kernel-Smoothed Machine Translation with Retrieved Examples.” In Proc. Of EMNLP, 7280–90. https://aclanthology.org/2021.emnlp-main.579.
  2. Khandelwal, Urvashi, Angela Fan, Dan Jurafsky, Luke Zettlemoyer, and Mike Lewis. 2021. “Nearest Neighbor Machine Translation.” In Proc. Of ICLR. https://openreview.net/forum?id=7wCBOfJ8hJM.
  3. Liang, Jianze, Chengqi Zhao, Mingxuan Wang, Xipeng Qiu, and Lei Li. 2021. “Finding Sparse Structure for Domain Specific Neural Machine Translation.” In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI).
  4. Sun, Zewei, Mingxuan Wang, Hao Zhou, Chengqi Zhao, Shujian Huang, Jiajun Chen, and Lei Li. 2022. “Rethinking Document-Level Neural Machine Translation.” In Findings of the Association for Computational Linguistics: ACL 2022, 3537–48. https://aclanthology.org/2022.findings-acl.279.
  5. Xu, Jingjing, Hao Zhou, Chun Gan, Zaixiang Zheng, and Lei Li. 2021. “Vocabulary Learning via Optimal Transport for Neural Machine Translation.” In Proc. Of ACL, 7361–73. https://aclanthology.org/2021.acl-long.571.

Multilingual Translation

  1. Yuan, Fei, Yinquan Lu, Wenhao Zhu, Lingpeng Kong, Lei Li, and Jingjing Xu. 2023. “Lego-MT: Learning Detachable Models for Massively Multilingual Machine Translation” In The 61st Annual Meeting of the Association for Computational Linguistics - Findings (ACL-Findings).
  2. Costa-jussà, Marta Ruiz, James Cross, Onur cCelebi, Maha Elbayad, Kenneth Heafield, Kevin Heffernan, Elahe Kalbassi, et al. 2022. “No Language Left Behind: Scaling Human-Centered Machine Translation.” ArXiv Preprint abs/2207.04672. https://arxiv.org/abs/2207.04672.
  3. Pan, Xiao, Mingxuan Wang, Liwei Wu, and Lei Li. 2021. “Contrastive Learning for Many-to-Many Multilingual Neural Machine Translation.” In Proc. Of ACL, 244–58. https://aclanthology.org/2021.acl-long.21.
  4. Fan, Angela, Shruti Bhosale, Holger Schwenk, Zhiyi Ma, Ahmed El-Kishky, Siddharth Goyal, Mandeep Baines, et al. 2021. “Beyond English-Centric Multilingual Machine Translation.” J. Mach. Learn. Res. 22: 107:1–48. http://jmlr.org/papers/v22/20-1307.html.
  5. Garcia, Xavier, Yamini Bansal, Colin Cherry, George Foster, Maxim Krikun, Fangxiaoyu Feng, Melvin Johnson, and Orhan Firat. 2023. “The Unreasonable Effectiveness of Few-Shot Learning for Machine Translation.” ArXiv Preprint abs/2302.01398. https://arxiv.org/abs/2302.01398.
  6. Guerreiro, Nuno M, Duarte Alves, Jonas Waldendorf, Barry Haddow, Alexandra Birch, Pierre Colombo, and André FT Martins. 2023. “Hallucinations in Large Multilingual Translation Models.” ArXiv Preprint abs/2303.16104. https://arxiv.org/abs/2303.16104.
  7. Lin, Zehui, Xiao Pan, Mingxuan Wang, Xipeng Qiu, Jiangtao Feng, Hao Zhou, and Lei Li. 2020. “Pre-Training Multilingual Neural Machine Translation by Leveraging Alignment Information.” In Proc. Of EMNLP, 2649–63. https://aclanthology.org/2020.emnlp-main.210.
  8. Lin, Zehui, Liwei Wu, Mingxuan Wang, and Lei Li. 2021. “Learning Language Specific Sub-Network for Multilingual Machine Translation.” In Proc. Of ACL, 293–305. https://aclanthology.org/2021.acl-long.25.
  9. Zhang, Biao, Ankur Bapna, Rico Sennrich, and Orhan Firat. 2021. “Share or Not? Learning to Schedule Language-Specific Capacity for Multilingual Translation.” In Proc. Of ICLR. https://openreview.net/forum?id=Wj4ODo0uyCF.

Large Language Models for MT

  1. Zhu, Wenhao, Hongyi Liu, Qingxiu Dong, Jingjing Xu, shujian Huang, Lingpeng Kong, Jiajun Chen, and Lei Li. 2023. “Multilingual Machine Translation with Large Language Models: Empirical Results and Analysis.” ArXiv Preprint abs/2304.04675. https://arxiv.org/abs/2304.04675.
  2. Agrawal, Sweta, Chunting Zhou, Mike Lewis, Luke Zettlemoyer, and Marjan Ghazvininejad. 2023. “In-Context Examples Selection for Machine Translation.” In Findings of the Association for Computational Linguistics: ACL 2023, 8857–73. https://aclanthology.org/2023.findings-acl.564.
  3. Vilar, David, Markus Freitag, Colin Cherry, Jiaming Luo, Viresh Ratnakar, and George Foster. 2023. “Prompting PaLM for Translation: Assessing Strategies and Performance.” In Proc. Of ACL, 15406–27. https://aclanthology.org/2023.acl-long.859.
  4. Sun, Zewei, Mingxuan Wang, and Lei Li. 2021. “Multilingual Translation via Grafting Pre-Trained Language Models.” In Findings of the Association for Computational Linguistics: EMNLP 2021, 2735–47. https://aclanthology.org/2021.findings-emnlp.233.
  5. Yang, Jiacheng, Mingxuan Wang, Hao Zhou, Chengqi Zhao, Weinan Zhang, Yong Yu, and Lei Li. 2020. “Towards Making the Most of BERT in Neural Machine Translation.” In The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, the Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, the Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, 9378–85. https://aaai.org/ojs/index.php/AAAI/article/view/6479.

Non-Autoregressive Translation

  1. Huang, Fei, Hao Zhou, Yang Liu, Hang Li, and Minlie Huang. 2022. “Directed Acyclic Transformer for Non-Autoregressive Machine Translation.” In International Conference on Machine Learning, ICML 2022, 17-23 July 2022, Baltimore, Maryland, USA, edited by Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvári, Gang Niu, and Sivan Sabato, 162:9410–28. Proceedings of Machine Learning Research. https://proceedings.mlr.press/v162/huang22m.html.
  2. Yu Bao, Hao Zhou, Shujian Huang, Dongqi Wang, Lihua Qian, Xinyu Dai, Jiajun Chen, and Lei Li. "latent-GLAT: Glancing at Latent Variables for Parallel Text Generation." ACL 2022. https://aclanthology.org/2022.acl-long.575/
  3. Qian, Lihua, Hao Zhou, Yu Bao, Mingxuan Wang, Lin Qiu, Weinan Zhang, Yong Yu, and Lei Li. 2021. “Glancing Transformer for Non-Autoregressive Neural Machine Translation.” In Proc. Of ACL, 1993–2003. https://aclanthology.org/2021.acl-long.155.
  4. Huang, Chenyang, Hao Zhou, Osmar R. Zaı̈ane, Lili Mou, and Lei Li. 2022. “Non-Autoregressive Translation with Layer-Wise Prediction and Deep Supervision.” In Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 2022, Thirty-Fourth Conference on Innovative Applications of Artificial Intelligence, IAAI 2022, the Twelveth Symposium on Educational Advances in Artificial Intelligence, EAAI 2022 Virtual Event, February 22 - March 1, 2022, 10776–84. https://ojs.aaai.org/index.php/AAAI/article/view/21323.
  5. Song, Zhenqiao, Hao Zhou, Lihua Qian, Jingjing Xu, Shanbo Cheng, Mingxuan Wang, and Lei Li. 2022. “Switch-GLAT: Multilingual Parallel Machine Translation via Code-Switch Decoder.” In Proc. Of ICLR. https://openreview.net/forum?id=5HvpvYd68b.
  6. Zhu, Yaoming, Jiangtao Feng, Chengqi Zhao, Mingxuan Wang, and Lei Li. 2021. “Counter-Interference Adapter for Multilingual Machine Translation.” In Findings of the Association for Computational Linguistics: EMNLP 2021, 2812–23. https://aclanthology.org/2021.findings-emnlp.240.
  7. Zheng, Zaixiang, Hao Zhou, Shujian Huang, Jiajun Chen, Jingjing Xu, and Lei Li. 2021. “Duplex Sequence-to-Sequence Learning for Reversible Machine Translation.” In Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, December 6-14, 2021, Virtual, edited by Marc’Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan, 21070–84. https://proceedings.neurips.cc/paper/2021/hash/afecc60f82be41c1b52f6705ec69e0f1-Abstract.html.

Multilingual Representation Learning

  1. Conneau, Alexis, Kartikay Khandelwal, Naman Goyal, Vishrav Chaudhary, Guillaume Wenzek, Francisco Guzmán, Edouard Grave, Myle Ott, Luke Zettlemoyer, and Veselin Stoyanov. 2020. “Unsupervised Cross-Lingual Representation Learning at Scale.” In Proc. Of ACL, 8440–51. https://aclanthology.org/2020.acl-main.747.
  2. Lewis, Mike, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Veselin Stoyanov, and Luke Zettlemoyer. 2020. BART: Denoising Sequence-to-Sequence Pre-Training for Natural Language Generation, Translation, and Comprehension.” In Proc. Of ACL, 7871–80. https://aclanthology.org/2020.acl-main.703.
  3. Liu, Yinhan, Jiatao Gu, Naman Goyal, Xian Li, Sergey Edunov, Marjan Ghazvininejad, Mike Lewis, and Luke Zettlemoyer. 2020. “Multilingual Denoising Pre-Training for Neural Machine Translation.” Transactions of the Association for Computational Linguistics 8: 726–42. https://aclanthology.org/2020.tacl-1.47.
  4. Ouyang, Xuan, Shuohuan Wang, Chao Pang, Yu Sun, Hao Tian, Hua Wu, and Haifeng Wang. 2021. ERNIE-M: Enhanced Multilingual Representation by Aligning Cross-Lingual Semantics with Monolingual Corpora.” In Proc. Of EMNLP, 27–38. https://aclanthology.org/2021.emnlp-main.3.
  5. Xue, Linting, Noah Constant, Adam Roberts, Mihir Kale, Rami Al-Rfou, Aditya Siddhant, Aditya Barua, and Colin Raffel. 2021. “MT5: A Massively Multilingual Pre-Trained Text-to-Text Transformer.” In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 483–98. https://aclanthology.org/2021.naacl-main.41.

Cross-lingual Transfer

  1. Artetxe, Mikel, Vedanuj Goswami, Shruti Bhosale, Angela Fan, and Luke Zettlemoyer. 2023. “Revisiting Machine Translation for Cross-Lingual Classification.” ArXiv Preprint abs/2305.14240. https://arxiv.org/abs/2305.14240.
  2. Machel Reid, Mikel Artetxe. 2023. “On the Role of Parallel Data in Cross-Lingual Transfer Learning.” In Findings of the Association for Computational Linguistics: ACL 2023, 5999–6006. https://aclanthology.org/2023.findings-acl.372.
  3. Wang, Danqing, Jiaze Chen, Hao Zhou, Xipeng Qiu, and Lei Li. 2021. “Contrastive Aligned Joint Learning for Multilingual Summarization.” In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, 2739–50. https://aclanthology.org/2021.findings-acl.242.
  4. Wu, Qianhui, Zijia Lin, Börje F. Karlsson, Biqing Huang, and Jianguang Lou. 2020. “UniTrans : Unifying Model Transfer and Data Transfer for Cross-Lingual Named Entity Recognition with Unlabeled Data.” In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI 2020, edited by Christian Bessiere, 3926–32. https://doi.org/10.24963/ijcai.2020/543.
  5. Xia, Mengzhou, Guoqing Zheng, Subhabrata Mukherjee, Milad Shokouhi, Graham Neubig, and Ahmed Hassan Awadallah. 2021. MetaXL: Meta Representation Transformation for Low-Resource Cross-Lingual Learning.” In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 499–511. https://aclanthology.org/2021.naacl-main.42.
  6. Yang, Huiyun, Huadong Chen, Hao Zhou, and Lei Li. 2022. “Enhancing Cross-Lingual Transfer by Manifold Mixup.” In Proc. Of ICLR. https://openreview.net/forum?id=OjPmfr9GkVv.

Speech Translation

  1. Dong, Qianqian, Yaoming Zhu, Mingxuan Wang, and Lei Li. 2022. “Learning When to Translate for Streaming Speech.” In Proc. Of ACL, 680–94. https://aclanthology.org/2022.acl-long.50.
  2. Rong Ye, Mingxuan Wang, Lei Li. 2022. “Cross-Modal Contrastive Learning for Speech Translation.” In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 5099–5113. https://aclanthology.org/2022.naacl-main.376.
  3. Dong, Qianqian, Mingxuan Wang, Hao Zhou, Shuang Xu, Bo Xu, and Lei Li. 2021. “Consecutive Decoding for Speech-to-Text Translation.” In Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, IAAI 2021, the Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, Virtual Event, February 2-9, 2021, 12738–48. https://ojs.aaai.org/index.php/AAAI/article/view/17508.
  4. Han, Chi, Mingxuan Wang, Heng Ji, and Lei Li. 2021. “Learning Shared Semantic Space for Speech-to-Text Translation.” In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, 2214–25. https://aclanthology.org/2021.findings-acl.195.
  5. Ouyang, Siqi, Rong Ye, and Lei Li. 2023. WACO: Word-Aligned Contrastive Learning for Speech Translation.” In Proc. Of ACL, 3891–3907. https://aclanthology.org/2023.acl-long.216.
  6. Ye, Rong, Mingxuan Wang, and Lei Li. 2021. “End-to-End Speech Translation via Cross-Modal Progressive Training.” In Proc. Of INTERSPEECH, edited by Hynek Hermansky, Honza Cernocký, Lukás Burget, Lori Lamel, Odette Scharenborg, and Petr Motlı́cek, 2267–71. https://doi.org/10.21437/Interspeech.2021-1065.
  7. Zheng, Renjie, Junkun Chen, Mingbo Ma, and Liang Huang. 2021. “Fused Acoustic and Text Encoding for Multimodal Bilingual Pretraining and Speech Translation.” In Proc. Of ICML, edited by Marina Meila and Tong Zhang, 139:12736–46. Proceedings of Machine Learning Research. http://proceedings.mlr.press/v139/zheng21a.html.
  8. Ma, Xutai, Yongqiang Wang, Mohammad Javad Dousti, Philipp Koehn, and Juan Pino. 2021. “Streaming Simultaneous Speech Translation with Augmented Memory Transformer,” 7523–27.

Speech Representation Learning

  1. Babu, Arun, Changhan Wang, Andros Tjandra, Kushal Lakhotia, Qiantong Xu, Naman Goyal, Kritika Singh, et al. 2021. “XLS-r: Self-Supervised Cross-Lingual Speech Representation Learning at Scale.” ArXiv Preprint. https://arxiv.org/abs/2111.09296.
  2. Baevski, Alexei, Yuhao Zhou, Abdelrahman Mohamed, and Michael Auli. 2020. “Wav2vec 2.0: A Framework for Self-Supervised Learning of Speech Representations.” In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, Virtual, edited by Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin. https://proceedings.neurips.cc/paper/2020/hash/92d1e1eb1cd6f9fba3227870bb6d7f07-Abstract.html.
  3. Hsu, Wei-Ning, Benjamin Bolte, Yao-Hung Hubert Tsai, Kushal Lakhotia, Ruslan Salakhutdinov, and Abdelrahman Mohamed. 2021. “HuBERT: Self-Supervised Speech Representation Learning by Masked Prediction of Hidden Units.” IEEE/ACM Transactions on Audio, Speech, and Language Processing 29: 3451–60.
  4. Tang, Yun, Hongyu Gong, Ning Dong, Changhan Wang, Wei-Ning Hsu, Jiatao Gu, Alexei Baevski, et al. 2022. “Unified Speech-Text Pre-Training for Speech Translation and Recognition.” In Proc. Of ACL, 1488–99. https://aclanthology.org/2022.acl-long.105.
  5. Wang, Chengyi, Yu Wu, Yao Qian, Ken’ichi Kumatani, Shujie Liu, Furu Wei, Michael Zeng, and Xuedong Huang. 2021. “UniSpeech: Unified Speech Representation Learning with Labeled and Unlabeled Data.” In Proc. Of ICML, edited by Marina Meila and Tong Zhang, 139:10937–47. Proceedings of Machine Learning Research. http://proceedings.mlr.press/v139/wang21y.html.
  6. Zhang, Ziqiang, Long Zhou, Junyi Ao, Shujie Liu, Lirong Dai, Jinyu Li, and Furu Wei. 2022. SpeechUT: Bridging Speech and Text with Hidden-Unit for Encoder-Decoder Based Speech-Text Pre-Training.” In Proc. Of EMNLP, 1663–76. https://aclanthology.org/2022.emnlp-main.108.

Automatic Speech Recognition

  1. Alec Radford, Jong Wook Kim, Tao Xu, Greg Brockman, Christine Mcleavey, Ilya Sutskever. Robust Speech Recognition via Large-Scale Weak Supervision. Proceedings of the 40th International Conference on Machine Learning, PMLR 202:28492-28518, 2023. https://proceedings.mlr.press/v202/radford23a.html
  2. Viet Anh Trinh, Pegah Ghahremani, Brian King, Jasha Droppo, Andreas Stolcke, Roland Maas. Reducing Geographic Disparities in Automatic Speech Recognition via Elastic Weight Consolidation. Interspeech 2022.
  3. Vineel Pratap, Andros Tjandra, Bowen Shi, Paden Tomasello, Arun Babu, Sayani Kundu, Ali Elkahky, Zhaoheng Ni, Apoorv Vyas, Maryam Fazel-Zarandi, Alexei Baevski, Yossi Adi, Xiaohui Zhang, Wei-Ning Hsu, Alexis Conneau, Michael Auli. Scaling Speech Technology to 1,000+ Languages. 2023. https://arxiv.org/abs/2305.13516
  4. Zhang et al. Google USM: Scaling Automatic Speech Recognition Beyond 100 Languages, 2023.
  5. Kazuya Kawakami, Luyu Wang, Chris Dyer, Phil Blunsom, Aaron van den Oord. Learning Robust and Multilingual Speech Representations. EMNLP 2020.

Speech Synthesis

  1. Aidan Pine, Dan Wells, Nathan Brinklow, Patrick Littell, Korin Richmond. Requirements and Motivations of Low-Resource Speech Synthesis for Language Revitalization. ACL 2022.
  2. Xu Tan, Jiawei Chen, Haohe Liu, Jian Cong, Chen Zhang, Yanqing Liu, Xi Wang, Yichong Leng, Yuanhao Yi, Lei He, Frank Soong, Tao Qin, Sheng Zhao, Tie-Yan Liu. NaturalSpeech: End-to-End Text to Speech Synthesis with Human-Level Quality, 2022. https://arxiv.org/abs/2205.04421

Corpus

  1. Chen, Yiran, Zhenqiao Song, Xianze Wu, Danqing Wang, Jingjing Xu, Jiaze Chen, Hao Zhou, and Lei Li. 2022. MTG: A Benchmark Suite for Multilingual Text Generation.” In Findings of the Association for Computational Linguistics: NAACL 2022, 2508–27. https://aclanthology.org/2022.findings-naacl.192.
  2. Goyal, Naman, Cynthia Gao, Vishrav Chaudhary, Peng-Jen Chen, Guillaume Wenzek, Da Ju, Sanjana Krishnan, Marc’Aurelio Ranzato, Francisco Guzmán, and Angela Fan. 2022. “The Flores-101 Evaluation Benchmark for Low-Resource and Multilingual Machine Translation.” Transactions of the Association for Computational Linguistics 10: 522–38. https://aclanthology.org/2022.tacl-1.30.
  3. Ruder, Sebastian, Noah Constant, Jan Botha, Aditya Siddhant, Orhan Firat, Jinlan Fu, Pengfei Liu, et al. 2021. XTREME-R: Towards More Challenging and Nuanced Multilingual Evaluation.” In Proc. Of EMNLP, 10215–45. https://aclanthology.org/2021.emnlp-main.802.
  4. Wang, Changhan, Anne Wu, Jiatao Gu, and Juan Pino. 2021. “CoVoST 2 and Massively Multilingual Speech Translation.” In Proc. Of INTERSPEECH, edited by Hynek Hermansky, Honza Cernocký, Lukás Burget, Lori Lamel, Odette Scharenborg, and Petr Motlı́cek, 2247–51. https://doi.org/10.21437/Interspeech.2021-2027.

Evaluation

  1. Xu, Wenda, Danqing Wang, Liangming Pan, Zhenqiao Song, Markus Freitag, William Yang Wang, and Lei Li. 2023. “Instructscore: Towards Explainable Text Generation Evaluation with Automatic Feedback.” ArXiv Preprint abs/2305.14282. https://arxiv.org/abs/2305.14282.
  2. Xu, Wenda, Xian Qian, Mingxuan Wang, Lei Li, and William Yang Wang. 2023. “SESCORE2: Learning Text Generation Evaluation via Synthesizing Realistic Mistakes.” In The 61st Annual Meeting of the Association for Computational Linguistics (ACL).
  3. Xu, Wenda, Yi-Lin Tuan, Yujie Lu, Michael Saxon, Lei Li, and William Yang Wang. 2022. “Not All Errors Are Equal: Learning Text Generation Metrics Using Stratified Error Synthesis.” In Findings of the Association for Computational Linguistics: EMNLP 2022, 6559–74. https://aclanthology.org/2022.findings-emnlp.489.
  4. Fomicheva, Marina, Shuo Sun, Lisa Yankovskaya, Frédéric Blain, Francisco Guzmán, Mark Fishel, Nikolaos Aletras, Vishrav Chaudhary, and Lucia Specia. 2020. “Unsupervised Quality Estimation for Neural Machine Translation.” Transactions of the Association for Computational Linguistics 8: 539–55. https://aclanthology.org/2020.tacl-1.35.
  5. Zhang*, Tianyi, Varsha Kishore*, Felix Wu*, Kilian Q. Weinberger, and Yoav Artzi. “BERTScore: Evaluating Text Generation with BERT,” ICLR 2020. https://openreview.net/forum?id=SkeHuCVFDr.
  6. Thibault Sellam, Dipanjan Das, Ankur Parikh. BLEURT: Learning Robust Metrics for Text Generation, ACL 2020.
  7. Ricardo Rei, Craig Stewart, Ana C Farinha, Alon Lavie. COMET: A Neural Framework for MT Evaluation, EMNLP 2020.

Additional Reading

  1. Wang, Xiaohui, Ying Xiong, Yang Wei, Mingxuan Wang, and Lei Li. 2021. LightSeq: A High Performance Inference Library for Transformers.” In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Industry Papers, 113–20. https://aclanthology.org/2021.naacl-industry.15.
  2. Zhao, Chengqi, Mingxuan Wang, Qianqian Dong, Rong Ye, and Lei Li. 2021. NeurST: Neural Speech Translation Toolkit.” In Proc. Of ACL, 55–62. https://aclanthology.org/2021.acl-demo.7.
  3. Artetxe, Mikel, and Holger Schwenk. 2019. “Massively Multilingual Sentence Embeddings for Zero-Shot Cross-Lingual Transfer and Beyond.” Transactions of the Association for Computational Linguistics 7: 597–610. https://aclanthology.org/Q19-1038.
  4. Bahdanau, Dzmitry, Kyunghyun Cho, and Yoshua Bengio. 2015. “Neural Machine Translation by Jointly Learning to Align and Translate.” In Proc. Of ICLR, edited by Yoshua Bengio and Yann LeCun. http://arxiv.org/abs/1409.0473.
  5. Bapna, Ankur, and Orhan Firat. 2019. “Simple, Scalable Adaptation for Neural Machine Translation.” In Proc. Of EMNLP, 1538–48. https://aclanthology.org/D19-1165.
  6. Di Gangi, Mattia A., Roldano Cattoni, Luisa Bentivogli, Matteo Negri, and Marco Turchi. 2019. MuST-C: A Multilingual Speech Translation Corpus.” In Proc. Of NAACL-HLT, 2012–17. https://aclanthology.org/N19-1202.
  7. Conneau, Alexis, and Guillaume Lample. 2019. “Cross-Lingual Language Model Pretraining.” In Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada, edited by Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d’Alché-Buc, Emily B. Fox, and Roman Garnett, 7057–67. https://proceedings.neurips.cc/paper/2019/hash/c04c19c2c2474dbf5f7ac4372c5b9af1-Abstract.html.
  8. Conneau, Alexis, Shijie Wu, Haoran Li, Luke Zettlemoyer, and Veselin Stoyanov. 2020. “Emerging Cross-Lingual Structure in Pretrained Language Models.” In Proc. Of ACL, 6022–34. https://aclanthology.org/2020.acl-main.536.
  9. Dong, Linhao, Shuang Xu, and Bo Xu. 2018. “Speech-Transformer: A No-Recurrence Sequence-to-Sequence Model for Speech Recognition.” In 2018 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2018, Calgary, AB, Canada, April 15-20, 2018, 5884–88. https://doi.org/10.1109/ICASSP.2018.8462506.
  10. Dalvi, Fahim, Nadir Durrani, Hassan Sajjad, and Stephan Vogel. 2018. “Incremental Decoding and Training Methods for Simultaneous Translation in Neural Machine Translation.” In Proc. Of NAACL-HLT, 493–99. https://aclanthology.org/N18-2079.
  11. Elbayad, Maha, Laurent Besacier, and Jakob Verbeek. 2020. “Efficient Wait-k Models for Simultaneous Machine Translation.” In Proc. Of INTERSPEECH, edited by Helen Meng, Bo Xu, and Thomas Fang Zheng, 1461–65. https://doi.org/10.21437/Interspeech.2020-1241.
  12. Firat, Orhan, Kyunghyun Cho, and Yoshua Bengio. 2016. “Multi-Way, Multilingual Neural Machine Translation with a Shared Attention Mechanism.” In Proc. Of NAACL-HLT, 866–75. https://aclanthology.org/N16-1101.
  13. Arivazhagan, Naveen, Ankur Bapna, Orhan Firat, Dmitry Lepikhin, Melvin Johnson, Maxim Krikun, Mia Xu Chen, et al. 2019. “Massively Multilingual Neural Machine Translation in the Wild: Findings and Challenges.” ArXiv Preprint abs/1907.05019. https://arxiv.org/abs/1907.05019.
  14. Saxon, Michael, Samridhi Choudhary, Joseph P. McKenna, and Athanasios Mouchtaris. 2021. “End-to-End Spoken Language Understanding for Generalized Voice Assistants.” In Proc. Of INTERSPEECH, edited by Hynek Hermansky, Honza Cernocký, Lukás Burget, Lori Lamel, Odette Scharenborg, and Petr Motlı́cek, 4738–42. https://doi.org/10.21437/Interspeech.2021-1826.
  15. Gu, Jiatao, James Bradbury, Caiming Xiong, Victor O. K. Li, and Richard Socher. 2018. “Non-Autoregressive Neural Machine Translation.” In Proc. Of ICLR. https://openreview.net/forum?id=B1l8BtlCb.
  16. He, Xiaodong, Li Deng, and Alex Acero. 2011. “Why Word Error Rate Is Not a Good Metric for Speech Recognizer Training for the Speech Translation Task?” In Proc. Of ICASSP, 5632–35. IEEE.
  17. Hu, Junjie, Sebastian Ruder, Aditya Siddhant, Graham Neubig, Orhan Firat, and Melvin Johnson. 2020. XTREME: A Massively Multilingual Multi-Task Benchmark for Evaluating Cross-Lingual Generalisation.” In Proc. Of ICML, 119:4411–21. Proceedings of Machine Learning Research. http://proceedings.mlr.press/v119/hu20b.html.
  18. Jia, Ye, Ron J. Weiss, Fadi Biadsy, Wolfgang Macherey, Melvin Johnson, Zhifeng Chen, and Yonghui Wu. 2019. “Direct Speech-to-Speech Translation with a Sequence-to-Sequence Model.” In Proc. Of INTERSPEECH, edited by Gernot Kubin and Zdravko Kacic, 1123–27. https://doi.org/10.21437/Interspeech.2019-1951.
  19. Johnson, Melvin, Mike Schuster, Quoc V. Le, Maxim Krikun, Yonghui Wu, Zhifeng Chen, Nikhil Thorat, et al. 2017. Googles Multilingual Neural Machine Translation System: Enabling Zero-Shot Translation.” Transactions of the Association for Computational Linguistics 5: 339–51. https://aclanthology.org/Q17-1024.
  20. Libovický, Jindřich, and Jindřich Helcl. 2018. “End-to-End Non-Autoregressive Neural Machine Translation with Connectionist Temporal Classification.” In Proc. Of EMNLP, 3016–21. https://aclanthology.org/D18-1336.
  21. Pan, Xiaoman, Boliang Zhang, Jonathan May, Joel Nothman, Kevin Knight, and Heng Ji. 2017. “Cross-Lingual Name Tagging and Linking for 282 Languages.” In Proc. Of ACL, 1946–58. https://aclanthology.org/P17-1178.
  22. Park, Daniel S., William Chan, Yu Zhang, Chung-Cheng Chiu, Barret Zoph, Ekin D. Cubuk, and Quoc V. Le. 2019. “SpecAugment: A Simple Data Augmentation Method for Automatic Speech Recognition.” In Proc. Of INTERSPEECH, edited by Gernot Kubin and Zdravko Kacic, 2613–17. https://doi.org/10.21437/Interspeech.2019-2680.
  23. Pires, Telmo, Eva Schlinger, and Dan Garrette. 2019. “How Multilingual Is Multilingual BERT?” In Proc. Of ACL, 4996–5001. https://aclanthology.org/P19-1493.
  24. Post, Matt. 2018. “A Call for Clarity in Reporting BLEU Scores.” In Proceedings of the Third Conference on Machine Translation: Research Papers, 186–91. https://aclanthology.org/W18-6319.
  25. Qi, Ye, Devendra Sachan, Matthieu Felix, Sarguna Padmanabhan, and Graham Neubig. 2018. “When and Why Are Pre-Trained Word Embeddings Useful for Neural Machine Translation?” In Proc. Of NAACL-HLT, 529–35. https://aclanthology.org/N18-2084.
  26. Schneider, Steffen, Alexei Baevski, Ronan Collobert, and Michael Auli. 2019. “Wav2vec: Unsupervised Pre-Training for Speech Recognition.” In Proc. Of INTERSPEECH, edited by Gernot Kubin and Zdravko Kacic, 3465–69. https://doi.org/10.21437/Interspeech.2019-1873.
  27. Schuster, Tal, Ori Ram, Regina Barzilay, and Amir Globerson. 2019. “Cross-Lingual Alignment of Contextual Word Embeddings, with Applications to Zero-Shot Dependency Parsing.” In Proc. Of NAACL-HLT, 1599–1613. https://aclanthology.org/N19-1162.
  28. ———. 2016b. “Neural Machine Translation of Rare Words with Subword Units.” In Proc. Of ACL, 1715–25. https://aclanthology.org/P16-1162.
  29. Sutskever, Ilya, Oriol Vinyals, and Quoc V Le. 2014. “Sequence to Sequence Learning with Neural Networks.” In Advances in Neural Information Processing Systems, 3104–12. https://papers.nips.cc/paper/5346-sequence-to-sequence-learning-with-neural-networks.pdf.
  30. Thompson, Brian, Jeremy Gwinnup, Huda Khayrallah, Kevin Duh, and Philipp Koehn. 2019. “Overcoming Catastrophic Forgetting During Domain Adaptation of Neural Machine Translation.” In Proc. Of NAACL-HLT, 2062–68. https://aclanthology.org/N19-1209.
  31. Wang, Changhan, Juan Pino, Anne Wu, and Jiatao Gu. 2020. CoVoST: A Diverse Multilingual Speech-to-Text Translation Corpus.” In Proceedings of the Twelfth Language Resources and Evaluation Conference, 4197–4203. https://aclanthology.org/2020.lrec-1.517.
  32. Wang, Qiang, Bei Li, Tong Xiao, Jingbo Zhu, Changliang Li, Derek F. Wong, and Lidia S. Chao. 2019. “Learning Deep Transformer Models for Machine Translation.” In Proc. Of ACL, 1810–22. https://aclanthology.org/P19-1176.
  33. Zheng, Zaixiang, Hao Zhou, Shujian Huang, Lei Li, Xinyu Dai, and Jiajun Chen. 2020. “Mirror Generative Models for Neural Machine Translation.” In International Conference on Learning Representations (ICLR).
  34. Alinejad, Ashkan, and Anoop Sarkar. 2020. “Effectively Pretraining a Speech Translation Decoder with Machine Translation Data.” In Proc. Of EMNLP, 8014–20. https://aclanthology.org/2020.emnlp-main.644.
  35. Anastasopoulos, Antonios, and David Chiang. 2018. “Tied Multitask Learning for Neural Speech Translation.” In Proc. Of NAACL-HLT, 82–91. https://aclanthology.org/N18-1008.
  36. Ao, Junyi, Rui Wang, Long Zhou, Chengyi Wang, Shuo Ren, Yu Wu, Shujie Liu, et al. 2022. SpeechT5: Unified-Modal Encoder-Decoder Pre-Training for Spoken Language Processing.” In Proc. Of ACL, 5723–38. https://aclanthology.org/2022.acl-long.393.
  37. Aharoni, Roee, Melvin Johnson, and Orhan Firat. 2019. “Massively Multilingual Neural Machine Translation.” In Proc. Of NAACL-HLT, 3874–84. https://aclanthology.org/N19-1388.
  38. Arivazhagan, Naveen, Colin Cherry, Wolfgang Macherey, Chung-Cheng Chiu, Semih Yavuz, Ruoming Pang, Wei Li, and Colin Raffel. 2019. “Monotonic Infinite Lookback Attention for Simultaneous Machine Translation.” In Proc. Of ACL, 1313–23. https://aclanthology.org/P19-1126.
  39. Arivazhagan, Naveen, Colin Cherry, Wolfgang Macherey, and George Foster. 2020. “Re-Translation Versus Streaming for Simultaneous Translation.” In Proceedings of the 17th International Conference on Spoken Language Translation, 220–27. https://aclanthology.org/2020.iwslt-1.27.
  40. Artetxe, Mikel, Sebastian Ruder, and Dani Yogatama. 2020. “On the Cross-Lingual Transferability of Monolingual Representations.” In Proc. Of ACL, 4623–37. https://aclanthology.org/2020.acl-main.421.
  41. Baevski, Alexei, Steffen Schneider, and Michael Auli. 2020. “Vq-Wav2vec: Self-Supervised Learning of Discrete Speech Representations.” In Proc. Of ICLR. https://openreview.net/forum?id=rylwJxrYDS.
  42. Bahar, Parnia, Albert Zeyer, Ralf Schlüter, and Hermann Ney. 2019. “On Using SpecAugment for End-to-End Speech Translation.” In Proceedings of the 16th International Conference on Spoken Language Translation. https://aclanthology.org/2019.iwslt-1.22.
  43. Bansal, Sameer, Herman Kamper, Karen Livescu, Adam Lopez, and Sharon Goldwater. 2019. “Pre-Training on High-Resource Speech Recognition Improves Low-Resource Speech-to-Text Translation.” In Proc. Of NAACL-HLT, 58–68. https://aclanthology.org/N19-1006.
  44. Bao, Yu, Shujian Huang, Tong Xiao, Dongqi Wang, Xinyu Dai, and Jiajun Chen. 2021. “Non-Autoregressive Translation by Learning Target Categorical Codes.” In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 5749–59. https://aclanthology.org/2021.naacl-main.458.
  45. Bapna, Ankur, Colin Cherry, Yu Zhang, Ye Jia, Melvin Johnson, Yong Cheng, Simran Khanuja, Jason Riesa, and Alexis Conneau. 2022. “mSLAM: Massively Multilingual Joint Pre-Training for Speech and Text.” ArXiv Preprint. https://arxiv.org/abs/2202.01374.
  46. Bapna, Ankur, Yu-an Chung, Nan Wu, Anmol Gulati, Ye Jia, Jonathan H. Clark, Melvin Johnson, Jason Riesa, Alexis Conneau, and Yu Zhang. 2021. “SLAM: A Unified Encoder for Speech and Language Modeling via Speech-Text Joint Pre-Training.” ArXiv Preprint. https://arxiv.org/abs/2110.10329.
  47. Bentivogli, Luisa, Mauro Cettolo, Marco Gaido, Alina Karakanta, Alberto Martinelli, Matteo Negri, and Marco Turchi. 2021. “Cascade Versus Direct Speech Translation: Do the Differences Still Make a Difference?” In Proc. Of ACL, 2873–87. https://aclanthology.org/2021.acl-long.224.
  48. Blevins, Terra, and Luke Zettlemoyer. 2022. “Language Contamination Helps Explains the Cross-Lingual Capabilities of English Pretrained Models.” In Proc. Of EMNLP, 3563–74. https://aclanthology.org/2022.emnlp-main.233.
  49. Bornea, Mihaela A., Lin Pan, Sara Rosenthal, Radu Florian, and Avirup Sil. 2021. “Multilingual Transfer Learning for QA Using Translation as Data Augmentation.” In Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, IAAI 2021, the Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, Virtual Event, February 2-9, 2021, 12583–91. https://ojs.aaai.org/index.php/AAAI/article/view/17491.
  50. Cao, Steven, Nikita Kitaev, and Dan Klein. 2020. “Multilingual Alignment of Contextual Word Representations.” In Proc. Of ICLR. https://openreview.net/forum?id=r1xCMyBtPS.
  51. Chen, Guanhua, Shuming Ma, Yun Chen, Li Dong, Dongdong Zhang, Jia Pan, Wenping Wang, and Furu Wei. 2021. “Zero-Shot Cross-Lingual Transfer of Neural Machine Translation with Multilingual Pretrained Encoders.” In Proc. Of EMNLP, 15–26. https://aclanthology.org/2021.emnlp-main.2.
  52. Chen, Junkun, Mingbo Ma, Renjie Zheng, and Liang Huang. 2020. “MAM: Masked Acoustic Modeling for End-to-End Speech-to-Text Translation.” ArXiv Preprint abs/2010.11445. https://arxiv.org/abs/2010.11445.
  53. ———. 2021. “Direct Simultaneous Speech-to-Text Translation Assisted by Synchronized Streaming ASR.” In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, 4618–24. https://aclanthology.org/2021.findings-acl.406.
  54. Chen, Xilun, Ahmed Hassan Awadallah, Hany Hassan, Wei Wang, and Claire Cardie. 2019. “Multi-Source Cross-Lingual Model Transfer: Learning What to Share.” In Proc. Of ACL, 3098–3112. https://aclanthology.org/P19-1299.
  55. Cheng, Yong, Zhaopeng Tu, Fandong Meng, Junjie Zhai, and Yang Liu. 2018. “Towards Robust Neural Machine Translation.” In Proc. Of ACL, 1756–66. https://aclanthology.org/P18-1163.
  56. Chi, Zewen, Li Dong, Furu Wei, Nan Yang, Saksham Singhal, Wenhui Wang, Xia Song, Xian-Ling Mao, Heyan Huang, and Ming Zhou. 2021. InfoXLM: An Information-Theoretic Framework for Cross-Lingual Language Model Pre-Training.” In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 3576–88. https://aclanthology.org/2021.naacl-main.280.
  57. Chuang, Yung-Sung, Chi-Liang Liu, Hung-yi Lee, and Lin-Shan Lee. 2020. “SpeechBERT: An Audio-and-Text Jointly Learned Language Model for End-to-End Spoken Question Answering.” In Proc. Of INTERSPEECH, edited by Helen Meng, Bo Xu, and Thomas Fang Zheng, 4168–72. https://doi.org/10.21437/Interspeech.2020-1570.
  58. Conneau, Alexis, Ruty Rinott, Guillaume Lample, Adina Williams, Samuel Bowman, Holger Schwenk, and Veselin Stoyanov. 2018. XNLI: Evaluating Cross-Lingual Sentence Representations.” In Proc. Of EMNLP, 2475–85. https://aclanthology.org/D18-1269.
  59. Dong, Qianqian, Rong Ye, Mingxuan Wang, Hao Zhou, Shuang Xu, Bo Xu, and Lei Li. 2021. “Listen, Understand and Translate: Triple Supervision Decouples End-to-End Speech-to-Text Translation.” In Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, IAAI 2021, the Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, Virtual Event, February 2-9, 2021, 12749–59. https://ojs.aaai.org/index.php/AAAI/article/view/17509.
  60. Du, Yichao, Weizhi Wang, Zhirui Zhang, Boxing Chen, Tong Xu, Jun Xie, and Enhong Chen. 2022. “Non-Parametric Domain Adaptation for End-to-End Speech Translation.” In Proc. Of EMNLP, 306–20. https://aclanthology.org/2022.emnlp-main.21.
  61. Emezue, Chris Chinenye, and Bonaventure F. P. Dossou. 2021. MMTAfrica: Multilingual Machine Translation for African Languages.” In Proceedings of the Sixth Conference on Machine Translation, 398–411. https://aclanthology.org/2021.wmt-1.48.
  62. Fang, Qingkai, Rong Ye, Lei Li, Yang Feng, and Mingxuan Wang. 2022. STEMM: Self-Learning with Speech-Text Manifold Mixup for Speech Translation.” In Proc. Of ACL, 7050–62. https://aclanthology.org/2022.acl-long.486.
  63. Fang, Yuwei, Shuohang Wang, Zhe Gan, Siqi Sun, and Jingjing Liu. 2021. FILTER: An Enhanced Fusion Method for Cross-Lingual Language Understanding.” In Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, IAAI 2021, the Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, Virtual Event, February 2-9, 2021, 12776–84. https://ojs.aaai.org/index.php/AAAI/article/view/17512.
  64. Firat, Orhan, Baskaran Sankaran, Yaser Al-onaizan, Fatos T. Yarman Vural, and Kyunghyun Cho. 2016. “Zero-Resource Translation with Multi-Lingual Neural Machine Translation.” In Proc. Of EMNLP, 268–77. https://aclanthology.org/D16-1026.
  65. Gaido, Marco, Mattia A. Di Gangi, Matteo Negri, and Marco Turchi. 2020. “End-to-End Speech-Translation with Knowledge Distillation: FBK@IWSLT2020.” In Proceedings of the 17th International Conference on Spoken Language Translation, 80–88. https://aclanthology.org/2020.iwslt-1.8.
  66. Gangi, Mattia Antonino Di, Matteo Negri, and Marco Turchi. 2019. “Adapting Transformer to End-to-End Spoken Language Translation.” In Proc. Of INTERSPEECH, edited by Gernot Kubin and Zdravko Kacic, 1133–37. https://doi.org/10.21437/Interspeech.2019-3045.
  67. Gangi, Mattia Antonino Di, Viet-Nhat Nguyen, Matteo Negri, and Marco Turchi. 2020. “Instance-Based Model Adaptation for Direct Speech Translation.” In 2020 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2020, Barcelona, Spain, May 4-8, 2020, 7914–18. https://doi.org/10.1109/ICASSP40776.2020.9053901.
  68. Ghazvininejad, Marjan, Vladimir Karpukhin, Luke Zettlemoyer, and Omer Levy. 2020. “Aligned Cross Entropy for Non-Autoregressive Machine Translation.” In Proc. Of ICML, 119:3515–23. Proceedings of Machine Learning Research. http://proceedings.mlr.press/v119/ghazvininejad20a.html.
  69. Gordon, Mitchell A, Kevin Duh, and Jared Kaplan. 2021. “Data and Parameter Scaling Laws for Neural Machine Translation.” In Proc. Of EMNLP, 5915–22. https://aclanthology.org/2021.emnlp-main.478.
  70. Gu, Jiatao, Hany Hassan, Jacob Devlin, and Victor O. K. Li. 2018. “Universal Neural Machine Translation for Extremely Low Resource Languages.” In Proc. Of NAACL-HLT, 344–54. https://aclanthology.org/N18-1032.
  71. Guo, Junliang, Xu Tan, Linli Xu, Tao Qin, Enhong Chen, and Tie-Yan Liu. 2020. “Fine-Tuning by Curriculum Learning for Non-Autoregressive Neural Machine Translation.” In The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, the Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, the Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, 7839–46. https://aaai.org/ojs/index.php/AAAI/article/view/6289.
  72. Guo, Junliang, Linli Xu, and Enhong Chen. 2020. “Jointly Masked Sequence-to-Sequence Model for Non-Autoregressive Neural Machine Translation.” In Proc. Of ACL, 376–85. https://aclanthology.org/2020.acl-main.36.
  73. Ha, Thanh-Le, Jan Niehues, and Alex Waibel. 2016. “Toward Multilingual Neural Machine Translation with Universal Encoder and Decoder.” In Proceedings of the 13th International Conference on Spoken Language Translation. https://aclanthology.org/2016.iwslt-1.6.
  74. Helcl, Jindřich, Barry Haddow, and Alexandra Birch. 2022. “Non-Autoregressive Machine Translation: Its Not as Fast as It Seems.” In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 1780–90. https://aclanthology.org/2022.naacl-main.129.
  75. Hovy, Dirk, Federico Bianchi, and Tommaso Fornaciari. 2020. You Sound Just Like Your Father Commercial Machine Translation Systems Include Stylistic Biases.” In Proc. Of ACL, 1686–90. https://aclanthology.org/2020.acl-main.154.
  76. Hu, Junjie, Melvin Johnson, Orhan Firat, Aditya Siddhant, and Graham Neubig. 2021. “Explicit Alignment Objectives for Multilingual Bidirectional Encoders.” In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 3633–43. https://aclanthology.org/2021.naacl-main.284.
  77. Hu, Junjie, Mengzhou Xia, Graham Neubig, and Jaime Carbonell. 2019. “Domain Adaptation of Neural Machine Translation by Lexicon Induction.” In Proc. Of ACL, 2989–3001. https://aclanthology.org/P19-1286.
  78. Huang, Haoyang, Yaobo Liang, Nan Duan, Ming Gong, Linjun Shou, Daxin Jiang, and Ming Zhou. 2019. Unicoder: A Universal Language Encoder by Pre-Training with Multiple Cross-Lingual Tasks.” In Proc. Of EMNLP, 2485–94. https://aclanthology.org/D19-1252.
  79. Inaguma, Hirofumi, Shun Kiyono, Kevin Duh, Shigeki Karita, Nelson Yalta, Tomoki Hayashi, and Shinji Watanabe. 2020. ESPnet-ST: All-in-One Speech Translation Toolkit.” In Proc. Of ACL, 302–11. https://aclanthology.org/2020.acl-demos.34.
  80. Indurthi, Sathish Reddy, Houjeung Han, Nikhil Kumar Lakumarapu, Beomseok Lee, Insoo Chung, Sangha Kim, and Chanwoo Kim. 2020. “End-End Speech-to-Text Translation with Modality Agnostic Meta-Learning.” In 2020 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2020, Barcelona, Spain, May 4-8, 2020, 7904–8. https://doi.org/10.1109/ICASSP40776.2020.9054759.
  81. Indurthi, Sathish, Houjeung Han, Nikhil Kumar Lakumarapu, Beomseok Lee, Insoo Chung, Sangha Kim, and Chanwoo Kim. 2020. “Data Efficient Direct Speech-to-Text Translation with Modality Agnostic Meta-Learning.” In Proc. Of ICASSP. IEEE.
  82. Ji, Baijun, Zhirui Zhang, Xiangyu Duan, Min Zhang, Boxing Chen, and Weihua Luo. 2020. “Cross-Lingual Pre-Training Based Transfer for Zero-Shot Neural Machine Translation.” In The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, the Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, the Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, 115–22. https://aaai.org/ojs/index.php/AAAI/article/view/5341.
  83. Jiang, Hui, Ziyao Lu, Fandong Meng, Chulun Zhou, Jie Zhou, Degen Huang, and Jinsong Su. 2022. “Towards Robust k-Nearest-Neighbor Machine Translation.” In Proc. Of EMNLP, 5468–77. https://aclanthology.org/2022.emnlp-main.367.
  84. Kale, Mihir, Aditya Siddhant, Rami Al-Rfou, Linting Xue, Noah Constant, and Melvin Johnson. 2021. “NmT5 - Is Parallel Data Still Relevant for Pre-Training Massively Multilingual Language Models?” In Proc. Of ACL, 683–91. https://aclanthology.org/2021.acl-short.87.
  85. Kasai, Jungo, James Cross, Marjan Ghazvininejad, and Jiatao Gu. 2020. “Non-Autoregressive Machine Translation with Disentangled Context Transformer.” In Proc. Of ICML, 119:5144–55. Proceedings of Machine Learning Research. http://proceedings.mlr.press/v119/kasai20a.html.
  86. Keung, Phillip, Yichao Lu, and Vikas Bhardwaj. 2019. “Adversarial Learning with Contextual Embeddings for Zero-Resource Cross-Lingual Classification and NER.” In Proc. Of EMNLP, 1355–60. https://aclanthology.org/D19-1138.
  87. Kim, Yunsu, Yingbo Gao, and Hermann Ney. 2019. “Effective Cross-Lingual Transfer of Neural Machine Translation Models Without Shared Vocabularies.” In Proc. Of ACL, 1246–57. https://aclanthology.org/P19-1120.
  88. Lam, Tsz Kin, Shigehiko Schamoni, and Stefan Riezler. 2021. “Cascaded Models with Cyclic Feedback for Direct Speech Translation.” In ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 7508–12.
  89. ———. 2022. “Sample, Translate, Recombine: Leveraging Audio Alignments for Data Augmentation in End-to-End Speech Translation.” In Proc. Of ACL, 245–54. https://aclanthology.org/2022.acl-short.27.
  90. Lauscher, Anne, Vinit Ravishankar, Ivan Vulić, and Goran Glavaš. 2020. “From Zero to Hero: On the Limitations of Zero-Shot Language Transfer with Multilingual Transformers.” In Proc. Of EMNLP, 4483–99. https://aclanthology.org/2020.emnlp-main.363.
  91. Le, Hang, Juan Pino, Changhan Wang, Jiatao Gu, Didier Schwab, and Laurent Besacier. 2020. “Dual-Decoder Transformer for Joint Automatic Speech Recognition and Multilingual Speech Translation.” In Proceedings of the 28th International Conference on Computational Linguistics, 3520–33. https://aclanthology.org/2020.coling-main.314.
  92. ———. 2021. “Lightweight Adapter Tuning for Multilingual Speech Translation.” In Proc. Of ACL, 817–24. https://aclanthology.org/2021.acl-short.103.
  93. Lewis, Patrick, Barlas Oguz, Ruty Rinott, Sebastian Riedel, and Holger Schwenk. 2020. MLQA: Evaluating Cross-Lingual Extractive Question Answering.” In Proc. Of ACL, 7315–30. https://aclanthology.org/2020.acl-main.653.
  94. Li, Xian, and Hongyu Gong. 2021. “Robust Optimization for Multilingual Translation with Imbalanced Data.” In Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, December 6-14, 2021, Virtual, edited by Marc’Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan, 25086–99. https://proceedings.neurips.cc/paper/2021/hash/d324a0cc02881779dcda44a675fdcaaa-Abstract.html.
  95. Li, Xian, Changhan Wang, Yun Tang, Chau Tran, Yuqing Tang, Juan Pino, Alexei Baevski, Alexis Conneau, and Michael Auli. 2021. “Multilingual Speech Translation with Efficient Finetuning of Pretrained Models.” In Proc. Of ACL.
  96. Liang, Sheng, Philipp Dufter, and Hinrich Schütze. 2020. “Monolingual and Multilingual Reduction of Gender Bias in Contextualized Representations.” In Proceedings of the 28th International Conference on Computational Linguistics, 5082–93. https://aclanthology.org/2020.coling-main.446.
  97. Libovický, Jindřich, Rudolf Rosa, and Alexander Fraser. 2020. “On the Language Neutrality of Pre-Trained Multilingual Representations.” In Findings of the Association for Computational Linguistics: EMNLP 2020, 1663–74. https://aclanthology.org/2020.findings-emnlp.150.
  98. Lin, Xi Victoria, Todor Mihaylov, Mikel Artetxe, Tianlu Wang, Shuohui Chen, Daniel Simig, Myle Ott, et al. 2022. “Few-Shot Learning with Multilingual Generative Language Models.” In Proc. Of EMNLP, 9019–52. https://aclanthology.org/2022.emnlp-main.616.
  99. Liu, Yuchen, Hao Xiong, Jiajun Zhang, Zhongjun He, Hua Wu, Haifeng Wang, and Chengqing Zong. 2019. “End-to-End Speech Translation with Knowledge Distillation.” In Proc. Of INTERSPEECH, edited by Gernot Kubin and Zdravko Kacic, 1128–32. https://doi.org/10.21437/Interspeech.2019-2582.
  100. Liu, Yuchen, Jiajun Zhang, Hao Xiong, Long Zhou, Zhongjun He, Hua Wu, Haifeng Wang, and Chengqing Zong. 2020. “Synchronous Speech Recognition and Speech-to-Text Translation with Interactive Decoding.” In The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, the Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, the Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, 8417–24. https://aaai.org/ojs/index.php/AAAI/article/view/6360.
  101. Luo, Fuli, Wei Wang, Jiahao Liu, Yijia Liu, Bin Bi, Songfang Huang, Fei Huang, and Luo Si. 2021. VECO: Variable and Flexible Cross-Lingual Pre-Training for Language Understanding and Generation.” In Proc. Of ACL, 3980–94. https://aclanthology.org/2021.acl-long.308.
  102. Lyu, Sungwon, Bokyung Son, Kichang Yang, and Jaekyoung Bae. 2020. “Revisiting Modularized Multilingual NMT to Meet Industrial Demands.” In Proc. Of EMNLP, 5905–18. https://aclanthology.org/2020.emnlp-main.476.
  103. Ma, Mingbo, Liang Huang, Hao Xiong, Kaibo Liu, Chuanqiang Zhang, Zhongjun He, Hairong Liu, Xing Li, and Haifeng Wang. 2019. “Stacl: Simultaneous Translation with Integrated Anticipation and Controllable Latency,” 3025–36.
  104. Ma, Xutai, Mohammad Javad Dousti, Changhan Wang, Jiatao Gu, and Juan Pino. 2020. SIMULEVAL: An Evaluation Toolkit for Simultaneous Translation.” In Proc. Of EMNLP, 144–50. https://aclanthology.org/2020.emnlp-demos.19.
  105. Ma, Xutai, Juan Pino, and Philipp Koehn. 2020. SimulMT to SimulST: Adapting Simultaneous Text Translation to End-to-End Simultaneous Speech Translation.” In Proceedings of the 1st Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 10th International Joint Conference on Natural Language Processing, 582–87. https://aclanthology.org/2020.aacl-main.58.
  106. Martins, Pedro Henrique, Zita Marinho, and André F. T. Martins. 2022. “Chunk-Based Nearest Neighbor Machine Translation.” In Proc. Of EMNLP, 4228–45. https://aclanthology.org/2022.emnlp-main.284.
  107. McCarthy, Arya D., Liezl Puzon, and Juan Pino. 2020. “SkinAugment: Auto-Encoding Speaker Conversions for Automatic Speech Translation.” In 2020 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2020, Barcelona, Spain, May 4-8, 2020, 7924–28. https://doi.org/10.1109/ICASSP40776.2020.9053406.
  108. Meng, Yuxian, Xiaoya Li, Xiayu Zheng, Fei Wu, Xiaofei Sun, Tianwei Zhang, and Jiwei Li. 2022. “Fast Nearest Neighbor Machine Translation.” In Findings of the Association for Computational Linguistics: ACL 2022, 555–65. https://aclanthology.org/2022.findings-acl.47.
  109. Mi, Chenggang, Lei Xie, and Yanning Zhang. 2022. “Improving Data Augmentation for Low Resource Speech-to-Text Translation with Diverse Paraphrasing.” Neural Netw. 148 (C): 194–205. https://doi.org/10.1016/j.neunet.2022.01.016.
  110. Moslem, Yasmin, Rejwanul Haque, and Andy Way. 2023. “Adaptive Machine Translation with Large Language Models.” ArXiv Preprint abs/2301.13294. https://arxiv.org/abs/2301.13294.
  111. Müller, Mathias, Annette Rios, and Rico Sennrich. 2020. “Domain Robustness in Neural Machine Translation.” In Proceedings of the 14th Conference of the Association for Machine Translation in the Americas (Volume 1: Research Track), 151–64. https://aclanthology.org/2020.amta-research.14.
  112. Neubig, Graham, and Junjie Hu. 2018. “Rapid Adaptation of Neural Machine Translation to New Languages.” In Proc. Of EMNLP, 875–80. https://aclanthology.org/D18-1103.
  113. Nguyen, Ha, Fethi Bougares, Natalia A. Tomashenko, Yannick Estève, and Laurent Besacier. 2020. “Investigating Self-Supervised Pre-Training for End-to-End Speech Translation.” In Proc. Of INTERSPEECH, edited by Helen Meng, Bo Xu, and Thomas Fang Zheng, 1466–70. https://doi.org/10.21437/Interspeech.2020-1835.
  114. Ni, Minheng, Haoyang Huang, Lin Su, Edward Cui, Taroon Bharti, Lijuan Wang, Dongdong Zhang, and Nan Duan. 2021. M3P: Learning Universal Representations via Multitask Multilingual Multimodal Pre-Training.” In IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, Virtual, June 19-25, 2021, 3977–86. https://openaccess.thecvf.com/content/CVPR2021/html/Ni\_M3P\_Learning\_Universal\_Representations\_via\_Multitask\_Multilingual\_Multimodal\_Pre-Training\_CVPR\_2021\_paper.html.
  115. Passban, Peyman, Tanya Roosta, Rahul Gupta, Ankit Chadha, and Clement Chung. 2022. “Training Mixed-Domain Translation Models via Federated Learning.” In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2576–86. https://aclanthology.org/2022.naacl-main.186.
  116. Pino, Juan Miguel, Qiantong Xu, Xutai Ma, Mohammad Javad Dousti, and Yun Tang. 2020. “Self-Training for End-to-End Speech Translation.” In Proc. Of INTERSPEECH, edited by Helen Meng, Bo Xu, and Thomas Fang Zheng, 1476–80. https://doi.org/10.21437/Interspeech.2020-2938.
  117. Ponti, Edoardo Maria, Goran Glavaš, Olga Majewska, Qianchu Liu, Ivan Vulić, and Anna Korhonen. 2020. XCOPA: A Multilingual Dataset for Causal Commonsense Reasoning.” In Proc. Of EMNLP, 2362–76. https://aclanthology.org/2020.emnlp-main.185.
  118. Qian, Lihua, Yi Zhou, Zaixiang Zheng, Yaoming Zhu, Zehui Lin, Jiangtao Feng, Shanbo Cheng, Lei Li, Mingxuan Wang, and Hao Zhou. 2021. “The Volctrans GLAT System: Non-Autoregressive Translation Meets WMT21.” In Proceedings of the Sixth Conference on Machine Translation, 187–96. https://aclanthology.org/2021.wmt-1.17.
  119. Qin, Libo, Minheng Ni, Yue Zhang, and Wanxiang Che. 2020. “CoSDA-ML: Multi-Lingual Code-Switching Data Augmentation for Zero-Shot Cross-Lingual NLP.” In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI 2020, edited by Christian Bessiere, 3853–60. https://doi.org/10.24963/ijcai.2020/533.
  120. Ran, Qiu, Yankai Lin, Peng Li, and Jie Zhou. 2021. “Guiding Non-Autoregressive Neural Machine Translation Decoding with Reordering Information.” In Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, IAAI 2021, the Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, Virtual Event, February 2-9, 2021, 13727–35. https://ojs.aaai.org/index.php/AAAI/article/view/17618.
  121. Reid, Machel, and Mikel Artetxe. 2022. PARADISE: Exploiting Parallel Data for Multilingual Sequence-to-Sequence Pretraining.” In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 800–810. https://aclanthology.org/2022.naacl-main.58.
  122. Ren, Yi, Jinglin Liu, Xu Tan, Chen Zhang, Tao Qin, Zhou Zhao, and Tie-Yan Liu. 2020. SimulSpeech: End-to-End Simultaneous Speech to Text Translation.” In Proc. Of ACL, 3787–96. https://aclanthology.org/2020.acl-main.350.
  123. Sachan, Devendra, and Graham Neubig. 2018. “Parameter Sharing Methods for Multilingual Self-Attentional Translation Models.” In Proceedings of the Third Conference on Machine Translation: Research Papers, 261–71. https://aclanthology.org/W18-6327.
  124. Saharia, Chitwan, William Chan, Saurabh Saxena, and Mohammad Norouzi. 2020. “Non-Autoregressive Machine Translation with Latent Alignments.” In Proc. Of EMNLP, 1098–108. https://aclanthology.org/2020.emnlp-main.83.
  125. Salesky, Elizabeth, Matthias Sperber, and Alexander Waibel. 2019. “Fluent Translations from Disfluent Speech in End-to-End Speech Translation.” In Proc. Of NAACL-HLT, 2786–92. https://aclanthology.org/N19-1285.
  126. Schneider, Felix, and Alexander Waibel. 2020. “Towards Stream Translation: Adaptive Computation Time for Simultaneous Machine Translation.” In Proceedings of the 17th International Conference on Spoken Language Translation, 228–36. https://aclanthology.org/2020.iwslt-1.28.
  127. Sennrich, Rico, Barry Haddow, and Alexandra Birch. 2016a. “Improving Neural Machine Translation Models with Monolingual Data.” In Proc. Of ACL, 86–96. https://aclanthology.org/P16-1009.
  128. Shao, Chenze, Jinchao Zhang, Yang Feng, Fandong Meng, and Jie Zhou. 2020. “Minimizing the Bag-of-Ngrams Difference for Non-Autoregressive Neural Machine Translation.” In The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, the Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, the Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, 198–205. https://aaai.org/ojs/index.php/AAAI/article/view/5351.
  129. Shu, Raphael, Jason Lee, Hideki Nakayama, and Kyunghyun Cho. 2020. “Latent-Variable Non-Autoregressive Neural Machine Translation with Deterministic Inference Using a Delta Posterior.” In The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, the Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, the Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, 8846–53. https://aaai.org/ojs/index.php/AAAI/article/view/6413.
  130. Siddhant, Aditya, Ankur Bapna, Yuan Cao, Orhan Firat, Mia Chen, Sneha Kudugunta, Naveen Arivazhagan, and Yonghui Wu. 2020. “Leveraging Monolingual Data with Self-Supervision for Multilingual Neural Machine Translation.” In Proc. Of ACL, 2827–35. https://aclanthology.org/2020.acl-main.252.
  131. Siddhant, Aditya, Melvin Johnson, Henry Tsai, Naveen Ari, Jason Riesa, Ankur Bapna, Orhan Firat, and Karthik Raman. 2020. “Evaluating the Cross-Lingual Effectiveness of Massively Multilingual Neural Machine Translation.” In The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, the Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, the Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, 8854–61. https://aaai.org/ojs/index.php/AAAI/article/view/6414.
  132. Sperber, Matthias, Graham Neubig, Jan Niehues, and Alex Waibel. 2019. “Attention-Passing Models for Robust and Data-Efficient End-to-End Speech Translation.” Transactions of the Association for Computational Linguistics 7: 313–25. https://aclanthology.org/Q19-1020.
  133. Stoian, Mihaela C., Sameer Bansal, and Sharon Goldwater. 2020. “Analyzing ASR Pretraining for Low-Resource Speech-to-Text Translation.” In 2020 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2020, Barcelona, Spain, May 4-8, 2020, 7909–13. https://doi.org/10.1109/ICASSP40776.2020.9053847.
  134. Sung, Tzu-Wei, Jun-You Liu, Hung-yi Lee, and Lin-Shan Lee. 2019. “Towards End-to-End Speech-to-Text Translation with Two-Pass Decoding.” In IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2019, Brighton, United Kingdom, May 12-17, 2019, 7175–79. https://doi.org/10.1109/ICASSP.2019.8682801.
  135. Tang, Yun, Juan Pino, Xian Li, Changhan Wang, and Dmitriy Genzel. 2021. “Improving Speech Translation by Understanding and Learning from the Auxiliary Text Translation Task.” In Proc. Of ACL, 4252–61. https://aclanthology.org/2021.acl-long.328.
  136. Tang, Yun, Juan Pino, Changhan Wang, Xutai Ma, and Dmitriy Genzel. 2021. “A General Multi-Task Learning Framework to Leverage Text Data for Speech to Text Tasks.” In Proc. of ICASSP, 6209–13. IEEE.
  137. Tang, Yuqing, Chau Tran, Xian Li, Peng-Jen Chen, Naman Goyal, Vishrav Chaudhary, Jiatao Gu, and Angela Fan. 2021. “Multilingual Translation from Denoising Pre-Training.” In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, 3450–66. https://aclanthology.org/2021.findings-acl.304.
  138. Tran, Chau, Yuqing Tang, Xian Li, and Jiatao Gu. 2020. “Cross-Lingual Retrieval for Iterative Self-Supervised Training.” In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, Virtual, edited by Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin. https://proceedings.neurips.cc/paper/2020/hash/1763ea5a7e72dd7ee64073c2dda7a7a8-Abstract.html.
  139. Vázquez, Raúl, Alessandro Raganato, Jörg Tiedemann, and Mathias Creutz. 2019. “Multilingual NMT with a Language-Independent Attention Bridge.” In Proceedings of the 4th Workshop on Representation Learning for NLP (RepL4NLP-2019), 33–39. https://aclanthology.org/W19-4305.
  140. Vydana, Hari Krishna, Martin Karafiát, Katerina Zmolikova, Lukáš Burget, and Honza Černockỳ. 2021. “Jointly Trained Transformers Models for Spoken Language Translation.” In Proc. of ICASSP, 7513–17. IEEE.
  141. Wang, Chengyi, Yu Wu, Shujie Liu, Zhenglu Yang, and Ming Zhou. 2020. “Bridging the Gap Between Pre-Training and Fine-Tuning for End-to-End Speech Translation.” In The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, the Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, the Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, 9161–68. https://aaai.org/ojs/index.php/AAAI/article/view/6452.
  142. Wang, Chengyi, Yu Wu, Shujie Liu, Ming Zhou, and Zhenglu Yang. 2020. “Curriculum Pre-Training for End-to-End Speech Translation.” In Proc. Of ACL, 3728–38. https://aclanthology.org/2020.acl-main.344.
  143. Wang, Chen, Yuchen Liu, Boxing Chen, Jiajun Zhang, Wei Luo, Zhongqiang Huang, and Chengqing Zong. 2022. “Discrete Cross-Modal Alignment Enables Zero-Shot Speech Translation.” In Proc. Of EMNLP, 5291–5302. https://aclanthology.org/2022.emnlp-main.354.
  144. Wang, Dexin, Kai Fan, Boxing Chen, and Deyi Xiong. 2022. “Efficient Cluster-Based k-Nearest-Neighbor Machine Translation.” In Proc. Of ACL, 2175–87. https://aclanthology.org/2022.acl-long.154.
  145. Wang, Mingxuan, Jun Xie, Zhixing Tan, Jinsong Su, Deyi Xiong, and Lei Li. 2019. “Towards Linear Time Neural Machine Translation with Capsule Networks.” In Proc. Of EMNLP, 803–12. https://aclanthology.org/D19-1074.
  146. Wang, Tao, Chengqi Zhao, Mingxuan Wang, Lei Li, Hang Li, and Deyi Xiong. 2021. “Secoco: Self-Correcting Encoding for Neural Machine Translation.” In Findings of the Association for Computational Linguistics: EMNLP 2021, 4639–44. https://aclanthology.org/2021.findings-emnlp.396.
  147. Wang, Xinyi, Hieu Pham, Philip Arthur, and Graham Neubig. 2019. “Multilingual Neural Machine Translation with Soft Decoupled Encoding.” In Proc. Of ICLR. https://openreview.net/forum?id=Skeke3C5Fm.
  148. Wang, Yining, Jiajun Zhang, Feifei Zhai, Jingfang Xu, and Chengqing Zong. 2018. “Three Strategies to Improve One-to-Many Multilingual Translation.” In Proc. Of EMNLP, 2955–60. https://aclanthology.org/D18-1326.
  149. Wang, Yiren, ChengXiang Zhai, and Hany Hassan. 2020. “Multi-Task Learning for Multilingual Neural Machine Translation.” In Proc. Of EMNLP, 1022–34. https://aclanthology.org/2020.emnlp-main.75.
  150. Weng, Rongxiang, Heng Yu, Shujian Huang, Shanbo Cheng, and Weihua Luo. 2020. “Acquiring Knowledge from Pre-Trained Model to Neural Machine Translation.” In The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, the Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, the Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, 9266–73. https://aaai.org/ojs/index.php/AAAI/article/view/6465.
  151. Wu, Anne, Changhan Wang, Juan Miguel Pino, and Jiatao Gu. 2020. “Self-Supervised Representations Improve End-to-End Speech Translation.” In Proc. Of INTERSPEECH, edited by Helen Meng, Bo Xu, and Thomas Fang Zheng, 1491–95. https://doi.org/10.21437/Interspeech.2020-3094.
  152. Wu, Liwei, Shanbo Cheng, Mingxuan Wang, and Lei Li. 2021. “Language Tags Matter for Zero-Shot Neural Machine Translation.” In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, 3001–7. https://aclanthology.org/2021.findings-acl.264.
  153. Wu, Shijie, and Mark Dredze. 2020a. “Are All Languages Created Equal in Multilingual BERT?” In Proceedings of the 5th Workshop on Representation Learning for NLP, 120–30. https://aclanthology.org/2020.repl4nlp-1.16.
  154. Shijie Wu, Mark Dredze. 2020b. “Do Explicit Alignments Robustly Improve Multilingual Encoders?” In Proc. Of EMNLP, 4471–82. https://aclanthology.org/2020.emnlp-main.362.
  155. Wu, Yonghui, Mike Schuster, Zhifeng Chen, Quoc V Le, Mohammad Norouzi, Wolfgang Macherey, Maxim Krikun, et al. 2016. “Google’s Neural Machine Translation System: Bridging the Gap Between Human and Machine Translation.” ArXiv Preprint abs/1609.08144. https://arxiv.org/abs/1609.08144.
  156. Xu, Chen, Bojie Hu, Yanyang Li, Yuhao Zhang, Shen Huang, Qi Ju, Tong Xiao, and Jingbo Zhu. 2021. “Stacked Acoustic-and-Textual Encoding: Integrating the Pre-Trained Models into Speech Translation Encoders.” In Proc. Of ACL, 2619–30. https://aclanthology.org/2021.acl-long.204.
  157. Yang, Jian, Shuming Ma, Dongdong Zhang, Shuangzhi Wu, Zhoujun Li, and Ming Zhou. 2020. “Alternating Language Modeling for Cross-Lingual Pre-Training.” In The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, the Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, the Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, 9386–93. https://aaai.org/ojs/index.php/AAAI/article/view/6480.
  158. Yang, Zhixian, Renliang Sun, and Xiaojun Wan. 2022. “Nearest Neighbor Knowledge Distillation for Neural Machine Translation.” In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 5546–56. https://aclanthology.org/2022.naacl-main.406.
  159. Zhang, Biao, Barry Haddow, and Rico Sennrich. 2022. “Revisiting End-to-End Speech-to-Text Translation from Scratch.” In International Conference on Machine Learning, ICML 2022, 17-23 July 2022, Baltimore, Maryland, USA, edited by Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvári, Gang Niu, and Sivan Sabato, 162:26193–205. Proceedings of Machine Learning Research. https://proceedings.mlr.press/v162/zhang22i.html.
  160. Zhang, Biao, Ivan Titov, Barry Haddow, and Rico Sennrich. 2020. “Adaptive Feature Selection for End-to-End Speech Translation.” In Findings of the Association for Computational Linguistics: EMNLP 2020, 2533–44. https://aclanthology.org/2020.findings-emnlp.230.
  161. Zhang, Biao, Philip Williams, Ivan Titov, and Rico Sennrich. 2020. “Improving Massively Multilingual Neural Machine Translation and Zero-Shot Translation.” In Proc. Of ACL, 1628–39. https://aclanthology.org/2020.acl-main.148.
  162. Zhang, Ruiqing, and Chuanqiang Zhang. 2020. “Dynamic Sentence Boundary Detection for Simultaneous Translation.” In Proceedings of the First Workshop on Automatic Simultaneous Translation, 1–9. https://aclanthology.org/2020.autosimtrans-1.1.
  163. Zhang, Ruiqing, Chuanqiang Zhang, Zhongjun He, Hua Wu, and Haifeng Wang. 2020. “Learning Adaptive Segmentation Policy for Simultaneous Translation.” In Proc. Of EMNLP, 2280–89. https://aclanthology.org/2020.emnlp-main.178.
  164. Zhao, Chengqi, Zhicheng Liu, Jian Tong, Tao Wang, Mingxuan Wang, Rong Ye, Qianqian Dong, Jun Cao, and Lei Li. 2021. “The Volctrans Neural Speech Translation System for IWSLT 2021.” In Proceedings of the 18th International Conference on Spoken Language Translation (IWSLT 2021), 64–74. https://aclanthology.org/2021.iwslt-1.6.
  165. Zhao, Wei, Steffen Eger, Johannes Bjerva, and Isabelle Augenstein. 2021. “Inducing Language-Agnostic Multilingual Representations.” In Proceedings of *SEM 2021: The Tenth Joint Conference on Lexical and Computational Semantics, 229–40. https://aclanthology.org/2021.starsem-1.22.
  166. Zheng, Bo, Li Dong, Shaohan Huang, Wenhui Wang, Zewen Chi, Saksham Singhal, Wanxiang Che, Ting Liu, Xia Song, and Furu Wei. 2021. “Consistency Regularization for Cross-Lingual Fine-Tuning.” In Proc. Of ACL, 3403–17. https://aclanthology.org/2021.acl-long.264.
  167. Zheng, Renjie, Mingbo Ma, Baigong Zheng, Kaibo Liu, and Liang Huang. 2020. “Opportunistic Decoding with Timely Correction for Simultaneous Translation.” In Proc. Of ACL, 437–42. https://aclanthology.org/2020.acl-main.42.
  168. Zheng, Renjie, Mingbo Ma, Baigong Zheng, Kaibo Liu, Jiahong Yuan, Kenneth Church, and Liang Huang. 2020. “Fluent and Low-Latency Simultaneous Speech-to-Speech Translation with Self-Adaptive Training.” In Findings of the Association for Computational Linguistics: EMNLP 2020, 3928–37. https://aclanthology.org/2020.findings-emnlp.349.
  169. Zheng, Xin, Zhirui Zhang, Junliang Guo, Shujian Huang, Boxing Chen, Weihua Luo, and Jiajun Chen. 2021. “Adaptive Nearest Neighbor Machine Translation.” In Proc. Of ACL, 368–74. https://aclanthology.org/2021.acl-short.47.
  170. Zheng, Xin, Zhirui Zhang, Shujian Huang, Boxing Chen, Jun Xie, Weihua Luo, and Jiajun Chen. 2021. “Non-Parametric Unsupervised Domain Adaptation for Neural Machine Translation.” In Findings of the Association for Computational Linguistics: EMNLP 2021, 4234–41. https://aclanthology.org/2021.findings-emnlp.358.
  171. Zhu, Changfeng, Heng Yu, Shanbo Cheng, and Weihua Luo. 2020. “Language-Aware Interlingua for Multilingual Neural Machine Translation.” In Proc. Of ACL, 1650–55. https://aclanthology.org/2020.acl-main.150.
  172. Zhu, Jinhua, Yingce Xia, Lijun Wu, Di He, Tao Qin, Wengang Zhou, Houqiang Li, and Tie-Yan Liu. 2020. “Incorporating BERT into Neural Machine Translation.” In Proc. Of ICLR. https://openreview.net/forum?id=Hyl7ygStwB.