Abstract
In recent years, neural network-based machine translation (MT) approaches have steadily superseded the statistical MT (SMT) methods, and represents the current state-of-the-art in MT research. Neural MT (NMT) is a data-driven end-to-end learning protocol whose training routine usually requires a large amount of parallel data in order to build a reasonable-quality MT system. This is particularly problematic for those language pairs that do not have enough parallel text for training. In order to counter the data sparsity problem of the NMT training, MT researchers have proposed various strategies, e.g. augmenting training data, exploiting training data from other languages, alternative learning strategies that use only monolingual data. This paper presents a survey on recent advances of NMT research from the perspective of low-resource scenarios.
Similar content being viewed by others
Explore related subjects
Discover the latest articles and news from researchers in related subjects, suggested using machine learning.Notes
The 6th Workshop on Asian Translation: http://lotus.kuee.kyoto-u.ac.jp/WAT/WAT2019/index.html.
References
Agić Ž, Johannsen A, Plank B, Martínez Alonso H, Schluter N, Søgaard A (2016) Multilingual projection for parsing truly low-resource languages. TACL 4:301–312
Artetxe M, Labaka G, Agirre E (2018a) Unsupervised statistical machine translation. In: Proceedings of the 2018 conference on empirical methods in natural language processing, pp 3632–3642, Brussels, Belgium. Association for Computational Linguistics
Artetxe M, Labaka G, Agirre E (2019) An effective approach to unsupervised machine translation. CoRR, arxiv:abs/1902.01313
Artetxe M, Labaka G, Agirre E, Cho K (2018b) Unsupervised neural machine translation. In: Unsupervised neural machine translation
Bahdanau D, Cho K, Bengio Y (2014) Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv: 1409.0473
Bahdanau D, Cho K, Bengio Y (2015) Neural machine translation by jointly learning to align and translate. In: 3rd International conference on learning representations (ICLR 2015), San Diego, CA
Barrault L, Bojar O, Costa-jussà MR, Federmann C, Fishel M, Graham Y, Haddow B, Huck M, Koehn P, Malmasi S, Monz C, Müller M, Pal S, Post M, Zampieri M (2019) Findings of the 2019 conference on machine translation (wmt19). In: Proceedings of the fourth conference on machine translation (vol 2: shared task papers, Day 1), pp 1–61, Florence, Italy. Association for Computational Linguistics
Bogoychev N, Sennrich R (2019) Domain, translationese and noise in synthetic data for neural machine translation
Bojar O, Chatterjee R, Federmann C, Graham Y, Haddow B, Huang S, Huck M, Koehn P, Liu Q, Logacheva V, Monz C, Negri M, Post M, Rubino R, Specia L, Turchi M (2017) Findings of the 2017 conference on machine translation (wmt17). In: Proceedings of the second conference on machine translation, volume 2: shared task papers, pp 169–214, Copenhagen, Denmark. Association for Computational Linguistics
Bojar O, Chatterjee R, Federmann C, Graham Y, Haddow B, Huck M, Jimeno Yepes A, Koehn P, Logacheva V, Monz C, Negri M, Neveol A, Neves M, Popel M, Post M, Rubino R, Scarton C, Specia L, Turchi M, Verspoor K, Zampieri M (2016) Findings of the 2016 conference on machine translation. In: Proceedings of the first conference on machine translation, pp 131–198, Berlin, Germany
Bojar O, Diatka V, Rychlý P, Straňák P, Suchomel V, Tamchyna A, Zeman D (2014) HindEnCorp - Hindi-English and Hindi-only Corpus for Machine Translation. In Chair), N. C. C., Choukri, K., Declerck, T., Loftsson, H., Maegaard, B., Mariani, J., Moreno, A., Odijk, J., and Piperidis, S (eds) Proceedings of the ninth international conference on language resources and evaluation (LREC’14), Reykjavik, Iceland. European Language Resources Association (ELRA)
Buck C, Heafield K, van Ooyen B (2014) N-gram counts and language models from the common crawl. In: Proceedings of the language resources and evaluation conference, Reykjavik, Iceland
Burlot F, Yvon F (2018) Using monolingual data in neural machine translation: a systematic study. In: Proceedings of the third conference on machine translation: research papers, pp 144–155, Belgium, Brussels. Association for Computational Linguistics
Callison-Burch C, Koehn P, Monz C, Zaidan OF (2011) Findings of the 2011 workshop on statistical machine translation. In: Proceedings of the sixth workshop on statistical machine translation, pp 22–64
Caswell I, Chelba C, Grangier D (2019) Tagged back-translation. In: Proceedings of the fourth conference on machine translation (volume 1: research papers), pp 53–63, Florence, Italy. Association for Computational Linguistics
Chen P-J, Shen J, Le M, Chaudhary V, El-Kishky A, Wenzek G, Ott M, Ranzato M (2019) Facebook AI’s WAT19 Myanmar-English translation task submission. In: Proceedings of the 6th workshop on Asian translation, pp 112–122, Hong Kong, China
Chen Y, Liu Y, Cheng Y, Li VO (2017) A teacher-student framework for zero-resource neural machine translation. In: Proceedings of the 55th annual meeting of the ACL (volume 1: long papers), vol. 1, pp 1925–1935
Chen Y, Liu Y, Li VO (2018) Zero-resource neural machine translation with multi-agent communication game. In: Thirty-second AAAI conference on artificial intelligence
Cheng Y, Xu W, He Z, He W, Wu H, Sun M, Liu Y (2016) Semi-supervised learning for neural machine translation. In: Proceedings of the 54th annual meeting of the ACL (volume 1: long papers). Association for Computational Linguistics, Berlin, Germany, pp 1965–1974
Cheng Y, Yang Q, Liu Y, Sun M, Xu W (2017) Joint training for pivot-based neural machine translation. In: Proceedings of the twenty-sixth international joint conference on artificial intelligence (IJCAI-17), pp 3974–3980, Melbourne, Australia
Chinea-Ríos M, Peris Á, Casacuberta F (2017) Adapting neural machine translation with parallel synthetic data. In: Proceedings of the second conference on machine translation, pp 138–147, Copenhagen, Denmark. Association for Computational Linguistics
Cho K, van Merriënboer B, Bahdanau D, Bengio Y (2014a) On the properties of neural machine translation: Encoder–decoder approaches. In: Proceedings of SSST-8, eighth workshop on syntax, semantics and structure in statistical translation, pp 103–111, Doha, Qatar. Association for Computational Linguistics
Cho K, Van Merriënboer B, Bahdanau D, Bengio Y (2014b) On the properties of neural machine translation: encoder-decoder approaches. arXiv preprint arXiv:1409.1259
Cho K, van Merriënboer B, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014c) Learning phrase representations using RNN encoder–decoder for statistical machine translation. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP)
Christodouloupoulos C, Steedman M (2015) A massively parallel corpus: the bible in 100 languages. Language Resour Eval 49(2):375–395
Currey A, Heafield K (2018) Unsupervised source hierarchies for low-resource neural machine translation. In: Proceedings of the workshop on the relevance of linguistic structure in neural architectures for NLP, pp 6–12
Currey A, Miceli Barone AV, Heafield K (2017) Copied monolingual data improves low-resource neural machine translation. In: Proceedings of the second conference on machine translation, pp 148–156, Copenhagen, Denmark. Association for Computational Linguistics
Dabre R, Chu C, Kunchukuttan A (2020) A survey of multilingual neural machine translation. ACM Comput Surv 53(5):99:1–99:38
Dabre R, Nakagawa T, Kazawa H (2017) An empirical study of language relatedness for transfer learning in neural machine translation. In: Proceedings of the 31st Pacific Asia conference on language, information and computation, pp 282–286
Devlin J, Chang M, Lee K, Toutanova K (2018) BERT: pre-training of deep bidirectional transformers for language understanding. CoRR, arxiv:abs/1810.04805
Durrani N, Koehn P (2014) Improving machine translation via triangulation and transliteration. In: Proceedings of the 17th annual conference of the European Association for Machine Translation (EAMT), Dubrovnik, Croatia
Dyer C, Chahuneau V, Smith NA (2013) A simple, fast, and effective reparameterization of IBM model 2. In: Proceedings of the 2013 Conference of the North American chapter of the ACL: human language technologies. Association for Computational Linguistics, Atlanta, Georgia, pp 644–648
Edunov S, Ott M, Auli M, Grangier D (2018) Understanding back-translation at scale. In: Proceedings of the 2018 conference on EMNLP. Association for Computational Linguistics, Brussels, Belgium, pp 489–500
Edunov S, Ott M, Ranzato M, Auli M (2020) On the evaluation of machine translation systems trained with back-translation. In: Proceedings of the 58th annual meeting of the ACL, Online. Association for Computational Linguistics, pp 2836–2846
Elliott D, Frank S, Sima’an K, Specia L (2016) Multi30K: Multilingual English-German image descriptions. In: Proceedings of the 5th workshop on vision and language, pp 70–74, Berlin, Germany. Association for Computational Linguistics
Fadaee M, Bisazza A, Monz C (2017) Data augmentation for low-resource neural machine translation. In: Proceedings of the 55th annual meeting of the ACL (volume 2: short papers). Association for Computational Linguistics, Vancouver, Canada, pp 567–573
Fadaee M, Monz C (2018) Back-translation sampling by targeting difficult words in neural machine translation. In: Proceedings of the 2018 conference on EMNLP. Association for Computational Linguistics, Brussels, Belgium, pp 436–446
Farajian MA, Bertoldi N, Negri M, Turchi M, Federico M (2018) Evaluation of terminology translation in instance-based neural mt adaptation
Farajian MA, Turchi M, Negri M, Federico M (2017) Multi-domain neural machine translation through unsupervised adaptation. In: Proceedings of the second conference on machine translation, pp 127–137
Federico M, Bentivogli L, Paul M, Stuker S (2011) Overview of the iwslt 2011 evaluation campaign. In: IWSLT 2011: proceedings of the international workshop on spoken language translation, pp 11–27, San Francisco
Finn C, Abbeel P, Levine S (2017) Model-agnostic meta-learning for fast adaptation of deep networks. In: Proceedings of the 34th international conference on machine learning-volume 70, pp 1126–1135. JMLR. org
Firat O, Cho K, Sankaran B, Vural FTY, Bengio Y (2017) Multi-way, multilingual neural machine translation. Comput Speech Language 45:236–252
Gao F, Zhu J, Wu L, Xia Y, Qin T, Cheng X, Zhou W, Liu T-Y (2019) Soft contextual data augmentation for neural machine translation. In: Proceedings of the 57th annual meeting of the ACL, pp 5539–5544
Graham Y, Haddow B, Koehn P (2019) Translationese in machine translation evaluation. CoRR, arxiv:abs/1906.09833
Gu J, Hassan H, Devlin J, Li VO (2018) Universal neural machine translation for extremely low resource languages. arXiv preprint arXiv:1802.05368
Gulcehre C, Firat O, Xu K, Cho K, Barrault L, Lin H-C, Bougares F, Schwenk H, Bengio Y (2015) On using monolingual corpora in neural machine translation. arXiv preprintarXiv:1503.03535
Guzmán F, Chen P-J, Ott M, Pino J, Lample G, Koehn P, Chaudhary V, Ranzato M (2019) Two new evaluation datasets for low-resource machine translation: Nepali-english and sinhala-english. arXiv preprint arXiv:1902.01382
Ha T-L, Niehues J, Waibel A (2017) Effective strategies in zero-shot neural machine translation. arXiv preprintarXiv:1711.07893
Haddow B, Kirefu F, (2020) Pmindia—a collection of parallel corpora of languages of india. arXiv preprintarXiv:2001.09907
Haque R, Penkale S, Jiang J, Way A (2012) Source-side suffix stripping for bengali-to-english smt. In: 2012 International conference on Asian language processing, pp 193–196. IEEE
He D, Xia Y, Qin T, Wang L, Yu N, Liu T-Y, Ma W-Y (2016) Dual learning for machine translation. In: Advances in neural information processing systems, pp 820–828
Hoang V CD, Koehn P, Haffari G, Cohn T (2018) Iterative back-translation for neural machine translation. In: Proceedings of the 2nd workshop on neural machine translation and generation, pp 18–24, Melbourne, Australia. Association for Computational Linguistics
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780
Huck M, Hangya V, Fraser A (2019) Better oov translation with bilingual terminology mining. In: Proceedings of the 57th annual meeting of the ACL, pp 5809–5815
Imamura K, Fujita A, Sumita E (2018) Enhancement of encoder and attention using target monolingual corpora in neural machine translation. In: Proceedings of the 2nd workshop on neural machine translation and generation, pp 55–63, Melbourne, Australia
Johnson M, Schuster M, Le QV, Krikun M, Wu Y, Chen Z, Thorat N, Viégas F, Wattenberg M, Corrado G, Hughes M, Dean J (2017) Google’s multilingual neural machine translation system: enabling zero-shot translation. Trans Assoc Comput Linguist 5:339–351
Jozefowicz R, Zaremba W, Sutskever I (2015) An empirical exploration of recurrent network architectures. In: International conference on machine learning, pp 2342–2350
Kim Y, Geng J, Ney H (2019) Improving unsupervised word-by-word translation with language model and denoising autoencoder. CoRR, arxiv:abs/1901.01590
Kocmi T, Bojar O (2018) Trivial transfer learning for low-resource neural machine translation. In: Proceedings of the third conference on machine translation: research papers, pp 244–252, Brussels, Belgium. Association for Computational Linguistics
Koehn P, Knowles R (2017) Six challenges for neural machine translation. CoRR, arxiv:abs/1706.03872
Koehn P, Och FJ, Marcu D (2003) Statistical phrase-based translation. HLT-NAACL 2003: conference combining Human Language Technology conference series and the North American Chapter of the Association for Computational Linguistics conference series. Edmonton, AB, pp 48–54
Kunchukuttan A, Mehta P, Bhattacharyya P (2018) The IIT Bombay English-Hindi parallel corpus. In Proceedings of the eleventh international conference on language resources and evaluation (LREC (2018) Miyazaki. Japan, European Language Resources Association (ELRA)
Lample G, Conneau A, Denoyer L, Ranzato M (2018a) Unsupervised machine translation using monolingual corpora only. In: International conference on learning representations (ICLR)
Lample G, Ott M, Conneau A, Denoyer L, Ranzato M (2018b) Phrase-based & neural unsupervised machine translation. In: Proceedings of the 2018 conference on Empirical Methods in Natural Language Processing (EMNLP)
Lan Z, Chen M, Goodman S, Gimpel K, Sharma P, Soricut R (2019) Albert: A lite bert for self-supervised learning of language representations. arXiv preprint arXiv:1909.11942
Leng Y, Tan X, Qin T, Li X-Y, Liu T-Y (2019) Unsupervised pivot translation for distant languages. In: Proceedings of the 57th annual meeting of the ACL, pp 175–183
Lewis M, Liu Y, Goyal N, Ghazvininejad M, Mohamed A, Levy O, Stoyanov V, Zettlemoyer L (2019) Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461
Liu Y, Gu J, Goyal N, Li X, Edunov S, Ghazvininejad M, Lewis M, Zettlemoyer L (2020) Multilingual denoising pre-training for neural machine translation. arXiv preprint arXiv:2001.08210
Liu Y, Ott M, Goyal N, Du J, Joshi M, Chen D, Levy O, Lewis M, Zettlemoyer L, Stoyanov V (2019) Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692
Luong M-T, Le QV, Sutskever I, Vinyals O, Kaiser L (2016) Multi-task sequence to sequence learning. In: International conference on learning representations (ICLR), San Juan, Puerto Rico
Marie B, Fujita A (2018) Unsupervised neural machine translation initialized by unsupervised statistical machine translation. arXiv preprint arXiv:1810.12703
Marie B, Rubino R, Fujita A (2020) Tagged back-translation revisited: Why does it really work? In: Proceedings of the 58th annual meeting of the ACL, Online. Association for Computational Linguistics, pp 5990–5997
Miura A, Neubig G, Sudoh K, Nakamura S (2017) Tree as a pivot: Syntactic matching methods in pivot translation. In: Proceedings of the second conference on machine translation, volume 1: research paper, pp 90–98, Copenhagen, Denmark
Mueller A, Lal YK (2019) Sentence-level adaptation for low-resource neural machine translation. In: Proceedings of the 2nd workshop on technologies for MT of low resource languages, pp 39–47
Nakazawa T, Ding C, Dabre R, Kunchukuttan A, Doi N, Oda Y, Bojar O, Parida S, Goto I, Mino H (2019) Proceedings of the 6th workshop on Asian translation. In: Proceedings of the 6th workshop on Asian translation
Neubig G, Hu J (2018) Rapid adaptation of neural machine translation to new languages. In: Proceedings of the 2018 conference on EMNLP. Association for Computational Linguistics, Brussels, Belgium, pp 875–880
Nguyen T, Chiang D (2018) Improving lexical choice in neural machine translation. In: Proceedings of the 2018 conference of the North American chapter of the ACL: human language technologies, volume 1 (long papers). New Orleans, Louisiana, pp 334–343
Nguyen TQ, Chiang D (2017) Transfer learning across low-resource, related languages for neural machine translation. In: Proceedings of the eighth international joint conference on natural language processing (volume 2: short papers), pp 296–301, Taipei, Taiwan. Asian Federation of Natural Language Processing
Niehues J, Cho E (2017) Exploiting linguistic resources for neural machine translation using multi-task learning. In: Proceedings of the second conference on machine translation, pp 80–89, Copenhagen, Denmark. Association for Computational Linguistics
Nießen S, Ney H (2004) Statistical machine translation with scarce resources using morpho-syntactic information. Comput Linguist 30(2):181–204
Nishimura Y, Sudoh K, Neubig G, Nakamura S (2018) Multi-source neural machine translation with data augmentation. arXiv preprint arXiv:1810.06826
Nishimura Y, Sudoh K, Neubig G, Nakamura S (2019) Multi-source neural machine translation with missing data. IEEE/ACM Trans Audio Speech Language Process 28:569–580
Niu X, Denkowski M, Carpuat M (2018) Bi-directional neural machine translation with synthetic parallel data. CoRR, arxiv:abs/1805.11213
Östling R, Tiedemann J (2017) Neural machine translation for low-resource languages. CoRR, arxiv:abs/1708.05729
Pan SJ, Yang Q (2009) A survey on transfer learning. IEEE Trans knowl Data Eng 22(10):1345–1359
Passban P, Liu Q, Way A (2017) Translating low-resource languages by vocabulary adaptation from close counterparts. ACM Trans Asian Low-Resour Language Inf Process (TALLIP) 16(4):1–14
Passban P, Liu Q, Way A (2018a) Improving character-based decoding using target-side morphological information for neural machine translation. In: Proceedings of the 2018 conference of the North American chapter of the ACL: human language technologies, volume 1 (long papers). New Orleans, Louisiana, pp 58–68
Passban P, Way A, Liu Q (2018b) Tailoring neural architectures for translating from morphologically rich languages. In: Proceedings of the 27th International Conference on Computational Linguistics. Santa Fe, New Mexico, USA, pp 3134–3145
Poncelas A, Shterionov D, Way A, de Buy Wenniger GM, Passban P (2018) Investigating backtranslation in neural machine translation. In: Proceedings of The 21st annual conference of the European Association for Machine Translation (EAMT 2018). Alicante, Spain, pp 249–258
Post M, Callison-Burch C, Osborne M (2012) Constructing parallel corpora for six indian languages via crowdsourcing. In: Proceedings of the seventh workshop on statistical machine translation, pp 401–409, Montréal, Canada. Association for Computational Linguistics
Pourdamghani N, Aldarrab N, Ghazvininejad M, Knight K, May J (2019) Translating translationese: a two-step approach to unsupervised machine translation. In: Proceedings of the 57th annual meeting of the ACL. Association for Computational Linguistics, Florence, Italy, pp 3057–3062
Qi Y, Sachan D, Felix M, Padmanabhan S, Neubig G (2018) When and why are pre-trained word embeddings useful for neural machine translation? In: Proceedings of the 2018 conference of the North American Chapter of the ACL: human language technologies, volume 2 (short papers). Association for Computational Linguistics, New Orleans, Louisiana, pp 529–535
Ramasamy L, Bojar O, Žabokrtskỳ Z (2014) Entam: An english-tamil parallel corpus (entam v2. 0). LINDAT/CLARIN digital library at the Institute of Formal and Applied Linguistics (ÚFAL), Faculty of Mathematics and Physics, Charles University
Ramesh SH, Sankaranarayanan KP (2018) Neural machine translation for low resource languages using bilingual lexicon induced from comparable corpora. In: Proceedings of the 2018 conference of the North American chapter of the ACL: student research workshop. Louisiana, USA, New Orleans, pp 112–119
Resnik P, Smith NA (2003) The web as a parallel corpus. Comput Linguist 29(3):349–380
Schwenk H, Chaudhary V, Sun S, Gong H, Guzmán F (2019) Wikimatrix: Mining 135m parallel sentences in 1620 language pairs from wikipedia. arXiv preprint arXiv:1907.05791
Sennrich R, Haddow B, Birch A (2016a) Improving neural machine translation models with monolingual data. In: Proceedings of the 54th annual meeting of the ACL (volume 1: long papers). Association for Computational Linguistics, Berlin, Germany, pp 86–96
Sennrich R, Haddow B, Birch A (2016b) Neural machine translation of rare words with subword units. In: Proceedings of the 54th annual meeting of the ACL (volume 1: long papers). Berlin, Germany, pp 1715–1725
Sennrich R, Zhang B (2019) Revisiting low-resource neural machine translation: A case study. In: Proceedings of the 57th annual meeting of the ACL. Association for Computational Linguistics, Florence, Italy, pp 211–221
Shearing S, Kirov C, Khayrallah H, Yarowsky D (2018) Improving low resource machine translation using morphological glosses (non-archival extended abstract). In: Proceedings of the 13th conference of the Association for Machine Translation in the Americas (vol 1: research papers). Boston, MA, pp 132–139
Søgaard A, Ruder S, Vulić I (2018) On the limitations of unsupervised bilingual dictionary induction. In: Proceedings of the 56th annual meeting of the ACL (vol 1: long papers). Association for Computational Linguistics, Melbourne, Australia, pp 778–788
Strassel S, Tracey J (2016) LORELEI language packs: Data, tools, and resources for technology development in low resource languages. In: Proceedings of the tenth international conference on language resources and evaluation (LREC 2016), pp 3273–3280, Portorož, Slovenia. European Language Resources Association (ELRA)
Sutskever I, Vinyals O, Le QV (2014) Sequence to sequence learning with neural networks. In: Advances in neural information processing systems, pp 3104–3112
Tiedemann J (2012) Parallel data, tools and interfaces in OPUS. In: Proceedings of the 8th international conference on language resources and evaluation (LREC’2012), pp 2214–2218, Istanbul, Turkey
Toral A, Sánchez-Cartagena VM (2017) A multifaceted evaluation of neural versus phrase-based machine translation for 9 language directions. In: Proceedings of the 15th conference of the European chapter of the ACL: vol 1, long papers. Association for Computational Linguistics, Valencia, Spain, pp 1063–1073
Ueffing N, Haffari G, Sarkar A (2007) Transductive learning for statistical machine translation. In: Proceedings of the 45th annual meeting of the association of computational linguistics, pp 25–32, Prague, Czech Republic. Association for Computational Linguistics
Utiyama M, Isahara H (2007) A comparison of pivot methods for phrase-based statistical machine translation. In: Proceedings of human language technologies, the conference of the North American chapter of the ACL (NAACL 2007). Rochester, USA, pp 484–491
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017a) Attention is all you need. CoRR, arxiv:abs/1706.03762
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017b) Attention is all you need. In: Advances in neural information processing systems, pp 5998–6008
Wang X, Pham H, Dai Z, Neubig G (2018) SwitchOut: an efficient data augmentation algorithm for neural machine translation. In: Proceedings of the 2018 conference on EMNLP. Association for Computational Linguistics, Brussels, Belgium, pp 856–861
Wu H, Wang H (2007) Pivot language approach for phrase-based statistical machine translation. Mach Transl 21(3):165–181
Wuebker J, Simianer P, DeNero J (2018) Compact personalized models for neural machine translation. In: Proceedings of the 2018 conference on EMNLP. Association for Computational Linguistics, Brussels, Belgium, pp 881–886
Xia M, Kong X, Anastasopoulos A, Neubig G (2019) Generalized data augmentation for low-resource translation. In: Proceedings of the 57th annual meeting of the ACL. Association for Computational Linguistics, Florence, Italy, pp 5786–5796
Yang Z, Chen W, Wang F, Xu B (2018) Unsupervised neural machine translation with weight sharing. In: Proceedings of the 56th annual meeting of the ACL (vol 1: long papers). Association for Computational Linguistics, Melbourne, Australia, pp 46–55
Yang Z, Dai Z, Yang Y, Carbonell J, Salakhutdinov RR, Le QV (2019) XLNet: Generalized autoregressive pretraining for language understanding. In: Wallach H, Larochelle H, Beygelzimer A, d’Alché-Buc F, Fox E, Garnett R (eds) Advances in Neural Information Processing Systems, vol 32. Curran Associates Inc
Yarowsky D, Ngai G, Wicentowski R (2001) Inducing multilingual text analysis tools via robust projection across aligned corpora. In: Proceedings of the first international conference on human language technology research
Zaremoodi P, Buntine W, Haffari G (2018) Adaptive knowledge sharing in multi-task learning: Improving low-resource neural machine translation. In: Proceedings of the 56th annual meeting of the ACL (vol 2: short papers), pp 656–661
Zaremoodi P, Haffari G (2018) Neural machine translation for bilingually scarce scenarios: a deep multi-task learning approach. In: Proceedings of the 2018 conference of the North American chapter of the ACL: human language technologies, vol 1 (long papers). New Orleans, Louisiana, pp 1356–1365
Zhang J, Zong C (2016) Exploiting source-side monolingual data in neural machine translation. In: Proceedings of the 2016 conference on EMNLP. Association for Computational Linguistics, Austin, Texas, pp 1535–1545
Zhou C, Ma X, Hu J, Neubig G (2019) Handling syntactic divergence in low-resource machine translation. In: Proceedings of the 2019 conference on EMNLP and the 9th international joint conference on natural language processing (EMNLP-IJCNLP). Association for Computational Linguistics, Hong Kong, China, pp 1388–1394
Zhu X, He Z, Wu H, Wang H, Zhu C, Zhao T (2013) Improving pivot-based statistical machine translation using random walk. In: Proceedings of the 2013 conference on EMNLP. Seattle, USA, pp 524–534
Ziemski M, Junczys-Dowmunt M, Pouliquen B (2016) The united nations parallel corpus v1.0. In: Proceedings of the international conference on language resources and evaluation (LREC), pp 1–5, Portorož, Slovenia
Zoph B, Yuret D, May J, Knight K (2016a) Transfer learning for low-resource neural machine translation. arXiv preprint arXiv:1604.02201
Zoph B, Yuret D, May J, Knight K (2016b) Transfer learning for low-resource neural machine translation. In: Proceedings of the 2016 conference on empirical methods in natural language processing, pp 1568–1575, Austin, Texas
Acknowledgements
The ADAPT Centre for Digital Content Technology is funded under the Science Foundation Ireland (SFI) Research Centres Programme (Grant No. 13/RC/2106) and is co-funded under the European Regional Development Fund. This project has partially received funding from the European Union’s Horizon 2020 research and innovation programme under the Marie Skłodowska-Curie grant agreement No. 713567, and the publication has emanated from research supported in part by a research grant from SFI under Grant Number 13/RC/2077.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Haque, R., Liu, CH. & Way, A. Recent advances of low-resource neural machine translation. Machine Translation 35, 451–474 (2021). https://doi.org/10.1007/s10590-021-09281-1
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10590-021-09281-1