REFERENCES
- D. Bahdanau, K. Cho, and Y. Bengio, “Neural machine translation by jointly learning to align and translate,” arXiv preprint arXiv:1409.0473,2014.
- P. Koehn and R. Knowles, “Six challenges for neural machine translation,” arXiv preprint arXiv:1706.03872,2017.
- S. Ravi and K. Knight, “Deciphering foreign language,” in Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, 2011, pp. 12–21.
- M. Nuhn and H. Ney, “Decipherment complexity in 1: 1 substitution ciphers.” in Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2013, pp. 615–21.
- E. Kim, K. Huang, S. Jegelka, and E. Olivetti, “Virtual screening of inorganic materials synthesis parameters with deep learning,” NPJ Comput. Mater., Vol. 3, no. 1, pp. 1–9, 2017.
- M. Artetxe, G. Labaka, E. Agirre, and K. Cho, “Unsupervised neural machine translation,” arXiv preprint arXiv:1710.11041, 2017.
- G. Lample, M. Ott, A. Conneau, L. Denoyer, and M. Ranzato, “Phrase-based & neural unsupervised machine translation,” in Proceedings of the 2018 Conference on Empirical Methods.
- Y. Kim, J. Geng, and H. Ney, “Improving unsupervised word-by-word translation with language model and denoising autoencoder,” arXiv preprint arXiv:1901.01590, 2019.
- R. Sennrich, B. Haddow, and A. Birch, “Improving neural machine translation models with monolingual data,” in Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2016, pp. 86–96.
- D. He, Y. Xia, T. Qin, L. Wang, N. Yu, T. Liu, and W.-Y. Ma, “Dual learning for machine translation,” in Proceedings of Advances in Neural Information Processing Systems (NeurIPS), 2016, pp. 820–8.
- Y. Kim, M. Graça, and H. Ney, “When and why is unsupervised neural machine translation useless?,” arXiv preprint arXiv:2004.10581, 2020.
- Y. Liu, et al., “Multilingual denoising pre-training for neural machine translation,” Trans. Assoc. Comput. Linguist., Vol. 8, pp. 726–42, 2020.
- J. Khatri and P. Bhattacharyya, “Filtering back-translated data in unsupervised neural machine translation,”. in Proceedings of the 28th International Conference on Computational Linguistics, 2020, pp. 4334–9.
- P. Kishore, S. Roukos, T. Ward, and W. J. Zhu, “BLEU: A method for automatic evaluation of machine translation,” in Proceedings of the 40th annual meeting of the Association for Computational Linguistics, 2002, pp. 311–18.
- R. Sennrich, et al., “Nematus: A toolkit for neural machine translation.” arXiv preprint arXiv:1703.04357, 2017.
- G. Zuckermann, “Historical and moral arguments for language reclamation,” History and Philosophy of the Language Sciences, 2013, pp. 26.
- unesco.org (2017). Available: http://www.unesco.org/languages-atlas/ [Online; accessed 04-July-2017].
- Wikipedia. (2020). Kangri language. Available: https://en.wikipedia.org/wiki/Kangri_language
- M. Artetxe, G. Labaka, and E. Agirre, “A robust self-learning method for fully unsupervised cross-lingual mappings of word embeddings,” in Proceedings of the 56th Annual Meeting of the Association, for Computational Linguistics (Volume 1: Long Papers), 2018, pp. 789–98.
- R. Sennrich, B. Haddow, and A. Birch, “Improving neural machine translation models with monolingual data,” in Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2016, pp. 86–96.
- K. J. Heafield, “KenLM: Faster and smaller language model queries.” in Proceedings of the Sixth Workshop on Statistical Machine Translation (pp. 187–97). Association for Computational Linguistics, 2011.
- F. Hill, K. Cho, and A. Korhonen, “Learning distributed representations of sentences from unlabelled data,” arXiv preprint arXiv:1602.03483, 2016.
- R. Sennrich, B. Haddow, and A. Birch, “Neural machine translation of rare words with subword units.” arXiv preprint arXiv:1508.07909, 2015.
- N. Garneau, J. S. Leboeuf, and L. Lamontagne, “Contextual generation of word embeddings for out of vocabulary words in downstream tasks,” in Canadian Conference on Artificial Intelligence (pp. 563–9). Springer, Cham, 2019.
- S. Banerjee and A. Lavie, “METEOR: An automatic metric for MT evaluation with improved correlation with human judgments,” in Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization, 2005.
- M. Snover, et al., “A study of translation edit rate with targeted human annotation,” in Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers, 2006.
- C.-Y. Lin, “Rouge: A package for automatic evaluation of summaries,” in Text Summarization Branches Out, Lin Chin-Yew, Ed. Barcelona, Spain: Association for Computational Linguistics, 2004, pp. 74–81.
- M. Przybocki, et al., “The NIST 2008 metrics for machine translation challenge—overview, methodology, metrics, and results,” Mach. Transl., Vol. 23, no. 2, pp. 71–103, 2009.
- P. Li, C. Chen, W. Zheng, Y. Deng, F. Ye, and Z. Zheng, “STD: An automatic evaluation metric for machine translation based on word embeddings,” IEEE/ACM Trans. Audio Speech Lang. Process., Vol. 27, no. 10, pp. 1497–506, 2019.
- A. Joulin, E. Grave, P. Bojanowski, and T. Mikolov, “Bag of tricks for efficient text classification,” arXiv preprint arXiv:1607.01759, 2016.
- G. Dinu, A. Lazaridou, and M. Baroni. “Improving zero-shot learning by mitigating the hubness problem,” arXiv preprint arXiv:1412.6568, 2014.
- S. L. Smith, D. H. Turban, S. Hamblin, and N. Y. Hammerla. “Offline bilingual word vectors, orthogonal transformations and the inverted softmax,” arXiv preprint arXiv:1702.03859, 2017.
- H. Azarbonyad, A. Shakery, and H. Faili, “A learning to rank approach for cross-language information retrieval exploiting multiple translation resources,” Nat. Lang. Eng., Vol. 25, no. 3, pp. 363–84, 2019.
- A. Vaswani, et al., “Attention is all you need,” in Advances in Neural Information Processing Systems 30 (NIPS 2017), V. Ashish , S. Noam, P. Niki, U. Jakob, J. Llion, G. N. Aidan, K. Łukasz and P. Illia Eds. San Diego, CA: Curran Associates, Inc. 2017, pp. 5998–6008.
- F. Hieber, T. Domhan, M. Denkowski, D. Vilar, A. Sokolov, A. Clifton, and M. Post. “Sockeye: A toolkit for neural machine translation,” arXiv preprint arXiv:1712.05690, 2017.
- A. Kunchukuttan, P. Mehta, and P. Bhattacharyya. “The IIT Bombay English-Hindi parallel corpus,” in Proceedings of the Eleventh International Conference on Language Resources and Evaluation, 2018.
- S. Chauhan, S. Saxena, and P. Daniel. “Monolingual and parallel corpora for Kangri low resource language,” arXiv preprint arXiv:2103.11596, 2021.
- https://github.com/taranjeet/hindi-tokenizer
- D. P. Kingma and J. Ba. “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.
- S. Chauhan, S. Saxena, and P. Daniel, “Fully unsupervised word translation from cross-lingual word embeddings especially for healthcare professionals,” Int. J. Syst. Assur. Eng. Manag., Vol. 67, 1–10, 2021.
- S. Chauhan, D. Philemon, M. Archita, and A. Kumar, “AdaBLEU: A modified BLEU score for morphologically rich languages,” IETE. J. Res., Vol. 12, 1–12, 2021.
- S. Chauhan, U. Pant, M. Mustafa, and P. Daniel, “A robust unsupervised word by word translation for morphological rich languages using different retrieval techniques,” J. Crit. Rev., Vol. 7, no. 17, pp. 2677–84, 2020.