257
Views
3
CrossRef citations to date
0
Altmetric
Computers and Computing

Analysis of Neural Machine Translation KANGRI Language by Unsupervised and Semi Supervised Methods

ORCID Icon, ORCID Icon & ORCID Icon

REFERENCES

  • D. Bahdanau, K. Cho, and Y. Bengio, “Neural machine translation by jointly learning to align and translate,” arXiv preprint arXiv:1409.0473,2014.
  • P. Koehn and R. Knowles, “Six challenges for neural machine translation,” arXiv preprint arXiv:1706.03872,2017.
  • S. Ravi and K. Knight, “Deciphering foreign language,” in Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, 2011, pp. 12–21.
  • M. Nuhn and H. Ney, “Decipherment complexity in 1: 1 substitution ciphers.” in Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2013, pp. 615–21.
  • E. Kim, K. Huang, S. Jegelka, and E. Olivetti, “Virtual screening of inorganic materials synthesis parameters with deep learning,” NPJ Comput. Mater., Vol. 3, no. 1, pp. 1–9, 2017.
  • M. Artetxe, G. Labaka, E. Agirre, and K. Cho, “Unsupervised neural machine translation,” arXiv preprint arXiv:1710.11041, 2017.
  • G. Lample, M. Ott, A. Conneau, L. Denoyer, and M. Ranzato, “Phrase-based & neural unsupervised machine translation,” in Proceedings of the 2018 Conference on Empirical Methods.
  • Y. Kim, J. Geng, and H. Ney, “Improving unsupervised word-by-word translation with language model and denoising autoencoder,” arXiv preprint arXiv:1901.01590, 2019.
  • R. Sennrich, B. Haddow, and A. Birch, “Improving neural machine translation models with monolingual data,” in Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2016, pp. 86–96.
  • D. He, Y. Xia, T. Qin, L. Wang, N. Yu, T. Liu, and W.-Y. Ma, “Dual learning for machine translation,” in Proceedings of Advances in Neural Information Processing Systems (NeurIPS), 2016, pp. 820–8.
  • Y. Kim, M. Graça, and H. Ney, “When and why is unsupervised neural machine translation useless?,” arXiv preprint arXiv:2004.10581, 2020.
  • Y. Liu, et al., “Multilingual denoising pre-training for neural machine translation,” Trans. Assoc. Comput. Linguist., Vol. 8, pp. 726–42, 2020.
  • J. Khatri and P. Bhattacharyya, “Filtering back-translated data in unsupervised neural machine translation,”. in Proceedings of the 28th International Conference on Computational Linguistics, 2020, pp. 4334–9.
  • P. Kishore, S. Roukos, T. Ward, and W. J. Zhu, “BLEU: A method for automatic evaluation of machine translation,” in Proceedings of the 40th annual meeting of the Association for Computational Linguistics, 2002, pp. 311–18.
  • R. Sennrich, et al., “Nematus: A toolkit for neural machine translation.” arXiv preprint arXiv:1703.04357, 2017.
  • G. Zuckermann, “Historical and moral arguments for language reclamation,” History and Philosophy of the Language Sciences, 2013, pp. 26.
  • unesco.org (2017). Available: http://www.unesco.org/languages-atlas/ [Online; accessed 04-July-2017].
  • Wikipedia. (2020). Kangri language. Available: https://en.wikipedia.org/wiki/Kangri_language
  • M. Artetxe, G. Labaka, and E. Agirre, “A robust self-learning method for fully unsupervised cross-lingual mappings of word embeddings,” in Proceedings of the 56th Annual Meeting of the Association, for Computational Linguistics (Volume 1: Long Papers), 2018, pp. 789–98.
  • R. Sennrich, B. Haddow, and A. Birch, “Improving neural machine translation models with monolingual data,” in Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2016, pp. 86–96.
  • K. J. Heafield, “KenLM: Faster and smaller language model queries.” in Proceedings of the Sixth Workshop on Statistical Machine Translation (pp. 187–97). Association for Computational Linguistics, 2011.
  • F. Hill, K. Cho, and A. Korhonen, “Learning distributed representations of sentences from unlabelled data,” arXiv preprint arXiv:1602.03483, 2016.
  • R. Sennrich, B. Haddow, and A. Birch, “Neural machine translation of rare words with subword units.” arXiv preprint arXiv:1508.07909, 2015.
  • N. Garneau, J. S. Leboeuf, and L. Lamontagne, “Contextual generation of word embeddings for out of vocabulary words in downstream tasks,” in Canadian Conference on Artificial Intelligence (pp. 563–9). Springer, Cham, 2019.
  • S. Banerjee and A. Lavie, “METEOR: An automatic metric for MT evaluation with improved correlation with human judgments,” in Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization, 2005.
  • M. Snover, et al., “A study of translation edit rate with targeted human annotation,” in Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers, 2006.
  • C.-Y. Lin, “Rouge: A package for automatic evaluation of summaries,” in Text Summarization Branches Out, Lin Chin-Yew, Ed. Barcelona, Spain: Association for Computational Linguistics, 2004, pp. 74–81.
  • M. Przybocki, et al., “The NIST 2008 metrics for machine translation challenge—overview, methodology, metrics, and results,” Mach. Transl., Vol. 23, no. 2, pp. 71–103, 2009.
  • P. Li, C. Chen, W. Zheng, Y. Deng, F. Ye, and Z. Zheng, “STD: An automatic evaluation metric for machine translation based on word embeddings,” IEEE/ACM Trans. Audio Speech Lang. Process., Vol. 27, no. 10, pp. 1497–506, 2019.
  • A. Joulin, E. Grave, P. Bojanowski, and T. Mikolov, “Bag of tricks for efficient text classification,” arXiv preprint arXiv:1607.01759, 2016.
  • G. Dinu, A. Lazaridou, and M. Baroni. “Improving zero-shot learning by mitigating the hubness problem,” arXiv preprint arXiv:1412.6568, 2014.
  • S. L. Smith, D. H. Turban, S. Hamblin, and N. Y. Hammerla. “Offline bilingual word vectors, orthogonal transformations and the inverted softmax,” arXiv preprint arXiv:1702.03859, 2017.
  • H. Azarbonyad, A. Shakery, and H. Faili, “A learning to rank approach for cross-language information retrieval exploiting multiple translation resources,” Nat. Lang. Eng., Vol. 25, no. 3, pp. 363–84, 2019.
  • A. Vaswani, et al., “Attention is all you need,” in Advances in Neural Information Processing Systems 30 (NIPS 2017), V. Ashish , S. Noam, P. Niki, U. Jakob, J. Llion, G. N. Aidan, K. Łukasz and P. Illia Eds. San Diego, CA: Curran Associates, Inc. 2017, pp. 5998–6008.
  • F. Hieber, T. Domhan, M. Denkowski, D. Vilar, A. Sokolov, A. Clifton, and M. Post. “Sockeye: A toolkit for neural machine translation,” arXiv preprint arXiv:1712.05690, 2017.
  • A. Kunchukuttan, P. Mehta, and P. Bhattacharyya. “The IIT Bombay English-Hindi parallel corpus,” in Proceedings of the Eleventh International Conference on Language Resources and Evaluation, 2018.
  • S. Chauhan, S. Saxena, and P. Daniel. “Monolingual and parallel corpora for Kangri low resource language,” arXiv preprint arXiv:2103.11596, 2021.
  • https://github.com/taranjeet/hindi-tokenizer
  • D. P. Kingma and J. Ba. “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.
  • S. Chauhan, S. Saxena, and P. Daniel, “Fully unsupervised word translation from cross-lingual word embeddings especially for healthcare professionals,” Int. J. Syst. Assur. Eng. Manag., Vol. 67, 1–10, 2021.
  • S. Chauhan, D. Philemon, M. Archita, and A. Kumar, “AdaBLEU: A modified BLEU score for morphologically rich languages,” IETE. J. Res., Vol. 12, 1–12, 2021.
  • S. Chauhan, U. Pant, M. Mustafa, and P. Daniel, “A robust unsupervised word by word translation for morphological rich languages using different retrieval techniques,” J. Crit. Rev., Vol. 7, no. 17, pp. 2677–84, 2020.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.