This paper describes a systematic study of an approach to Farsi-Spanish low-resource Neural Machine Translation (NMT) that leverages monolingual data for joint learning of forward and backward translation models. As is standard for NMT systems, the training process begins using two pre-trained translation models that are iteratively updated by decreasing translation costs. In each iteration, either translation model is used to translate monolingual texts from one language to another, to generate synthetic datasets for the other translation model. Two new translation models are then learned from bilingual data along with the synthetic texts. The key distinguishing feature between our approachmore »
This content will become publicly available on May 1, 2023
A generic neural network model to estimate populational neural activity for robust neural decoding
- Award ID(s):
- 1847319
- Publication Date:
- NSF-PAR ID:
- 10319953
- Journal Name:
- Computers in Biology and Medicine
- Volume:
- 144
- Issue:
- C
- ISSN:
- 0010-4825
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
Neural Machine Translation (NMT) performs training of a neural network employing an encoder-decoder architecture. However, the quality of the neural-based translations predominantly depends on the availability of a large amount of bilingual training dataset. In this paper, we explore the performance of translations predicted by attention-based NMT systems for Spanish to Persian low-resource language pairs. We analyze the errors of NMT systems that occur in the Persian language and provide an in-depth comparison of the performance of the system based on variations in sentence length and size of the training dataset. We evaluate our translation results using BLEU and humanmore »