Professional Documents
Culture Documents
NMT Based Similar Language Translation For Hindi - Marathi
NMT Based Similar Language Translation For Hindi - Marathi
414
Proceedings of the 5th Conference on Machine Translation (WMT), pages 414–417
Online, November 19–20, 2020.
2020
c Association for Computational Linguistics
and monolingual) on which we carried out all exper- (3) aur jab maan@@ saa##haaree
iments. We deliberately excluded Indic WordNet pakshee loth##on par jha@@ pat##e ,
data from the training after doing manual quality tab ab##raam ne unhen uda diya .
check. As this is a constrained task, our experi- ‘And when the carnivorous birds swooped on
ments do not utilise any other available data. the carcasses, Abram blew them away.’
415
Model Feature BPE (Merge ops) BLEU
BiLSTM + LuongAttn Word level - 19.70
BiLSTM + LuongAttn Word + Shared Vocab (SV)+ POS - 20.49
BiLSTM + LuongAttn BPE 10K 20.1
BiLSTM + LuongAttn BPE+SV+MORPH Segmentation 10K 20.44
BiLSTM + LuongAttn BPE+SV+MORPH+POS 10K 20.62
BiLSTM + LuongAttn BPE+SV+MORPH+POS + BT 10K 16.49
• Morph + BPE based subword segmentation, on given training data for a direction (i.e, Marathi
POS tags as feature to Hindi) to enrich training data of the opposite
directional NMT training (i.e, Hindi - Marathi) by
• Embedding size : 500 populating synthetic data. We used around 5M
• RNN for encoder and decoder: bi-LSTM back translated pairs (after perplexity based prun-
ing with respect to sentence length) for both trans-
• Bi-LSTM dimension : 500 lation directions.
Using above described configuration, we per-
• encoder - decoder layers : 2 formed experiments based on different parameter
• Attention : luong (general) (feature) configurations. We trained and tested our
models on word level, BPE level and morph + BPE
• copy attention(Gu et al., 2016) on dynamically level for input and output. We also used POS tagger
generated dictionary and experimented with shared vocabulary across
the translation task. The results are discussed in
• label smoothing : 1.0 following Result section.
• dropout : 0.30
6 Result
• Optimizer : Adam
Table-2 and Table-3 show performance of systems
• Beam size : 4 (train) and 10 (test) with different configuration in terms of BLEU
score(Papineni et al., 2002) for Hindi-Marathi and
As these are two similar languages, share writing Marathi-Hindi respectively on the validation data.
scripts and large sets of named entities, we used We achieved 20.62 and 25.55 development and
shared vocab across training. We used Opennmt-py 5.94 and 18.14 test BLEU scores for Hindi-Marathi
(Klein et al., 2020) toolkit with above configuration and Marathi-Hindi systems respectively.
for our experiments.
The results show that for low resource similar
5 Back Translation
language settings, MT models based on sequence
Back translation is a widely used data augmentation to sequence neural network can be improved
method for low resource neural machine transla- with linguistic information like morph based
tion(Sennrich et al., 2016a). We utilised monolin- segmentation and POS features. The results also
gual data (i.e of Marathi) and a NMT model trained show that morph based segmentation along with
416
byte pair encoding improves BLEU score for both Philipp Koehn. 2009. Statistical machine translation.
directions. But Marathi-Hindi directed translation Cambridge University Press.
shows considerable improvement. Therefore our Minh-Thang Luong, Hieu Pham, and Christopher D
method shows improvement while translating Manning. 2015. Effective approaches to attention-
from morphologically richer language (Marathi) based neural machine translation. In Proceedings of
to comparatively less morphologically richer the 2015 Conference on Empirical Methods in Natu-
ral Language Processing, pages 1412–1421.
language (Hindi).
wikipedia Marathi. 2020. Maharashtri prakrit
The results also suggest that the use of back - wikipedia. https://en.wikipedia.org/
wiki/Maharashtri_Prakrit. (Accessed on
translated synthetic data for low resource language 08/15/2020).
pairs reduces the overall performance marginally.
The reason for this could be, due to low quantity of Kishore Papineni, Salim Roukos, Todd Ward, and Wei-
training data for NMT models, they could be over Jing Zhu. 2002. Bleu: a method for automatic eval-
uation of machine translation. In Proceedings of the
learning and back translation could be helping to 40th annual meeting of the Association for Compu-
do better generalization. tational Linguistics, pages 311–318.
417