Text Generation and Enhanced Evaluation of Metric for Machine Translation

  • Conference paper
  • First Online:
Data Intelligence and Cognitive Informatics

Part of the book series: Algorithms for Intelligent Systems ((AIS))

Abstract

Here the power of a recurrent neural network (RNN) has been exhibited for generating grammatically correct new text from given input text and translation of the new text to the Hindi language with modified bilingual evaluation understudy (BLEU) metric score. Our system aims to generate a grammatically correct new text from given input sentences or paragraphs and translate generated text to Hindi with high translation score. To accomplish a grammatically correct sentence, natural language toolkit (NLTK) is used for grammar correction at the end of text generation. RNN is not very useful for text generation of a gated connection decided to be used for this purpose. The generated text is transferred to machine translation (MT) module. For MT since evaluation is done by humans is a time-consuming task and results differ from evaluator to another evaluator. Hence, the need for assessment of translation system is emerged. The synonym issue is not considered by the BLEU metric. A synonym is treated as a separate word. A modified BLEU (M-BLEU) has been developed as evaluation metrics. It includes several features such as replacing synonym and shallow modules of parsing. The final score of translation is given by BLEU metric scores. Finally, two outputs are there: one is generated text (English) and second is translated text with an improved translation score (Hindi).

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
EUR 32.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or Ebook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
EUR 29.95
Price includes VAT (Spain)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
EUR 160.49
Price includes VAT (Spain)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
EUR 207.99
Price includes VAT (Spain)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free ship** worldwide - see info
Hardcover Book
EUR 207.99
Price includes VAT (Spain)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free ship** worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Bengio Y, Simard P, Frasconi P (1994) Learning long-term dependencies with gradient descent is difficult. IEEE Trans Neural Netw 5:157–166

    Article  Google Scholar 

  2. Pollastri G, Przybylski D, Rost B, Baldi P (2002) Improving the prediction of protein secondary structure in three and eight classes using recurrent neural networks and profiles. Proteins Struct Funct Genet 47:228–235

    Google Scholar 

  3. Chung J, Gulcehre C, Cho K, Bengio Y (2014) Empirical evaluation of gated recurrent neural networks on sequence modeling. ar**v preprint ar**v:1412.3555

  4. Souri A, El Maazouzi Z, Al Achhab M, El Mohajir B (2018) Arabic text generation using recurrent neural networks. In: Communications in computer and information science, pp 523–533

    Google Scholar 

  5. Gasthaus J, Wood F, Teh Y (2010) Lossless compression based on the sequence memoizer. In: 2010 data compression conference

    Google Scholar 

  6. Bahdanau D, Cho K, Bengio Y (2014) Neural machine translation by jointly learning to align and translate. ar**v preprint ar**v:1409.0473

  7. Luong M-T, Manning CD (2015) Stanford neural machine translation systems for spoken language domain. In: International workshop on spoken language translation

    Google Scholar 

  8. Papineni K, Roukos S, Ward T, Zhu W (2001) BLEU. In: Proceedings of the 40th annual meeting on association for computational linguistics - ACL 2002

    Google Scholar 

  9. Dwivedi SK, Sukhadeve PP (2010) Machine translation system in Indian perspectives. J Comp Sci 6(10):1111–1116. https://doi.org/10.3844/jcssp.2010.1111.1116

  10. Verwimp L, Renkens V, Wambacq P State gradients for RNN memory analysis. http://dx.doi.org/10.21437/Interspeech.2018-1153

  11. Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9:1735–1780

    Article  Google Scholar 

  12. Rozovskaya A, Roth D Grammatical error correction: machine translation and classifiers. https://www.aclweb.org/anthology/P16-1208

  13. Hermanto A, Adji T, Setiawan N (2015) Recurrent neural network language model for English-Indonesian machine translation: experimental study. In: 2015 International Conference on Science in Information Technology (ICSITech)

    Google Scholar 

  14. Castilho S, Doherty S, Gaspari F, Moorkens J (2018) Approaches to human and machine translation quality assessment. In: Machine translation: technologies and applications, pp 9–38

    Google Scholar 

  15. Bojanowski P, Joulin A, Mikolov T (2015) Alternative structures for character-level RNNs. ar**v preprint ar**v:1511.06303

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Sujit S. Amin .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Amin, S.S., Ragha, L. (2021). Text Generation and Enhanced Evaluation of Metric for Machine Translation. In: Jeena Jacob, I., Kolandapalayam Shanmugam, S., Piramuthu, S., Falkowski-Gilski, P. (eds) Data Intelligence and Cognitive Informatics. Algorithms for Intelligent Systems. Springer, Singapore. https://doi.org/10.1007/978-981-15-8530-2_1

Download citation

Publish with us

Policies and ethics

Navigation