International Journal For Multidisciplinary Research
E-ISSN: 2582-2160
•
Impact Factor: 9.24
A Widely Indexed Open Access Peer Reviewed Multidisciplinary Bi-monthly Scholarly International Journal
Home
Research Paper
Submit Research Paper
Publication Guidelines
Publication Charges
Upload Documents
Track Status / Pay Fees / Download Publication Certi.
Editors & Reviewers
View All
Join as a Reviewer
Get Membership Certificate
Current Issue
Publication Archive
Conference
Publishing Conf. with IJFMR
Upcoming Conference(s) ↓
Conferences Published ↓
IC-AIRCM-T3-2026
SPHERE-2025
AIMAR-2025
SVGASCA-2025
ICCE-2025
Chinai-2023
PIPRDA-2023
ICMRS'23
Contact Us
Plagiarism is checked by the leading plagiarism checker
Call for Paper
Volume 8 Issue 1
January-February 2026
Indexing Partners
A Comparative Analysis of BLEU and BERT Score for Evaluating Machine Translation of a Telugu Poem
| Author(s) | Prof. Madhu Jyoti Kolakaluri, Dr. Revathi Kudumula, Prasanna Guduru |
|---|---|
| Country | India |
| Abstract | The proliferation of Machine Translation (MT) systems has necessitated robust automatic evaluation metrics. While BLEU (Bi-Lingual Evaluation Understudy) and BERT (Bidirectional Encoder Representations from Transformers) These are the longstanding standards for assessing lexical and n-gram overlap. Its efficacy in evaluating semantically complex and culturally nuanced texts, such as poetry and prose, remains limited. This study presents a comparative evaluation of five NMT systems—Google Translate, Microsoft Translator, Quill Bot, Gork, and Bhashini on a corpus of Telugu literary texts, the poem "Bandipotlu". This research employs both the traditional BLEU metric and the context-aware BERT metric scores to analyse machine translation quality. Our findings indicate a significant divergence in the rankings provided by these metrics. BLEU scores are heavily penalizing creative paraphrasing and stylistic variations, whereas BERTScores leverage semantic embeddings, and they demonstrate a higher correlation with human intuitions for literary translation. The results suggest that BERTScore is a more suitable metric for evaluating the preservation of meaning, tone, and cultural nuance in literary machine translation, and it advocates for a paradigm shift beyond surface-level n-gram matching. |
| Keywords | Machine Translation, Evaluation Metrics, BLEU, BERTScore, Telugu Literature, Literary Translation, Natural Language Processing |
| Field | Arts |
| Published In | Volume 8, Issue 1, January-February 2026 |
| Published On | 2026-01-31 |
| DOI | https://doi.org/10.36948/ijfmr.2026.v08i01.67737 |
Share this

E-ISSN 2582-2160
CrossRef DOI is assigned to each research paper published in our journal.
IJFMR DOI prefix is
10.36948/ijfmr
Downloads
All research papers published on this website are licensed under Creative Commons Attribution-ShareAlike 4.0 International License, and all rights belong to their respective authors/researchers.