International Journal For Multidisciplinary Research
E-ISSN: 2582-2160
•
Impact Factor: 9.24
A Widely Indexed Open Access Peer Reviewed Multidisciplinary Bi-monthly Scholarly International Journal
Home
Research Paper
Submit Research Paper
Publication Guidelines
Publication Charges
Upload Documents
Track Status / Pay Fees / Download Publication Certi.
Editors & Reviewers
View All
Join as a Reviewer
Get Membership Certificate
Current Issue
Publication Archive
Conference
Publishing Conf. with IJFMR
Upcoming Conference(s) ↓
Conferences Published ↓
DePaul-2026
IC-AIRCM-T3-2026
SPHERE-2025
AIMAR-2025
SVGASCA-2025
ICCE-2025
Chinai-2023
PIPRDA-2023
ICMRS'23
Contact Us
Plagiarism is checked by the leading plagiarism checker
Call for Paper
Volume 8 Issue 3
May-June 2026
Indexing Partners
Understanding the Limitations of Zero-Shot Large- Language Models on Hinglish and Tanglish Text
| Author(s) | Jaganathan B, Saravanan P |
|---|---|
| Country | India |
| Abstract | Large Language Models (LLMs) have shown remarkable skill in a wide range of Natural Language Processing (NLP) tasks. But we still don't know how well they perform in real life when there are more than one language, especially when code is mixed in. A lot of people in India utilise Hinglish (Hindi- English) and Tanglish (Tamil-English) on social media. People regularly switch languages in the middle of sentences and use grammar, transliteration, and slang from their area that isn't particularly professional. This study conducts an empirical error analysis of zero-shot LLMs employed for sentiment classification on code-mixed Indian texts. A comparative study is performed using two publicly available datasets: Hinglish and Tanglish. Sentiment categorisation employs the BART-Large-MNLI model in a zero-shot manner, lacking any task-specific training. To see how well the model performs, we look at its accuracy, precision, recall, F1-score, and confusion matrices. The results demonstrate that zero-shot LLMs don't work very well; they only get 31.1% correct on Hinglish and 43.0% correct on Tanglish datasets. Transliteration ambiguity, slang, irony, and complex code-switching are all challenges that come up again and again, according to error analysis. The work highlights the challenges faced by modern big language models in processing code-mixed Indian languages and stresses the imperative for language-specific adaptations in multilingual natural language processing systems. |
| Keywords | Large Language Models, Code-Mixed Languages, Hinglish, Tanglish, Sentiment Analysis, Error Analysis, Zero- Shot Learning, Multilingual NLP. |
| Field | Computer > Artificial Intelligence / Simulation / Virtual Reality |
| Published In | Volume 8, Issue 3, May-June 2026 |
| Published On | 2026-05-03 |
Share this

E-ISSN 2582-2160
CrossRef DOI is assigned to each research paper published in our journal.
IJFMR DOI prefix is
10.36948/ijfmr
Downloads
All research papers published on this website are licensed under Creative Commons Attribution-ShareAlike 4.0 International License, and all rights belong to their respective authors/researchers.
Powered by Sky Research Publication and Journals