International Journal For Multidisciplinary Research
E-ISSN: 2582-2160
•
Impact Factor: 9.24
A Widely Indexed Open Access Peer Reviewed Multidisciplinary Bi-monthly Scholarly International Journal
Home
Research Paper
Submit Research Paper
Publication Guidelines
Publication Charges
Upload Documents
Track Status / Pay Fees / Download Publication Certi.
Editors & Reviewers
View All
Join as a Reviewer
Get Membership Certificate
Current Issue
Publication Archive
Conference
Publishing Conf. with IJFMR
Upcoming Conference(s) ↓
Conferences Published ↓
DePaul-2026
IC-AIRCM-T3-2026
SPHERE-2025
AIMAR-2025
SVGASCA-2025
ICCE-2025
Chinai-2023
PIPRDA-2023
ICMRS'23
Contact Us
Plagiarism is checked by the leading plagiarism checker
Call for Paper
Volume 8 Issue 3
May-June 2026
Indexing Partners
Explainable Multimodal Deep Learning Framework For Early Disease Diagnosis
| Author(s) | Mr. Yash Raj, Mr. Ashmit Iyer, Mr. Gourab Bhadra, Ms. Moushumi Sarker, Dr. Anish Pandey |
|---|---|
| Country | India |
| Abstract | Early disease diagnosis remains a critical challenge in modern healthcare due to the fragmented nature of medical data, which often includes heterogeneous modalities such as medical imaging, clinical text, and physiological signals. Traditional deep learning models, while achieving high predictive performance, suffer from limited interpretability, thereby restricting their adoption in high-stakes clinical environments. This paper proposes an Explainable Multimodal Deep Learning Framework (EMDLF) designed to integrate and analyze diverse medical data sources while ensuring transparency in decision-making. The framework combines convolutional neural networks for image feature extraction, transformer-based architectures for textual understanding, and attention-driven fusion mechanisms for multimodal integration. To enhance interpretability, the model incorporates explainability techniques such as SHAP (Shapley Additive Explanations), Grad-CAM, and attention visualization, enabling clinicians to understand feature contributions and decision pathways. Experimental evaluation on benchmark healthcare datasets demonstrates improved diagnostic accuracy compared to unimodal baselines, along with meaningful explanation maps that align with clinical insights. The proposed approach not only improves predictive performance but also builds trust by offering interpretable outputs, making it suitable for real-world clinical deployment. This work contributes toward bridging the gap between high-performance AI systems and explainable, trustworthy healthcare solutions. |
| Keywords | Multimodal Deep Learning, Explainable Artificial Intelligence (XAI), Early Disease Diagnosis, Medical Data Fusion, Clinical Decision Support Systems |
| Field | Engineering |
| Published In | Volume 8, Issue 3, May-June 2026 |
| Published On | 2026-05-09 |
| DOI | https://doi.org/10.36948/ijfmr.2026.v08i03.77789 |
Share this

E-ISSN 2582-2160
CrossRef DOI is assigned to each research paper published in our journal.
IJFMR DOI prefix is
10.36948/ijfmr
Downloads
All research papers published on this website are licensed under Creative Commons Attribution-ShareAlike 4.0 International License, and all rights belong to their respective authors/researchers.
Powered by Sky Research Publication and Journals