
International Journal For Multidisciplinary Research
E-ISSN: 2582-2160
•
Impact Factor: 9.24
A Widely Indexed Open Access Peer Reviewed Multidisciplinary Bi-monthly Scholarly International Journal
Home
Research Paper
Submit Research Paper
Publication Guidelines
Publication Charges
Upload Documents
Track Status / Pay Fees / Download Publication Certi.
Editors & Reviewers
View All
Join as a Reviewer
Get Membership Certificate
Current Issue
Publication Archive
Conference
Publishing Conf. with IJFMR
Upcoming Conference(s) ↓
WSMCDD-2025
GSMCDD-2025
AIMAR-2025
Conferences Published ↓
ICCE (2025)
RBS:RH-COVID-19 (2023)
ICMRS'23
PIPRDA-2023
Contact Us
Plagiarism is checked by the leading plagiarism checker
Call for Paper
Volume 7 Issue 4
July-August 2025
Indexing Partners



















Trustworthy AI for Data Governance: Explaining Compliance Decisions Using SHAP and Causal Inference
Author(s) | Sai Kishore Chintakindhi |
---|---|
Country | United States |
Abstract | Integrating new technologies into data governance calls for a deep look at how compliance, transparency, and user trust all work together in AI systems. Recent progress stresses that we need to clarify why AI makes the decisions it does, especially when it comes to following complex legal and ethical rules. This study uses SHAP (SHapley Additive exPlanations), a method that comes from game theory and helps us understand model predictions by figuring out how much each feature adds to the outcome. By using SHAP along with causal inference, the goal is to make AI's decision-making clearer. It also highlights the rules set by things like the General Data Protection Regulation (GDPR) [1]. These kinds of rules say that automated decisions that affect people need to be clear, which gives us a standard for judging how well we can understand AI [2]. It’s really important to build trust in AI systems because they’re being used more and more in situations where compliance decisions have big effects. Explainability and accountability together are key to building trust with everyone involved. AI systems should not only follow data governance rules but also give explanations that make sense to users and regulators [3]. Causal inference is very important here, helping us look at not just correlations but the potential links between variables that influence decisions. This makes AI outputs easier to understand [4]. For example, using causal diagrams helps us see how specific features directly affect compliance results, which makes AI applications more trustworthy [5]. Moving from just looking at data statistically to thinking about it causally makes the conversation about AI’s reliability better and makes compliance stronger. This research leverages the SHAPs ability to provide localized explanations, to show how different features contribute to model predictions in compliance situations. This gives stakeholders insights that they can act on [6]. This effect of understandability and accountability working together is important for organizations that want to follow the rules while also building trust with their data subjects. SHAP and causal analyses together are a key method for reaching these goals, offering a way to include explainable AI models in regular governance practices [7][8]. The findings ultimately highlight how crucial it is to shift how we govern AI. We need transparent systems that prioritize involving stakeholders and following regulations. As organizations handle different datasets and compliance needs, the suggested framework aims to give a clear way to use AI technologies responsibly and ethically [9]. The study showcases how explainability and compliance can be smoothly combined in AI applications, demonstrated in the flowchart that maps out the data governance framework. This advances not just the theory of data governance but also has practical impacts for different industries. It prepares them to handle the complexities of using AI in areas where compliance is key [11][12]. By creating an environment that values trust and accountability, the study hopes to set new standards for responsible AI. This ensures that data governance changes along with technology and what society expects [13]. |
Field | Engineering |
Published In | Volume 7, Issue 4, July-August 2025 |
Published On | 2025-07-19 |
DOI | https://doi.org/10.36948/ijfmr.2025.v07i04.51762 |
Short DOI | https://doi.org/g9wnrz |
Share this

E-ISSN 2582-2160

CrossRef DOI is assigned to each research paper published in our journal.
IJFMR DOI prefix is
10.36948/ijfmr
Downloads
All research papers published on this website are licensed under Creative Commons Attribution-ShareAlike 4.0 International License, and all rights belong to their respective authors/researchers.
