DOI : 10.17577/IJERTCONV14IS020082- Open Access

- Authors : Miss. Sanika Narayan Dhande, Mr. Rahul Abhiman Sonawane
- Paper ID : IJERTCONV14IS020082
- Volume & Issue : Volume 14, Issue 02, NCRTCS – 2026
- Published (First Online) : 21-04-2026
- ISSN (Online) : 2278-0181
- Publisher Name : IJERT
- License:
This work is licensed under a Creative Commons Attribution 4.0 International License
Explainable Artificial Intelligence for Trustworthy Decision-Making in Critical Domains
Miss. Sanika Narayan Dhande, Student, SYMSc(CA)
Department of Computer Science
Ashoka Center for Business and Computer Studies, Chandsi Nashik, India
Mr. Rahul Abhiman Sonawane, Assistant Professor
Department of Computer Science
Ashoka Center for Business and Computer Studies, Chandsi Nashik, India
Abstract – The adoption of Artificial Intelligence (AI) and Machine Learning (ML) systems in such vital areas of life as healthcare, finance, and human resource management has enhanced the efficiency and accuracy of decision-making significantly. Most high-performing AI models are however black boxes which provide minimal transparency on the manner in which predictions are made [1]. This uninterpretability is highly questionable in terms of trusting, accountability, fairness and ethical adherence, particularly when the decisions carry the direct effects on the humankind. The lack of explainability can tend to limit the embrace of AI systems by stakeholders and regulators.
In this study, the problem of model opaqueness is discussed based on the research of Explainable Artificial Intelligence (XAI) methods to improve transparency and trust in AI systems used to make decisions. The experiment uses and tests the most popular XAI techniques including SHAP (Shapley Additive explanations)
-
and LIME (Local Interpretable Model-agnostic Explanations)
-
on machine learning models used on a critical dataset of decisions. These methods are employed to understand which features are important, clarify each individual prediction and give human comprehensible insights into the behaviour of a model.
The experimental evidence confirms that the technique of integrating the explainability techniques does not have a significant negative impact on the predictive performance, but still has a significant positive impact on the model interpretability. The suggested strategy empowers stakeholders to make informed, verified, and reliable decisions made by AI so as to embrace AI responsibility. The results of this project demonstrate the relevance of explainable models to the implementation of AI systems in sensitive and high-risk settings and future plans on how to combine accuracy, fairness, and transparency in intelligent systems.
Keywords – Explainable AI,Machine learning, Model Interpretability, Trustworthy AI, Critical Decision Systems
-
INTRODUCTION
-
Background
Artificial Intelligence (AI) systems are also being implemented in high stakes settings like medical diagnosis, credit scoring, criminal justice and recruitment. These systems are guaranteed to be efficient, more precise, and scalable. Nevertheless, their extensive use has also been accompanied by the increasing concerns as to their transparency. The predictions of complex models such as deep neural networks and ensemble methods can be regarded as black boxes that are hard to comprehend regarding the justification of the decisions
they take [4]. Such non-transparency can result in lack of trust in end-users, regulatory issues, ethical dilemmas, especially in areas where the decision-made is related to the well-being and rights of humans. The concept of Explainable Artificial Intelligence (XAI) has come to be the focus of important research in order to make AI systems more readable and responsible [5]. XAI techniques can close the performance-to- understandable gap in model decision-making processes to promote a feeling of trust and achieve compliance with legal and ethical conventions like the GDPR regarding the right to explanation [6].
-
Problem Statement
Although AI performance has improved, model interpretability is a major challenge towards their application in sensitive areas. Clinicians, loan officers and regulatory bodies are the main stakeholders who are reluctant to trust the use of AI systems whose decision-making process cannot be elucidated. This issue becomes even worse in the context where the biases, mistakes, or unjust results may be critical. The existing XAI techniques are different in terms of applicability, fidelity, and usability and therefore demand a systematic analysis of their effectiveness in actual practice.
-
Research Objectives
This study aims to:
-
Research how XAI can help to increase the transparency and trust of the AI-driven decision systems.
-
SHAP and LIME Implement and test on healthcare and financial benchmark datasets.
-
Determine the trade off between model interpretability and accuracy.
-
Offer practical recommendations on the implementation of the explainable AI in accordance with ethical and regulatory standards.
-
-
Scope and Limitations
This paper is based on post-hoc explainability methods of supervised classification models. The data sets are obtained in healthcare and finance as publicly available. Limitations in the research are the access to labeled data and generalizability of results in other domains. Future research can be expanded to real time explainability and unsupervised learning conditions.
-
-
LITERATURE REVIEW
-
Theoretical Foundations
The explainability requirement of AI is based on the ethical principles of AI, human-centered design, and regulations. Theories like Trust in Automation and Algorithmic Accountability state that users should know and justify automated decisions to trust them. In the same vein, the GDPR requiring model interpretability has rendered model interpretability legally binding in numerous jurisdictions such as the Right to Explanation.
-
Previous Research
The different XAI methods have been examined in the past:
-
LIME (Ribeiro et al., 2016) is a technique that comes up with local explanations through approximating complex models using interpretable ones [3].
-
Games theory is applied in SHAP (Lundberg and Lee, 2017) to assign the significance of features in a regular way [2].
-
In healthcare (e.g., Caruana et al., 2015), it has been demonstrated that the explainable models can enable clinicians to trust AI-assisted diagnoses [7].
-
Research also demonstrates that explainable credit scoring models enhance customer satisfaction and regulation in the financial industry [8].
-
Nevertheless, there are spaces of evaluation of the practical utility of these approaches in various critical areas and the perception of what effects they have on user confidence and decision-making [9].
-
-
Gaps in Current Research
-
Absence of comparative literature on local and global explainability methods [10].
-
Lack of emphasis on the implementation of XAI in the current decision-making processes.
-
Very little studies have investigated the longitudinal effect of explainability on user trust and adoption [11].
-
-
-
METHODOLOGY
-
Research Design
The research is based on the experimental design, which concurrently incorporates measurement of quantitative performance of the model and qualitative measure of the explainability. The study is carried out in three stages:
-
Training and validation of model training.
-
SHAP and LIME techniques of XAI.
-
Comparison on basis of accuracy measures and interpretability measures.
-
-
Datasets
There are two publicly available datasets that are utilised:
-
Healthcare Heart Disease UCI Dataset.
-
German Credit Data (Finance).
They are both binary classification using features pertaining to critical decision-making.
-
-
Models and Tools
-
Deep Neural Network, Gradient Boosting, Random Forest.
-
XAI Tools SHAP (python library), LIME (python library)
-
Measures of Evaluation: Accuracy, Precision, Recall, F1-Score, AUC-ROC; Explanation Fidelity, Consistency and User Comprehension Scores.
-
-
Implementation Steps
-
Pretraining and training models.
-
Make explanations based on SHAP (global and local) and LIME (local).
-
Comparison of model action with and without explanation.
-
The survey of stakeholders (N=20) to determine the clarity and usefulness of the explanations is conducted.
-
-
-
SYSTEM DESIGN / CONCEPTUAL FRAMEWORK
-
Framework Overview
It suggests a conceptual framework of integrating XAI as a critical decision system, which includes:
-
Data Layer- Processed, domain specific data.
-
Model Layer – ML models of high performance.
-
Explainability Layer Rules that explain the model.
-
interface Layer- Stakeholder-friendly dashboards.
-
Validation Layer- Networking with moral and regulatory regulations.
-
-
Component Integration
The framework makes sure that real-time explanations are created, displayed on user-friendly formats (visualizations, textual summaries), and registered to be audited. This leads to transparency, eases debugging and builds user trust.
-
-
EXPERIMENTAL RESULTS AND ANALYSIS.
-
Model Performance
Model
Accuracy (%)
F1-Score
AUC-ROC
Random Forest
91.2
0.89
0.94
Gradient Boosting
89.7
0.87
0.92
Deep Neural Network
88.5
0.86
0.91
Table 1: Predictive results on Heart Disease data.
-
Explainability Evaluation
XAI
Method
Fidelity Score
Consistency
User Comprehension (Avg. Score
/10)
SHAP
0.92
High
8.5
LIME
0.88
Medium
7.8
Table 2: Measures of explainability.
-
Stakeholder Feedback
Findings of the surveys showed that:
-
Shareholders rated SHAP summaries 85 percent useful in grasping the general model behavior.
-
Three-quarters (78%) found MIKE more favorable in case-specific explanations.
-
90% of them noted that they had more trust in AI decisions when they were explained.
-
-
-
PRIVACY, SECURITY, AND ETHICAL CONSIDERATIONS
-
Data Privacy
All data sets were anonymized and utilized according to the regulations of data protection. The explanations were made in such a way that there would not be leakage of sensitive information.
-
Ethical Implications
The paper highlights the significance of:
-
Not biased explanations [12].
-
Promoting equity in model interpretations [13].
-
Giving straightforward explanations to the non-expert users.
-
-
Regulatory Alignment
The suggested XAI integration is in line with GDPR, HIPAA, and future AI ethics standards and advocates accountability and transparency [6,16].
-
-
DISCUSSION AND CONCLUSION
-
Interpretation of Findings
The findings affirm that XAI methods can be highly effective to improve the interpretability of models without affecting the precision of the results [14]. SHAP was more consistent in global insights and LIME was more local, instance-level. Explainability provided more trust between stakeholders and made it easier to validate models.
-
Comparison with Existing Work
This research builds upon the work of other researchers by:
-
Providing a comparative study of SHAP and LIME on important areas [15].
-
Suggesting a deployable explainable AI system framework.
-
Adding user-feedback to determine usefulness.
-
-
Conclusion
Explainable AI is not a technical improvement but a necessity of reliable AI in areas of concern [1,5]. This study proves that approaches such as SHAP and LIME have the potential to reduce the distance between complex models and human cognition and make AI responsible.
Future work should focus on:
-
Explainability in a dynamic environment in real-time.
-
Visual explanations (explanations through text, descriptions, visual, interactive).
-
Explainability assessment measures which are standardized [17].
-
-
-
ACKNOWLEDGMENTS
-
The author appreciates the open-source communities of SHAP and LIME, their tools being of great help. No external funding was accorded to this research.
REFERENCES
-
Adadi, A., & Berrada, M. (2018). Peeking inside the black-box: A survey on explainable artificial intelligence (XAI). IEEE Access, 6, 52138-52160.
-
Arrieta, A. B.-R. (2020). Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Information Fusion, 58, 82115.
-
Barocas, S., Hardt, M., & Narayanan, A. (2019). Fairness and Machine Learning. Cambridge, MA: MIT Press.
-
Caruana, R. L. (2015). Intelligible models for healthcare: Predicting pneumonia risk and hospital 30-day readmission. 21st ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (pp. 17211730). New York, NY: ACM.
-
Doshi-Velez, F. &. (2017). Towards a rigorous science of interpretable machine learning. arXiv.
-
European Commission. (2019). Ethics Guidelines for Trustworthy AI. Brussels, Belgium: European Commission.
-
Goodman, B. &. (2017). European Union regulations on algorithmic decision-making and a "right to explanation". AI Magazine, 38(3), 5057.
-
Gunning, D. &. (2019). DARPAs explainable artificial
intelligence (XAI) program. AI Magazine, 40(2), 4458.
-
Hall, P., Gill, N., & Kurka, M. (2019). Machine learning interpretability with H2O driverless AI. Mountain View, CA: H2O.ai.
-
Lakkaraju, H., Kamar, E., Caruana, R., & Leskovec, J. (2019). Faithful and customizable explanations of black box models. AAAI/ACM Conference on AI, Ethics, and Society (pp. 131-138). Honolulu, HI: Association for Computing Machinery.
-
Lundberg, S. M. (2017). A unified approach to interpreting model predictions. Advances in Neural Information Processing Systems
30 (NeurIPS) (pp. 47654774). San Diego, CA: Neural Information Processing Systems Foundation.
-
Molnar, C. (2022). Interpretable Machine Learning: A Guide for Making Black Box Models Explainable (2nd ed.). Munich, Germany: Self-published / Online.
-
Ribeiro, M. T. (2016). Why should I trust you?" Explaining the predictions of any classifier. 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (pp. 1135 1144). San Francisco, CA: Association fo Computing Machinery (ACM).
-
Rudin, C. (2019). Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence, 1, 206-215.
-
Samek, W. W. (2017). Explainable artificial intelligence: Understanding, visualizing and interpreting deep learning models. arXiv.
-
Stepin, I., Alonso, J. M., Catala, A., & Pereira-Fariña, M. (2021). A survey of contrastive and counterfactual explanation generation methods for explainable artificial intelligence. IEEE Access, 9, 11974-12001.
-
Zhang, J., Liao, Y., & Wang, S. (2023). A survey on evaluation of explainable AI. ACM Computing Surveys, 55(9), 1-38.
