Explainable Artificial Intelligence In High-Stakes Decision-Making: A Systematic Review Of Methods, Applications, And Challenges
Keywords:
Explainable Artificial Intelligence, High-Stakes Decision-Making, Interpretability, Transparency, Trustworthy AI, LIME, SHAP, Fairness, Responsible AI, Systematic ReviewAbstract
Explainable Artificial Intelligence (XAI) has become a vital area of research as we see more complex machine learning models being used in critical decision-making fields like healthcare, finance, criminal justice, and public policy. While these advanced AI systems often deliver impressive predictive accuracy, their black-box nature raises important questions about transparency, accountability, fairness, and trust—especially when their decisions can significantly affect people's lives and societal outcomes. This systematic review dives into the current state of explainable AI methods, their practical applications, and the hurdles we face when trying to implement XAI in crucial settings. The review brings together existing research on key explainability strategies, such as intrinsic interpretable models, post-hoc explanation techniques like LIME and SHAP, saliency-based visualization methods, and the newer counterfactual and causal explanation frameworks. Additionally, the paper showcases how XAI enhances decision support systems by boosting interpretability, aiding in bias detection, and building user trust in sensitive applications. It pays special attention to how these methods are applied in specific sectors, including clinical diagnosis, credit scoring, fraud detection, and risk assessment tools in the legal system. Despite the progress made, several challenges still need to be addressed, such as balancing model accuracy with interpretability, the subjective nature of evaluating explanations, scalability issues in real-time systems, and the potential for misleading or adversarial explanations. The review also touches on ethical, legal, and regulatory aspects, stressing the importance of standardized evaluation metrics and governance frameworks to ensure responsible AI deployment. This systematic review highlights just how crucial explainable AI is for building decision-making systems that are trustworthy, transparent, and ethically sound. It also points out future research paths that could lead to more robust, human-centered, and domain-aware solutions for explainability.
References
Amodei, D., Olah, C., Steinhardt, J., Christiano, P., Schulman, J., & Mané, D. (2016). Concrete problems in AI safety. arXiv preprint arXiv:1606.06565.
Angwin, J., Larson, J., Mattu, S., & Kirchner, L. (2016). Machine bias. ProPublica. https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing
Arrieta, A. B., Díaz-Rodríguez, N., Del Ser, J., Bennetot, A., Tabik, S., Barbado, A., … Herrera, F. (2020). Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Information Fusion, 58, 82–115. https://doi.org/10.1016/j.inffus.2019.12.012
Barboza, F., Kimura, H., & Altman, E. (2017). Machine learning models and bankruptcy prediction. Expert Systems with Applications, 83, 405–417. https://doi.org/10.1016/j.eswa.2017.04.006
Caruana, R., Lou, Y., Gehrke, J., Koch, P., Sturm, M., & Elhadad, N. (2015). Intelligible models for healthcare: Predicting pneumonia risk and hospital readmission. Proceedings of the 21st ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 1721–1730. https://doi.org/10.1145/2783258.2788613
Doshi-Velez, F., & Kim, B. (2017). Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608.
European Union. (2016). General Data Protection Regulation (GDPR). Official Journal of the European Union.
Feng, X., Cai, Z., Li, X., & Li, Y. (2021). Explainable AI in finance: Applications and challenges. Journal of Risk and Financial Management, 14(9), 430. https://doi.org/10.3390/jrfm14090430
Guidotti, R., Monreale, A., Ruggieri, S., Turini, F., Giannotti, F., & Pedreschi, D. (2018). A survey of methods for explaining black box models. ACM Computing Surveys, 51(5), 1–42. https://doi.org/10.1145/3236009
Lipton, Z. C. (2018). The mythos of model interpretability. Communications of the ACM, 61(10), 36–43. https://doi.org/10.1145/3233231
Lundberg, S. M., & Lee, S.-I. (2017). A unified approach to interpreting model predictions. Advances in Neural Information Processing Systems, 30, 4765–4774.
Molnar, C. (2022). Interpretable machine learning (2nd ed.). Lulu Press.
National Institute of Standards and Technology. (2023). Artificial Intelligence Risk Management Framework (AI RMF 1.0). U.S. Department of Commerce.
Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). “Why should I trust you?” Explaining the predictions of any classifier. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 1135–1144. https://doi.org/10.1145/2939672.2939778
Rudin, C. (2019). Stop explaining black box machine learning models for high-stakes decisions and use interpretable models instead. Nature Machine Intelligence, 1, 206–215. https://doi.org/10.1038/s42256-019-0048-x
Simonyan, K., Vedaldi, A., & Zisserman, A. (2014). Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv preprint arXiv:1312.6034.
Slack, D., Hilgard, S., Jia, E., Singh, S., & Lakkaraju, H. (2020). Fooling LIME and SHAP: Adversarial attacks on post-hoc explanation methods. Proceedings of the AAAI Conference on Artificial Intelligence, 34(04), 1801–1808. https://doi.org/10.1609/aaai.v34i04.6046
Tonekaboni, S., Joshi, S., McCradden, M. D., & Goldenberg, A. (2019). What clinicians want: Contextualizing explainable machine learning for clinical end use. arXiv preprint arXiv:1905.05134.
Downloads
How to Cite
Issue
Section
License
Copyright (c) 2024 International Journal of Engineering Science & Humanities

This work is licensed under a Creative Commons Attribution 4.0 International License.


