Abstract
Challenges related to causal learning remain a major issue for artificial reasoning systems. Similar to other ML approaches, robust and trustworthy explainability is needed to support the underlying tasks. This paper aims to provide a novel perspective on causal explainability, creating a model which extracts quantitative causal knowledge and relationships from observational data via Average treatment effect (ATE) estimation to generate robust explanations through comparison and validation of the ranked causally relevant features with results from correlation-based feature relevance explanations. Average treatment effect estimation is calculated to provide a quantitative comparison of the causal features to the relevant features from Explainable AI (XAI). This approach provides a comprehensive method to generate explanations via validations from both causality and XAI to ensure trustworthiness, fairness, and bias detection from both within the data, as well as the AI/ML models themselves for artificial reasoning systems.
This work was supported in part by the DoD Center of Excellence in AI and Machine Learning (CoE-AIML) at Howard University under Contract Number W911NF-20-2-0277 with the U.S. Army Research Laboratory.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Beckers, S.: Causal explanations and xai. ar**v preprint ar**v:2201.13169 (2022)
Busuioc, M.: Accountable artificial intelligence: holding algorithms to account. Public Adm. Rev. 81(5), 825–836 (2021)
Chen, H., Harinen, T., Lee, J.Y., Yung, M., Zhao, Z.: Causalml: python package for causal machine learning. ar**v preprint ar**v:2002.11631 (2020)
Chou, Y.L., Moreira, C., Bruza, P., Ouyang, C., Jorge, J.: Counterfactuals and causability in explainable artificial intelligence: theory, algorithms, and applications. Inf. Fusion 81, 59–83 (2022)
Cui, P., Athey, S.: Stable learning establishes some common ground between causal inference and machine learning. Nature Mach. Intell. 4(2), 110–115 (2022)
Frye, C., Rowat, C., Feige, I.: Asymmetric shapley values: incorporating causal knowledge into model-agnostic explainability. Adv. Neural. Inf. Process. Syst. 33, 1229–1239 (2020)
Gelman, A.: Causality and statistical learning (2011)
Gunning, D., Aha, D.: Darpa’s explainable artificial intelligence (xai) program. AI Mag. 40(2), 44–58 (2019)
Guo, R., Cheng, L., Li, J., Hahn, P.R., Liu, H.: A survey of learning causality with data: problems and methods. ACM Comput. Surv. (CSUR) 53(4), 1–37 (2020)
Janzing, D., Minorics, L., Blöbaum, P.: Feature relevance quantification in explainable ai: a causal problem. In: International Conference on Artificial Intelligence and Statistics, pp. 2907–2916. PMLR (2020)
Lundberg, S.M., Lee, S.I.: A unified approach to interpreting model predictions. Advances in Neural Information Processing Systems 30 (2017)
Moraffah, R., Karami, M., Guo, R., Raglin, A., Liu, H.: Causal interpretability for machine learning-problems, methods and evaluation. ACM SIGKDD Explorations Newsl 22(1), 18–33 (2020)
Pearl, J.: Causal inference in statistics: an overview. Stat. Surv. 3, 96–146 (2009)
Pearl, J.: Theoretical impediments to machine learning with seven sparks from the causal revolution. ar**v preprint ar**v:1801.04016 (2018)
Pearl, J.: The seven tools of causal inference, with reflections on machine learning. Commun. ACM 62(3), 54–60 (2019)
Peters, J., Janzing, D., Schölkopf, B.: Elements of causal inference: foundations and learning algorithms (2017)
Rawal, A., Mccoy, J., Rawat, D.B., Sadler, B., Amant, R.: Recent advances in trustworthy explainable artificial intelligence: status, challenges and perspectives. IEEE Trans. Artif. Intell. 1(01), 1–1 (2021)
Ribeiro, M.T., Singh, S., Guestrin, C.: “Why should i trust you?” explaining the predictions of any classifier. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1135–1144 (2016)
Ribeiro, M.T., Singh, S., Guestrin, C.: Nothing else matters: model-agnostic explanations by identifying prediction invariance. ar**v preprint ar**v:1611.05817 (2016)
Sharma, A., Kiciman, E.: Dowhy: an end-to-end library for causal inference. ar**v preprint ar**v:2011.04216 (2020)
Shrikumar, A., Greenside, P., Kundaje, A.: Learning important features through propagating activation differences. In: Precup, D., Teh, Y.W. (eds.) Proceedings of the 34th International Conference on Machine Learning. Proceedings of Machine Learning Research, vol. 70, pp. 3145–3153. PMLR (06–11 Aug 2017)
Smith, S.C., Ramamoorthy, S.: Counterfactual explanation and causal inference in service of robustness in robot control. In: 2020 Joint IEEE 10th International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob), pp. 1–8. IEEE (2020)
Syrgkanis, V., et al.: Causal inference and machine learning in practice with econml and causalml: Industrial use cases at microsoft, tripadvisor, uber. In: Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pp. 4072–4073 (2021)
Wang, H.X., Fratiglioni, L., Frisoni, G.B., Viitanen, M., Winblad, B.: Smoking and the occurence of Alzheimer’s disease: cross-sectional and longitudinal data in a population-based study. Am. J. Epidemiol. 149(7), 640–644 (1999)
**an, Y., Fu, Z., Muthukrishnan, S., De Melo, G., Zhang, Y.: Reinforcement knowledge graph reasoning for explainable recommendation. In: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 285–294 (2019)
Xu, S., et al.: Learning causal explanations for recommendation. In: The 1st International Workshop on Causality in Search and Recommendation (2021)
Yao, L., Chu, Z., Li, S., Li, Y., Gao, J., Zhang, A.: A survey on causal inference. ACM Trans. Knowl. Discovery Data (TKDD) 15(5), 1–46 (2021)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Rawal, A., McCoy, J., Raglin, A., Rawat, D.B. (2023). A Quantitative Comparison of Causality and Feature Relevance via Explainable AI (XAI) for Robust, and Trustworthy Artificial Reasoning Systems. In: Degen, H., Ntoa, S. (eds) Artificial Intelligence in HCI. HCII 2023. Lecture Notes in Computer Science(), vol 14050. Springer, Cham. https://doi.org/10.1007/978-3-031-35891-3_17
Download citation
DOI: https://doi.org/10.1007/978-3-031-35891-3_17
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-35890-6
Online ISBN: 978-3-031-35891-3
eBook Packages: Computer ScienceComputer Science (R0)