Abstract
In recent years, artificial intelligence (AI) technology has been used in most if not all domains and has greatly benefited our lives. While AI can accurately extract critical features and valuable information from large amounts of data to help people complete tasks faster, there are growing concerns about the non-transparency of AI in the decision-making process. The emergence of explainable AI (XAI) has allowed humans to better understand and control AI systems, which is motivated to provide transparent explanations for the decisions made by AI. This article aims to present a comprehensive overview of recent research on XAI approaches from three well-defined taxonomies. We offer an in-depth analysis and summary of the status and prospects of XAI applications in several key areas where reliable explanations are urgently needed to avoid mistakes in decision-making. We conclude by discussing XAI’s limitations and future research directions.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Deep learning has been contributing to artificial intelligence (AI) systems to speed up and improve numerous tasks, including decision-making, predictions, identifying anomalies and patterns, and even recommendations and so on. Although the accuracy of deep learning models has dramatically improved during the last decade, this improved accuracy has often been achieved through increased model complexity, which may induce common sense mistakes in practice without providing any reasons for the mistakes, making it impossible to fully trust its decisions. It’s also challenging to achieve targeted model improvement and optimisation [1]. Without reliable explanations that accurately represent the current AI system processes, humans still consider AI untrustworthy due to a variety of dynamics and uncertainties [2] when deploying AI applications in real-world environments. This motivates the inherent need and expectation from human users that AI systems should be explainable to help confirm decisions.
Explainable AI (explainable artificial intelligence (XAI)) is often considered a set of processes and methods that are used to describe deep learning models, by characterizing model accuracy, transparency, and outcomes in AI systems [3]. XAI methods aim to provide human-readable explanations to help users comprehend and trust the outputs created by deep learning algorithms. Additionally, some regulations such as European General Data Protection Regulation (general data protection regulation (GDPR))[4] have been introduced to drive further XAI research, demanding the important ethics [5], justifications [6], trust [15]. In other domains, such as medical image or signal recognition, where accuracy is paramount, the focus may be more on predictive power than interpretability [16].
The current XAI methods exhibit various dimensions and descriptions to understand deep learning models and some survey papers [3, 17, 18] have summarized the methods and basic differences among different XAI approaches. However, the state-of-the-art analysis with respect to existing approaches and limitations for different XAI-enabled application domains still lacks investigation.
The field of explainable artificial intelligence (XAI) has witnessed the emergence of numerous methods and techniques aimed at comprehending the intricate workings of deep learning models. Currently, some survey papers have made efforts to summarize these methods and offer a fundamental understanding of the distinctions among various XAI approaches [3, 17, 18]. However, while certain survey papers have focused on specific domains like healthcare [19] or medical applications [20], there still exists a substantial gap in the state-of-the-art analysis pertaining to the existing approaches and their limitations across all XAI-enabled application domains. This gap necessitates a comprehensive investigation encompassing various aspects such as different requirements, suitable XAI approaches, and domain-specific limitations. Conducting such an analysis is crucial as it allows us to gain a deeper understanding of the performance of XAI techniques in real-world scenarios. Additionally, it helps us identify the challenges and opportunities that arise when applying these approaches to different application domains. By bridging this gap, we can make significant strides towards develo** more effective and reliable XAI systems tailored to specific domains and their unique characteristics.
In this survey, our primary objective is to provide a comprehensive overview of explainable artificial intelligence (XAI) approaches across various application domains by exploring and analysing the different methods and techniques employed in XAI and their application-specific considerations. We achieve this by utilizing three well-defined taxonomies, as depicted in Fig. 1. Unlike many existing surveys that solely focus on reviewing and comparing methods, we go beyond that by providing domain map**. This map** provides insights into how XAI methods are interconnected and utilized across various application domains, and even in cases where domains intersect. Additionally, we delve into a detailed discussion on the limitations of the existing methods, acknowledging the areas where further improvements are necessary. Lastly, we summarize the future directions in XAI research, highlighting potential avenues for advancements and breakthroughs. Our contributions in this survey can be summarized as follows:
-
Develop a new taxonomy for the description of XAI approaches based on three well-defined orientations with a wider range of explanation options;
-
Investigate and examine various XAI-enabled applications to identify the available XAI techniques and domain insights through case studies;
-
Discuss the limitations and gaps in the design of XAI methods for the future directions of research and development.
In order to comprehensively analyze XAI approaches, limitations, and future directions from application perspectives, our survey is structured around two main themes, as depicted in Fig. 1. The first theme focuses on general approaches and limitations in XAI, while the second theme aims to analyze the available XAI approaches and domain-specific insights.
Under each domain, we explore four main sub-themes: problem definition, available XAI approaches, case studies, and domain insights. Before delving into each application domain, it is important to review the general taxonomies of XAI approaches. This provides a foundation for understanding and categorizing the various XAI techniques. In each domain, we discuss the available and suitable XAI approaches that align with the proposed general taxonomies of XAI approaches. Additionally, we examine the domain-specific limitations and considerations, taking into account the unique challenges and requirements of each application area. We also explore cross-disciplinary techniques that contribute to XAI innovations. The findings from these discussions are summarized as limitations and future directions, providing valuable insights into current research trends and guiding future studies in the field of XAI.
2 Taxonomies of XAI Approaches
2.1 Review Scope and Execution
This work is mainly based on a scope of review refers to the specific boundaries and focus of the research being conducted. In the context of an XAI survey, the scope typically includes the following aspects:
-
XAI approaches: The review will focus on examining and analyzing different XAI approaches and methods that have been proposed in the literature. This include visualization techniques, symbolic explanations, ante-hoc explanations, post-hoc explanations, local explanations, global explanations and any other relevant techniques.
-
Application domains: The review may consider various application domains where XAI techniques have been applied, including medical and biomedical, healthcare, finance, law, cyber security, education and training, civil engineering. The scope involve exploring the usage of XAI techniques in these domains and analyzing their effectiveness and limitations across multiple domains.
-
Research papers: The review will involve studying and synthesizing research papers that are relevant to the chosen scope. These papers may include original research articles, survey papers and scholarly publications that contribute to the understanding of XAI approaches and their application in the selected domains through case studies.
-
Limitations and challenges: The scope also encompass examining the limitations and challenges of existing XAI methods and approaches. This could involve identifying common issues, gaps in the literature, and areas that require further research or improvement.
Having the scope of review established, the selected databases and a search engine include Scopus, Web of Science and Google Scholar (Search engine) and ar**v between 2013 and 2023. The search terms based on the scopes are:
-
XAI keywords: explainable, XAI, interpretable.
-
Review keywords: survey, review, overview, literature, bibliometric, challenge, prospect, trend, insight, opportunity, future direction.
-
Domain keywords: medical, biomedical, healthcare, wellness, civil, urban, transportation, cyber security, information security, education, training, learning and teaching, coaching, finance, economics, law, legal system.
With the selected search terms, the two-round search strings were designed to effectively retrieve relevant information and narrow down the search results.
The first round, focusing on general research papers, consisted of the following search string: (explainable OR XAI OR interpretable) AND (survey OR review OR overview OR literature OR bibliometric OR challenge OR prospect OR trend OR opportunity OR "future direction").
The second round, aimed at selecting specific application domains, utilized the following search string: (explainable OR XAI OR interpretable) AND (medical, biomedical OR healthcare OR wellness OR civil OR urban OR transportation OR “cyber security” OR “information security” OR education OR training OR “learning and teaching” OR coaching OR finance OR economics OR law OR “legal system”).
Publications that did not clearly align with the scopes based on their title or abstract were excluded from this review. While not all literature explicitly stated this information, the extracted data was organized and served as the foundation for our analysis.
2.2 XAI Approaches
The taxonomies in the existing survey papers generally categorised XAI approaches based on scope (local or global) [21], stage (ante-hoc or post-hoc) [17] and output format (numerical, visual, textual or mixed) [22]. The main difference between the existing study and our survey is that this paper focuses on the human perspective involving source, representation, and logic reasoning. We summarise the taxonomies categorised in this survey in Fig. 2:
Source-oriented (source-oriented (SO)) the sources that support building explanations can be either subjective (S) or objective (O) cognition, depending on whether the explanations are provided based on the fact or human experience. For example, in the medical field, if the explanation of a diagnosis is provided based on the patient’s clinical symptoms and explains the cause and pathology in detail during the AI learning process, this is from objective cognitive concern. In contrast, explanations with subjective cognitive consider patients’ current physical conditions and doctors’ medical knowledge.
Representation-oriented (representation-oriented (RO)) core representation among the XAI approaches can generally be classified into visualisation-based (V), symbolic-based (S) or even hybrid (H) methods. Visual-based methods are the most common representation ways including input visualisation and model visualisation. Input visualisation methods provide an accessible way to view and understand how input data affect model outputs, while model visualisation methods provide analysis based on the aspect of layers or features inside the model.
Besides visualization-based methods, other formats of explanations, including numerical, graphical, rules, and textual explanations, are covered in symbolic-based methods. Symbolic-based methods tend to describe the process of deep learning models by extracting insightful information, such as meaning and context, and representing them in different formats. The coding symbolic-based explanation is provided directly from the factual features, including numerical, graphical and textual explanations. For instance, a numerical method [36] performs explanation by highlighting the important regions in the image, which refers to objective cogitative. Some researchers also consider using subjective sources, such as in [85], authors presented the explanation by combining time series, histopathological images, knowledge databases as well as patient histories.
In terms of representation-oriented, visualisation methods emphasise the visualisation of training data rules and the visualisation inside the model, which is the most popular XAI approaches used in medical image analysis. Some typical examples include attributed-based and perturbation-based methods for model-agnostic explanations as well as CAM-based and concept attribution for model-specific explanations. Locally-interpretable model-agnostic explanations (locally-interpretable model-agnostic explanations (LIME)) [86] is utilised to generate explanations for the classification of medical image patches. Zhu et al. [87] used rule-based segmentation and perturbation-based analysis to generate the explanation for visualising the importance of each feature in the image. The concept attribution [37] is introduced by quantifying the contribution of features of interest to the CNN network’s decision-making. Symbolic methods focus on the symbolic information representations that simulate the doctor’s decision-making process with natural language, along with the generated decision results, such as primary diagnosis reports, etc. For example, Kim et al.[66] introduced concept activation vectors (CAVs), which provided textual interpretation of neural network internal state with user-friendly concepts. Lee et al. [73] provided explainable computer-aided diagnoses by combining a visual pointing map and diagnostic sentences based on the predefined knowledge base.
In terms of logical-oriented, explanations focused on end-end logic reasoning, such as the above-mentioned LIME, perturbation-based methods are utilised to explain the relationship between input medical images and predicted results. For example, a linear regression model is embedded into LIME [86] to identify relevant regions by plotting heat maps with varying color scales. Zhang et al. [3.7 Civil Engineering AI systems used in civil engineering research have a significant impact on the decision-making processes in road transport and power systems. In particular, autonomous driving techniques in road transport and power system analysis and power systems are the common areas used deep learning techniques, such as navigation and path planning, scene recognition, lane and obstacle detection, as well as planning, monitoring, and controlling the power system [150, 184]. In the field of autonomous driving, deep learning techniques are normally utilised to recognize scenes for digital images [184, 185]. While in the field of power system analysis, deep learning techniques are used to extract features from the underlying data for power system management, such as power grid synthesis, state estimation, and photovoltaic (PV) power prediction [150, 186]. Deep learning explainable techniques are used to automatically extract abstract features of images or depth non-linear features of underlying data through end-to-end predictive processing to obtain results, which is not sufficient to provide the evidence to trust and accept the result of autonomous driving and power system management. For example, one can use traffic lights and signal recognition for driving planning, in which the traffic lights at crosswalks and intersections are an essential function in following traffic rules and preventing traffic accidents. Deep learning methods have achieved prominence in traffic sign and light recognition, but they are hard to explain the correlation between inputs and outputs and lack an explanation to support reasoning in driving planning studies [187]. In power system management, deep learning methods may mislead the output explanations of power stability to provide unreliable recommendations, so explanations can increase user trust [150]. XAI can improve the management of autonomous driving and power system, providing an effective interaction to promote smart civil engineering. Deep learning interpretation research in autonomous driving and power systems is a common interpretable deep learning method because it is not only influenced by data, but also relates to expert knowledge and ethical principles. In terms of source-oriented, objective interpretability obtains visible or measurable results from 2D and 3D images or underlying datasets, while subjective interpretability requires consideration of the knowledge from automotive or electrical experts and the ethical standards of their fields. Currently, XAI proposals include objective and subjective cognitive aspects. For example, CAM, as an objective cognition method, is used to explain the highlight of important regions in 2D or 3D images. Time series, 2D images, 3D images, Lidar images, knowledge databases and ethical criteria are utilised as subject sources to explain the model [147, 185, 187]. In terms of representation-oriented, visual interpretation is the highest level semantics to understand which parts of the image impact the model, emphasing on visual structure of data and model, which is the primary XAI method used in autonomous driving. These XAI methods can be divided into gradient-based and back propagation-based. Gradient-based interpretation methods include CAM, and its enhanced variants such as Guided Grad-CAM, Grad-CAM, Grad-CAM++ and Smooth Grad CAM++. CAM can highlight the discriminative regions of a scene image used for scene detection [147]. Backpropagation-based methods contain guided backpropagation, layered relevance propagation, visual backprop and deep lift. Visual Backprop shows which input pixels set contributes to steering self-driving cars [144]. Symbolic interpretation uses understandable language to provide evidence for result recommendations in autonomous driving and power system management. In autonomous driving, proposed AI methods make decisions according to traffic rules. For example, “the traffic light ahead turned red,” thus “the car stopped” [185]. In power system management, it uses the data gathered from occupant actions for resources such as room lighting to forecast patterns of energy resource usage [188]. Hybrid interpretation combines visual interpretation and symbolic interpretation to provide steering determination in autonomous driving. For example, Berkeley Deep Drive-X (BDD-X) is introduced in autonomous driving which includes the description of driving pictures and annotations for textual interpretation [49]. In terms of logical-oriented, the end-end explanations are used to explain the relationship between input images including obstacle and scene images and the prediction. For example, LIME is utilised to explain the relationship between input radar image and prediction results [189]. Middle-end explanations reveal reasons behind the autoencoder-based assessment model and how they can help drivers reach a better understanding and trust in the model and its results. For example, a rule-based local surrogate interpretable method is proposed, namely MuRLoS, which focuses on the interaction between features [149]. Correlation expatriation is used in the risk management of self-driving and power systems. For example, SHAP is used to assess and explain collision risk using real-world driving data for self-driving [190]. Decisive vehicle actions Decisive vehicle actions in autonomous driving are based on multiple tasks, such as scene recognition, obstacle detection, lane recognition, and path planning. It can use attention mechanisms, heat maps, diagnostic models and texture descriptions to recognise obstacles, scenes and lanes and steer the car operation [147, 185, 187]. As mentioned before, CAM is used to highlight the main area for recognition [63]. Visual Backprop, unlike CAM-based, emphases highlighting pixel-level to filter features of scene images [144]. Grad-CAM is combined with existing fine-grained visualisations to provide a high-resolution class-discriminative visualisation [36]. Visual attention heat maps are used to explain the vehicle controller behaviour through segmenting and filtering simpler and more accurate maps while not degrading control accuracy [145]. A neural motion planner uses 3D detection instances with descriptive information for safe driving [146]. An interpretable tree-based representation as hybrid presentations combines rules, actions, and observation to generate multiple explanations for self-driving [147]. An architecture is used for joint scene prediction to explain object-induced actions [149]. An auto-discern system utilises surroundings observations and common-sense reasoning with answers for driving decisions [148]. Power system management Power system management normally consists of stability assessment, emergency control, power quality disturbance, and energy forecasting. CNN classifier, combined with non-intrusive load monitoring (NILM), is utilised to estimate the activation state and provide feedback for the consumer-user [150]. The shape method is firstly used in emergency control for reinforcement learning for grid control (RLGC) under three different outputs analysis [151]. Deep-SHAP is proposed for the under-voltage load shedding of power systems, and it adds feature classification of the inputs and probabilistic analysis of the outputs to increase clarity [152]. In terms of transportation systems, operators, such as drivers and passengers, are the primary end-users in scenarios involving decisive vehicle actions, because they may want to comprehend the reasoning behind the decisions made by the autonomous system. It is very important in high-stake domains which human lives are risk. XAI can provide explanations for AI decisions to enhance the system more transparent and fostering trust. Real-time explanations pose a significant challenge for XAI in decisive vehicle actions, because decisions need to be made with fractions of a second. Rapidly changing environments such as weather conditions, pedestrian movement and other vehicles actions promote XAI should ideally make quick and accurate decisions. Moreover, every driving situation can be unique. XAI needs suitable for diversity situation and adapt its explanations which based on context-aware interoperability. As previously mentioned, XAI demands more computational resources because of real-time explanations based on timely response. Moreover, deceive vehicle actions require high dimensional sensor data, such as the inputs from LiDAR and stereo cameras, which lead the methods, like LIME and SHAP, which adopts approximate local decision boundaries, are expensive for computation and especially for high-dimensional inputs. The requirements in XAI that can generate real-time, informative explanations without overburdening the computational resources of the system. In terms of infrastructure system management, such as power or water system management, general public, including governments and residents, are the key end-users in power system management. Government bodies want to oversee the safe and fair use of AI in power system management. Meanwhile, residents may be curious about the mechanics of AI used to manage power systems in the city. XAI can be used to evaluate AI systems for safety, fairness, transparency, and adherence to regulatory requirements. Interpretation complexity is a primary challenge for XAI in infrastructure system management due to the multidimensional nature of the data, which includes factors from power generators, transmission lines, and power consumers. Moreover, unlike the case of autonomous driving, power system operations demand more technical expertise and need to adhere to various regulatory requirements. Consequently, XAI is not only to provide coherent and insightful interpretations of the system’s operations but also to demonstrate that these operations comply with all relevant regulations. The entire process in infrastructure system management is starting from generation and distribution to monitor consumer usage patterns. The complexity is future amplified by the demands for load balancing and power outages, which influences the public life and the city operation. Moreover, it also need to fix the various regulations and standers. To evidence such compliance, XAI may need to generate more complex or detailed explanations, thus increasing the computational cost. XAI innovations for cross-disciplinary refers to the advancements and developments in explainable AI (XAI) that span multiple domains and disciplines. It involves the integration and adaptation of XAI techniques and methodologies to address complex problems and challenges that arise in diverse fields. One aspect of XAI Innovations for cross-disciplinary is the exploration and utilization of common XAI techniques across different domains. These techniques, such as attention-based models, model-agnostic methods, and rule-based methods, can be applied to various fields to provide transparent and interpretable explanations for AI models. Below are some examples of common XAI techniques: Regression-based partitioned methods:can be applied to any black-box model. For example, LIME approximates the decision boundaries of the model locally and generates explanations by highlighting the features that contribute most to the prediction for a specific instance. LIME can be used in domains such as healthcare, cyber security, finance, or education to provide instance-level interpretability and explainability. SHAP is another common technique based on cooperative game theory, which can be applied to different domains to explain the importance of features in the decision-making process. For example, in medical diagnostics, SHAP can help understand which medical parameters or biomarkers have the most impact on a particular diagnosis. Feature importance: Feature importance techniques assess the relevance and contribution of each feature in the model’s predictions. Methods like permutation importance, Gini importance, or gain-based importance are commonly used. Feature importance can be useful in various domains to identify the factors that drive specific outcomes or decisions. For instance, in finance, feature importance can help understand which financial indicators or market factors play a crucial role in investment decisions. Partial dependence plots: Partial dependence plots visualize the relationship between a feature and the model’s output while holding other features constant. These plots show how changing the value of a specific feature affects the model’s predictions. Partial dependence plots can be employed in domains such as healthcare, where they can provide insights into the impact of certain medical treatments or interventions on patient outcomes. Rule-based models: Rule-based models provide transparent and interpretable decision-making processes by expressing decision rules in the form of “if-then” statements. These models can be used in various domains to generate explanations that are easily understandable by humans. In legal applications, rule-based models can help explain legal reasoning by map** legal principles and regulations to decision rules. These are just a few examples of common XAI techniques that can be applied across different domains. The choice of technique depends on the specific requirements and characteristics of each domain. We summarise some typeical suitable XAI approaches for each domain shown in Table 5. By leveraging these techniques, domain experts and practitioners can gain insights into the inner workings of AI models and make informed decisions based on understandable and interpretable explanations. Another aspect of XAI innovations for cross-disciplinary involves the development of domain-specific XAI approaches. In Table 5, we summarize some typical suitable XAI approaches for different domains. These approaches can be tailored to the unique characteristics and requirements of specific domains, taking into account the specific challenges and complexities of each field. Domain-specific XAI approaches consider various factors, including domain knowledge, regulations, and ethical considerations, to create an XAI framework that is specifically designed for a particular domain. By incorporating domain expertise and contextual information, these approaches provide explanations that are not only interpretable but also relevant and meaningful within their respective domains. By tailoring XAI approaches to specific domains, practitioners can gain deeper insights into the behavior of AI models within the context of their field. This not only enhances transparency and trust in AI systems but also enables domain-specific considerations to be incorporated into the decision-making process, ensuring the explanations are relevant and aligned with the requirements and constraints of each domain. Furthermore, XAI innovations for cross-disciplinary emphasize the importance of collaboration and the integration of expertise from different fields. This approach recognizes that the challenges and complexities of XAI extend beyond individual domains and require a multidisciplinary perspective. Collaboration and integration of expertise enable a holistic approach to XAI, where insights from different disciplines can inform the development of innovative and effective solutions. For example, in the field of healthcare, collaboration between medical practitioners, data scientists, and AI researchers can lead to the development of XAI techniques that not only provide interpretable explanations but also align with medical guidelines and regulations. This integration of expertise ensures that the explanations generated by XAI systems are not only technically sound but also relevant and meaningful in the specific healthcare context. Similarly, in the domain of cybersecurity, collaboration between cybersecurity experts, AI specialists, and legal professionals can lead to the development of XAI techniques that address the unique challenges of cybersecurity threats. By combining knowledge from these different fields, XAI systems can provide interpretable explanations that enhance the understanding of AI-based security measures, assist in identifying vulnerabilities, and facilitate decision-making processes for cybersecurity professionals. The collaboration and integration of expertise from different fields also foster a cross-pollination of ideas and perspectives, driving innovation and the development of novel XAI techniques. By leveraging the diverse knowledge and experiences of experts from various domains, XAI can evolve and adapt to meet the evolving needs and challenges of different industries and societal contexts.3.7.1 Problem Definition
3.7.2 XAI Based Proposals
3.7.3 Cases Studies
3.7.4 Domain-Specific Insights
3.8 Cross-Disciplinary Techniques for XAI Innovations
4 Discussion
As the concerns on explainability and the attentions for XAI, regulations such as GDPR set out the transparency rules about the data processing. As most modern AI systems are data-driven AI, these requirements are actually applicable to all application domains. Not only the explainability is necessary, but also the way of explaining is required.
In this section, we will summarize the limitations of existing XAI approaches based on the above review in each application domain, and identify future research directions.
4.1 Limitations
Adaptive integration and explanation: many existing approaches provide explanations in a generic manner, without considering the diverse backgrounds (culture, context, etc.) and knowledge levels of users. This one-size-fits-all approach can lead to challenges in effective comprehension for both novice and expert users. Novice users may struggle to understand complex technical explanations, while expert users may find oversimplified explanations lacking in depth. These limitations hinder the ability of XAI techniques to cater to users with different levels of expertise and may impact the overall trust and usability of the system. Furthermore, the evaluation and assessment of XAI techniques often prioritize objective metrics, such as fidelity or faithfulness, which measure how well the explanations align with the model’s internal workings. While these metrics are important for evaluating the accuracy of the explanations, they may not capture the subjective aspects of user understanding and interpretation. The perceived quality of explanations can vary among users with different expertise levels, as well as under different situations or conditions.
Interactive explanation: in the current landscape of XAI research, there is recognition that a single explanation may not be sufficient to address all user concerns and questions in decision-making scenarios. As a result, the focus has shifted towards develo** interactive explanations that allow for a dynamic and iterative process. However, there are challenges that need to be addressed in order to effectively implement interactive explanation systems. One of the key challenges is the ability to handle a wide range of user queries and adapt the explanations accordingly. Users may have diverse information needs and may require explanations that go beyond superficial or generic responses. In particular, addressing queries that involve deep domain knowledge or intricate reasoning processes can be complex and requires sophisticated techniques. Another challenge is striking a balance between providing timely responses to user queries and maintaining computational efficiency. Interactive explanation systems need to respond quickly to user interactions to facilitate a smooth and engaging user experience. However, generating accurate and informative explanations within a short response time can be demanding, and trade-offs may need to be made depending on the specific domain and computational resources available. Moreover, the design and implementation of interactive explanation systems should also consider the context and domain-specific requirements. Different domains may have unique challenges and constraints that need to be taken into account when develo** interactive explanations. It is important to ensure that the interactive explanation systems are tailored to the specific domain and can effectively address the needs of users in that context.
Connection and consistency in hybrid explanation: in the context of hybrid explanations in XAI, it is crucial to ensure connection and consistency among different sources of explanations. Hybrid approaches aim to leverage multiple techniques to provide users in various domains with different application purposes, achieving robustness and interpretability. However, it is necessary to address potential conflicts and ensure coordinated integration of different components within these hybrid systems. Currently, many works focus on combining various explanation techniques to complement each other and enhance overall system performance. While this integration is valuable, it is important to acknowledge that different techniques may have inherent differences in their assumptions, methodologies, and outputs. These differences can result in conflicts or inconsistencies when combined within a hybrid explanation system. Therefore, careful attention should be given to the design of complex hybrid explanation systems. The structure and architecture need to be thoughtfully planned to ensure seamless connections between components. This involves identifying potential conflicts early on and develo** strategies to resolve them. Additionally, efforts should be made to establish a unified framework that allows for effective coordination and integration of the different techniques used in the hybrid system. Furthermore, the evaluation and validation of hybrid explanation systems should include assessing the consistency of explanations provided by different sources. This evaluation process helps identify any discrepancies or inconsistencies and guides the refinement of the system to ensure a coherent and unified user experience.
Balancing model interpretability with predictive accuracy: currently, researchers are develo** hybrid approaches that aim to strike a better balance between interpretability and accuracy, such as using post-hoc interpretability techniques with complex models or designing new model architectures that inherently provide both interpretability and high accuracy.However, they also come with their own limitations. Post-hoc interpretability techniques generate explanations after the model has made its predictions, which means they do not directly influence the model’s decision-making process. As a result, the explanations may not capture the full complexity and nuances of the model’s internal workings. Furthermore, post-hoc techniques can be computationally expensive and may not scale well to large datasets or complex models with high-dimensional inputs. Designing new model architectures such as rule-based models or attention mechanisms in neural networks may struggle to capture complex interactions and may require a significant amount of manual rule engineering. It is crucial to recognize that there is no universal solution to the interpretability-accuracy trade-off. The choice of approach depends on the specific requirements of the application, available resources, and acceptable trade-offs in the given context. Researchers and practitioners must carefully consider the limitations and benefits of different techniques to strike an appropriate balance based on their specific use cases.
Long-term usability and maintainability: the current XAI methods face several limitations when deployed in real-world scenarios. One significant limitation is the need for continuous explanation updates. XAI systems generate explanations based on training data, and as the underlying AI models or data evolve, the explanations may become outdated or less accurate. To ensure relevance and usefulness, XAI systems should be designed to incorporate mechanisms for updating explanations to reflect the latest model updates or data changes. Another limitation is the assumption of stationary data distributions. XAI methods are typically trained on historical data, assuming that the future data will follow a similar distribution. However, if the data distribution changes over time, the performance of the XAI system may deteriorate. Adapting XAI methods to handle shifting data distributions is essential for maintaining their effectiveness and ensuring reliable explanations in dynamic environments. Scalability is another crucial consideration, particularly for large-scale AI systems. XAI techniques that work well on small-scale or controlled datasets may face challenges when applied to large-scale AI systems with complex models and massive amounts of data. Efficient algorithms and sufficient computational resources are necessary to handle the increased computational demands of explaining large-scale AI systems without sacrificing performance or usability.
4.2 Future Directions
To address the first limitation, building the context-awareness XAI is important, we need to explore how to generate explanations by considering mission contexts (surrounding environment, situations, time-series datasets.), map** user roles (end-user, domain expert, business manager, AI developer, etc.) and targeted goals (refine the model, debugging system errors, detecting bias, understand AI learning process, etc.) regardless of the type of AI system. So far, most of these studies were still conceptual with limited consideration, the more general context-driven systems and practical implementations will be an important direction for future research.
Secondly, interactive explanations (e.g., conversation system Interfaces, games, using audio, visuals, video, etc.) should be explored further. This is a promising approach to building truly human-centred explanations by identifying users’ requirements and providing better human-AI collaboration. These incorporating theories and frameworks allow an iterative process from humans, which is a crucial aspect of building successful XAI systems.
Finally, the hybrid explanation should be applied by concerning fusing heterogeneous knowledge from different sources, managing time-sensitive data, inconsistency, uncertainty, etc. Among these conditions, hybrid explanation has been an interesting and increasing topic in recent years. This will also involve a wide range of criteria and strategies that target a clear structure and consensus on what constitutes success and trustworthy explanations.
5 Conclusion
This paper addresses a wide range of explainable AI topics. XAI is a rapidly growing field of research, as it fills a gap in current AI approaches, allowing people to better understand AI models and therefore trust their outputs. By summarising the current literature, we have proposed a new taxonomy for XAI from the human perspective. The taxonomy considers source-oriented, representation-oriented aspect, and logic-oriented perspectives.
It is very important that we have elaborated on the applications of XAI in multiple areas in the paper, including medical, healthcare, cybersecurity, finance and law, education and training, and civil engineering. We provide a comprehensive review of different XAI approaches and identify the key techniques for case studies. Finally, we discuss the limitations of existing XAI methods and present several corresponding areas for further research: (1) context-awareness XAI, (2) interactive explanations, and (3) hybrid explanations.
Overall, this paper provides a clear survey of the current XAI research and application status from the human perspective. We hope this article will provide a valuable reference for XAI-related researchers and practitioners. We believe XAI will build a bridge of trust between humans and AI.
Data availability
Not applicable.
Abbreviations
- AI:
-
Artificial intelligence
- XAI:
-
Explainable artificial intelligence
- GDPR:
-
General data protection regulation
- SO:
-
Source-oriented
- RO:
-
Representation-oriented
- CNN:
-
Convolutional neural network
- NNKX:
-
Neural network knowledge extraction
- REFNE:
-
Rule extraction from neural network ensemble
- ERE:
-
Electric rule extraction
- KG:
-
Knowledge graph
- LO:
-
Logic-oriented
- CT:
-
Computed tomography
- MRI:
-
Magnetic resonance imaging
- US:
-
Ultrasound
- Grad-CAM:
-
Gradient-weighted class activation map**
- LIME:
-
Locally-interpretable model-agnostic explanations
- SHAP:
-
Shapley additive explanations
- CMGE:
-
Counterfactual multi-granularity graph supporting fact extraction
- EHR:
-
Electronic health records
- LEMNA:
-
Local explanation method using nonlinear approximation
References
Zhang Y, Tiňo P, Leonardis A, Tang K. A survey on neural network interpretability. IEEE Trans Emerg Top Comput Intell. 2021;20:20.
Tomsett R, Preece A, Braines D, Cerutti F, Chakraborty S, Srivastava M, Pearson G, Kaplan L. Rapid trust calibration through interpretable and uncertainty-aware AI. Patterns. 2020;1(4):100049.
Arrieta AB, Díaz-Rodríguez N, Del Ser J, Bennetot A, Tabik S, Barbado A, García S, Gil-López S, Molina D, Benjamins R, et al. Explainable artificial intelligence (XAI): concepts, taxonomies, opportunities and challenges toward responsible AI. Inf Fusion. 2020;58:82–115.
Regulation GDP. General data protection regulation (GDPR). Intersoft Consult. 2018;24:1.
Bostrom N, Yudkowsky E. The ethics of artificial intelligence. In: Artificial intelligence safety and security. New York: Chapman and Hall; 2018. p. 57–69.
Weld DS, Bansal G. The challenge of crafting intelligible intelligence. Commun ACM. 2019;62(6):70–9.
Das A, Rad P. Opportunities and challenges in explainable artificial intelligence (XAI): a survey. ar**v:2006.11371 (ar**v preprint) (2020).
Challen R, Denny J, Pitt M, Gompels L, Edwards T, Tsaneva-Atanasova K. Artificial intelligence, bias and clinical safety. BMJ Qual Saf. 2019;28(3):231–7.
Patil MS, Främling K. Context, utility and influence of an explanation. ar**v:2303.13552 (ar**v preprint); 2023.
Ooge J, Verbert K. Explaining artificial intelligence with tailored interactive visualisations. In: 27th international conference on intelligent user interfaces; 2022. p. 120–3.
Saeed W, Omlin C. Explainable AI (XAI): a systematic meta-survey of current challenges and future opportunities. Knowl Based Syst. 2023;11:0273.
Förster M, Klier M, Kluge K, Sigler I. Fostering human agency: a process for the design of user-centric XAI systems; 2020.
Kotriwala A, Klöpper B, Dix M, Gopalakrishnan G, Ziobro D, Potschka A. Xai for operations in the process industry-applications, theses, and research directions. In: AAAI spring symposium: combining machine learning with knowledge engineering; 2021.
Albahri A, Duhaim AM, Fadhel MA, Alnoor A, Baqer NS, Alzubaidi L, Albahri O, Alamoodi A, Bai J, Salhi A, et al. A systematic review of trustworthy and explainable artificial intelligence in healthcare: assessment of quality, bias risk, and data fusion. Inf Fusion. 2023;20:20.
Kurshan E, Chen J, Storchan V, Shen H. On the current and emerging challenges of develo** fair and ethical AI solutions in financial services. In: Proceedings of the second ACM international conference on AI in finance; 2021. p. 1–8.
Komorowski P, Baniecki H, Biecek P. Towards evaluating explanations of vision transformers for medical imaging. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition; 2023. p. 3725–3731.
Adadi A, Berrada M. Peeking inside the black-box: a survey on explainable artificial intelligence (XAI). IEEE Access. 2018;6:52138–60.
Minh D, Wang HX, Li YF, Nguyen TN. Explainable artificial intelligence: a comprehensive review. Artif Intell Rev. 2021;20:1–66.
Chaddad A, Peng J, Xu J, Bouridane A. Survey of explainable AI techniques in healthcare. Sensors. 2023;23(2):634.
Tjoa E, Guan C. A survey on explainable artificial intelligence (XAI): toward medical XAI. IEEE Trans Neural Netw Learn Syst. 2020;32(11):4793–813.
Angelov PP, Soares EA, Jiang R, Arnold NI, Atkinson PM. Explainable artificial intelligence: an analytical review. Wiley Interdiscip Rev Data Min Knowl Discov. 2021;11(5):1424.
Vilone G, Longo L. Classification of explainable artificial intelligence methods through their output formats. Mach Learn Knowl Extract. 2021;3(3):615–61.
Alain G, Bengio Y. Understanding intermediate layers using linear classifier probes. ar**v:1610.01644 (ar**v preprint); 2016.
Zhang Q, Cao R, Shi F, Wu YN, Zhu S-C. Interpreting CNN knowledge via an explanatory graph. In: Proceedings of the AAAI conference on artificial intelligence, vol. 32; 2018.
Hendricks LA, Hu R, Darrell T, Akata Z. Grounding visual explanations. In: Proceedings of the European conference on computer vision (ECCV); 2018. p. 264–79.
Bondarenko A, Aleksejeva L, Jumutc V, Borisov A. Classification tree extraction from trained artificial neural networks. Proced Comput Sci. 2017;104:556–63.
Zhou Z-H, Jiang Y, Chen S-F. Extracting symbolic rules from trained neural network ensembles. AI Commun. 2003;16(1):3–15.
Barakat N, Diederich J. Eclectic rule-extraction from support vector machines. Int J Comput Intell. 2005;2(1):59–62.
Nikolov A, d’Aquin M. Uncovering semantic bias in neural network models using a knowledge graph. In: Proceedings of the 29th ACM international conference on information and knowledge management; 2020, p. 1175–84.
Riquelme F, De Goyeneche A, Zhang Y, Niebles JC, Soto A. Explaining VQA predictions using visual grounding and a knowledge base. Image Vis Comput. 2020;101:103968.
Erion G, Janizek JD, Sturmfels P, Lundberg SM, Lee S-I. Learning explainable models using attribution priors; 2019.
Robnik-Šikonja M, Bohanec M. Perturbation-based explanations of prediction models. In: Human and machine learning. Berlin: Springer; 2018. p. 159–75.
Laugel T, Lesot M-J, Marsala C, Renard X, Detyniecki M. The dangers of post-hoc interpretability: Unjustified counterfactual explanations. ar**v:1907.09294 (ar**v preprint); 2019.
Chefer H, Gur S, Wolf L. Transformer interpretability beyond attention visualization. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition; 2021. p. 782–91.
Jalaboi R, Faye F, Orbes-Arteaga M, Jørgensen D, Winther O, Galimzianova A. Dermx: an end-to-end framework for explainable automated dermatological diagnosis. Med Image Anal. 2023;83:102647.
Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-cam: visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE international conference on computer vision; 2017. p. 618–26.
Graziani M, Andrearczyk V, Marchand-Maillet S, Müller H. Concept attribution: explaining CNN decisions to physicians. Comput Biol Med. 2020;123:103865.
Zhang Q, Wu YN, Zhu S-C. Interpretable convolutional neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2018. p. 8827–836.
Liang X, Hu Z, Zhang H, Lin L, **ng EP. Symbolic graph reasoning meets convolutions. Adv Neural Inf Process Syst. 2018;31:25.
Li CY, Liang X, Hu Z, **ng EP. Knowledge-driven encode, retrieve, paraphrase for medical image report generation. In: Proceedings of the AAAI conference on artificial intelligence; 2019. vol. 33, p. 6666–73.
Ribeiro MT, Singh S, Guestrin C. Anchors: High-precision model-agnostic explanations. In: Proceedings of the AAAI conference on artificial intelligence; 2018. vol. 32.
Teng F, Yang W, Chen L, Huang L, Xu Q. Explainable prediction of medical codes with knowledge graphs. Front Bioeng Biotechnol. 2020;8:867.
Sun P, Gu L. Fuzzy knowledge graph system for artificial intelligence-based smart education. J Intell Fuzzy Syst. 2021;40(2):2929–40.
Panchenko A, Ruppert E, Faralli S, Ponzetto SP, Biemann C. Unsupervised does not mean uninterpretable: the case for word sense induction and disambiguatio; 2017. Association for Computational Linguistics.
Bennetot A, Laurent J-L, Chatila R, Díaz-Rodríguez N. Towards explainable neural-symbolic visual reasoning. ar**v:1909.09065 (ar**v preprint); 2019.
Tamagnini P, Krause J, Dasgupta A, Bertini E. Interpreting black-box classifiers using instance-level visual explanations. In: Proceedings of the 2nd workshop on human-in-the-loop data analytics; 2017. p. 1–6.
Spinner T, Schlegel U, Schäfer H, El-Assady M. Explainer: a visual analytics framework for interactive and explainable machine learning. IEEE Trans Visual Comput Graph. 2019;26(1):1064–74.
Hendricks LA, Akata Z, Rohrbach M, Donahue J, Schiele B, Darrell T. Generating visual explanations. In: European conference on computer vision. Springer; 2016. p. 3–19.
Kim J, Rohrbach A, Darrell T, Canny J, Akata Z. Textual explanations for self-driving vehicles. In: Proceedings of the European conference on computer vision (ECCV); 2018. p. 563–78.
Park DH, Hendricks LA, Akata Z, Rohrbach A, Schiele B, Darrell T, Rohrbach M. Multimodal explanations: justifying decisions and pointing to the evidence. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2018. p. 8779–8788.
Xu K, Ba J, Kiros R, Cho K, Courville A, Salakhudinov R, Zemel R, Bengio Y. Show, attend and tell: neural image caption generation with visual attention. In: International conference on machine learning; 2015. p. 2048–2057.
Gu D, Li Y, Jiang F, Wen Z, Liu S, Shi W, Lu G, Zhou C. Vinet: a visually interpretable image diagnosis network. IEEE Trans Multimed. 2020;22(7):1720–9.
Slack D, Hilgard S, Jia E, Singh S, Lakkaraju H. Fooling lime and shap: adversarial attacks on post hoc explanation methods. In: Proceedings of the AAAI/ACM conference on AI, ethics, and society; 2020. p. 180–86.
Zhang Z, Rudra K, Anand A. Explain and predict, and then predict again. In: Proceedings of the 14th ACM international conference on web search and data mining; 2021. p. 418–26.
Montavon G, Binder A, Lapuschkin S, Samek W, Müller K-R. Layer-wise relevance propagation: an overview. Explainable AI: interpreting, explaining and visualizing deep learning; 2019. 193–209.
Zhang Z, Chen P, McGough M, **ng F, Wang C, Bui M, **e Y, Sapkota M, Cui L, Dhillon J, et al. Pathologist-level interpretable whole-slide cancer diagnosis with deep learning. Nat Mach Intell. 2019;1(5):236–45.
Sarlin P-E, DeTone D, Malisiewicz T, Rabinovich A. Superglue: learning feature matching with graph neural networks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition; 2020. p. 4938–47.
Shen S, Han SX, Aberle DR, Bui AA, Hsu W. Explainable hierarchical semantic convolutional neural network for lung cancer diagnosis. In: CVPR workshops; 2019. p. 63–6.
Gozzi N, Malandri L, Mercorio F, Pedrocchi A. Xai for myo-controlled prosthesis: explaining EMG data for hand gesture classification. Knowl-Based Syst. 2022;240:108053.
Simonyan K, Vedaldi A, Zisserman A. Deep inside convolutional networks: visualising image classification models and saliency maps. ar**v:1312.6034 (ar**v preprint); 2013.
Zeiler MD, Fergus R. Visualizing and understanding convolutional networks. In: European conference on computer vision. Springer; 2014. p. 818–33.
Springenberg JT, Dosovitskiy A, Brox T, Riedmiller M. Striving for simplicity: the all convolutional net. ar**v:1412.6806 (ar**v preprint); 2014.
Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A. Learning deep features for discriminative localization. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2016. p. 2921–9.
Olah C, Mordvintsev A, Schubert L. Feature visualization. Distill. 2017;2(11):7.
Zhang Z, **e Y, **ng F, McGough M, Yang L. Mdnet: a semantically and visually interpretable medical image diagnosis network. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2017. p. 6428–36.
Kim B, Wattenberg M, Gilmer J, Cai C, Wexler J, Viegas F, et al. Interpretability beyond feature attribution: quantitative testing with concept activation vectors (TCAV). In: International conference on machine learning; 2018. p. 2668–77.
Wu B, Zhou Z, Wang J, Wang Y. Joint learning for pulmonary nodule segmentation, attributes and malignancy prediction. In: 2018 IEEE 15th international symposium on biomedical imaging (ISBI 2018); 2018. p. 1109–13. IEEE.
de Vos BD, Wolterink JM, Leiner T, de Jong PA, Lessmann N, Išgum I. Direct automatic coronary calcium scoring in cardiac and chest CT. IEEE Trans Med Imaging. 2019;38(9):2127–38.
Eitel F, Ritter K, ADNI. Testing the robustness of attribution methods for convolutional neural networks in MRI-based Alzheimer’s disease classification. In: Interpretability of machine intelligence in medical image computing and multimodal learning for clinical decision support. Berlin: Springer; 2019. p. 3–11.
Clough JR, Oksuz I, Puyol-Antón E, Ruijsink B, King AP, Schnabel J.A. Global and local interpretability for cardiac MRI classification. In: International conference on medical image computing and computer-assisted intervention; Springer. 2019. p. 656–4.
Gasimova A. Automated enriched medical concept generation for chest X-ray images. In: Interpretability of machine intelligence in medical image computing and multimodal learning for clinical decision support. Springer; 2019. p. 83–92.
Kim ST, Lee J-H, Ro YM. Visual evidence for interpreting diagnostic decision of deep neural network in computer-aided diagnosis. In: Medical imaging 2019: computer-aided diagnosis. 2019; vol. 10950, p. 139–47. SPIE.
Lee H, Kim ST, Ro YM. Generation of multimodal justification using visual word constraint model for explainable computer-aided diagnosis. In: Interpretability of machine intelligence in medical image computing and multimodal learning for clinical decision support. Springer; 2019. p. 21–9.
Shen S, Han SX, Aberle DR, Bui AA, Hsu W. An interpretable deep hierarchical semantic convolutional neural network for lung nodule malignancy classification. Expert Syst Appl. 2019;128:84–95.
Arun N, Gaw N, Singh P, Chang K, Aggarwal M, Chen B, et al. Assessing the (un) trustworthiness of saliency maps for localizing abnormalities in medical imaging (ar**v preprint); 2020.
Zeng X, Wen L, Xu Y, Ji C. Generating diagnostic report for medical image by high-middle-level visual information incorporation on double deep learning models. Comput Methods Programs Biomed. 2020;197:105700.
Yang S, Niu J, Wu J, Liu X. Automatic medical image report generation with multi-view and multi-modal attention mechanism. In: International conference on algorithms and architectures for parallel processing. Springer; 2020. p. 687–99.
Barnett AJ, Schwartz FR, Tao C, Chen C, Ren Y, Lo JY, Rudin C. A case-based interpretable deep learning model for classification of mass lesions in digital mammography. Nat Mach Intell. 2021;3(12):1061–70.
Saleem H, Shahid AR, Raza B. Visual interpretability in 3d brain tumor segmentation network. Comput Biol Med. 2021;133:104410.
Wang S, Yin Y, Wang D, Wang Y, ** Y. Interpretability-based multimodal convolutional neural networks for skin lesion diagnosis. IEEE Trans Cybern. 2021;20:20.
Ahmed U, Jhaveri RH, Srivastava G, Lin JC-W. Explainable deep attention active learning for sentimental analytics of mental disorder. Trans Asian Low-Resour Lang Inf Proces. 2022;20:22.
Lu Y, Perer A. An interactive interpretability system for breast cancer screening with deep learning. ar**v:2210.08979 (ar**v preprint); 2022.
Figueroa KC, Song B, Sunny S, Li S, Gurushanth K, Mendonca P, Mukhia N, Patrick S, Gurudath S, Raghavan S, et al. Interpretable deep learning approach for oral cancer classification using guided attention inference network. J Biomed Opt. 2022;27(1):015001.
Hicks SA, Eskeland S, Lux M, de Lange T, Randel KR, Jeppsson M, Pogorelov K, Halvorsen P, Riegler M. Mimir: an automatic reporting and reasoning system for deep learning based analysis in the medical domain. In: Proceedings of the 9th ACM multimedia systems conference; 2018. p. 369–74.
Holzinger A, Malle B, Saranti A, Pfeifer B. Towards multi-modal causability with graph neural networks enabling information fusion for explainable AI. Inf Fusion. 2021;71:28–37.
Palatnik de Sousa I, Maria Bernardes Rebuzzi Vellasco M, Costa da Silva E. Local interpretable model-agnostic explanations for classification of lymph node metastases. Sensors. 2019;19(13):2969.
Zhu P, Ogino M. Guideline-based additive explanation for computer-aided diagnosis of lung nodules. In: Interpretability of machine intelligence in medical image computing and multimodal learning for clinical decision support. Springer; 2019; p. 39–47.
Paschali M, Ferjadnaeem M, Simson W, et al. Improving the interpretability of medical imaging neural networks. In: Computer vision and pattern recognition; 2019.
Liao W, Zou B, Zhao R, Chen Y, He Z, Zhou M. Clinical interpretable deep learning model for glaucoma diagnosis. IEEE J Biomed Health Inform. 2019;24(5):1405–12.
Lee H, Yune S, Mansouri M, Kim M, Tajmir SH, Guerrier CE, Ebert SA, Pomerantz SR, Romero JM, Kamalian S, et al. An explainable deep-learning algorithm for the detection of acute intracranial haemorrhage from small datasets. Nat Biomed Eng. 2019;3(3):173–82.
Biffi C, Oktay O, Tarroni G, Bai W, De Marvao A, Doumou G, Rajchl M, Bedair R, Prasad S, Cook S, et al. Learning interpretable anatomical features through deep generative models: application to cardiac remodeling. In: International conference on medical image computing and computer-assisted intervention. Springer; 2018. p. 464–71.
Garcia-Peraza-Herrera LC, Everson M, Li W, Luengo I, Berger L, Ahmad O, Lovat L, Wang H-P, Wang W-L, Haidry R, et al. Interpretable fully convolutional classification of intrapapillary capillary loops for real-time detection of early squamous neoplasia. ar**v:1805.00632 (ar**v preprint); 2018.
Amoroso N, Pomarico D, Fanizzi A, Didonna V, Giotta F, La Forgia D, Latorre A, Monaco A, Pantaleo E, Petruzzellis N, et al. A roadmap towards breast cancer therapies supported by explainable artificial intelligence. Appl Sci. 2021;11(11):4881.
Sarp S, Kuzlu M, Wilson E, Cali U, Guler O. The enlightening role of explainable artificial intelligence in chronic wound classification. Electronics. 2021;10(12):1406.
Wu H, Chen W, Xu S, Xu B. Counterfactual supporting facts extraction for explainable medical record based diagnosis with graph network. In: Proceedings of the 2021 conference of the north American chapter of the association for computational linguistics: human language technologies; 2021. p. 1942–55.
Wang X, Peng Y, Lu L, Lu Z, Summers RM. Tienet: Text-image embedding network for common thorax disease classification and reporting in chest X-rays. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2018. p. 9049–58.
Lucieri A, Bajwa MN, Braun SA, Malik MI, Dengel A, Ahmed S. On interpretability of deep learning based skin lesion classifiers using concept activation vectors. In: 2020 international joint conference on neural networks (IJCNN); 2020. p. 1–10. IEEE.
Senior AW, Evans R, Jumper J, Kirkpatrick J, Sifre L, Green T, Qin C, Žídek A, Nelson AW, Bridgland A, et al. Improved protein structure prediction using potentials from deep learning. Nature. 2020;577(7792):706–10.
Jumper J, Evans R, Pritzel A, Green T, Figurnov M, Ronneberger O, Tunyasuvunakool K, Bates R, Žídek A, Potapenko A, et al. Highly accurate protein structure prediction with alphafold. Nature. 2021;596(7873):583–9.
Merk D, Friedrich L, Grisoni F, Schneider G. De novo design of bioactive small molecules by artificial intelligence. Mol Inf. 2018;37(1–2):1700153.
Zhavoronkov A, Ivanenkov YA, Aliper A, Veselov MS, Aladinskiy VA, Aladinskaya AV, Terentiev VA, Polykovskiy DA, Kuznetsov MD, Asadulaev A, et al. Deep learning enables rapid identification of potent ddr1 kinase inhibitors. Nat Biotechnol. 2019;37(9):1038–40.
Karimi M, Wu D, Wang Z, Shen Y. Explainable deep relational networks for predicting compound-protein affinities and contacts. J Chem Inf Model. 2020;61(1):46–66.
Ezzat D, Hassanien AE, Ella HA. An optimized deep learning architecture for the diagnosis of covid-19 disease based on gravitational search optimization. Appl Soft Comput. 2021;98:106742.
Segler MH, Kogej T, Tyrchan C, Waller MP. Generating focused molecule libraries for drug discovery with recurrent neural networks. ACS Cent Sci. 2018;4(1):120–31.
Preuer K, Renz P, Unterthiner T, Hochreiter S, Klambauer G. Fréchet chemnet distance: a metric for generative models for molecules in drug discovery. J Chem Inf Model. 2018;58(9):1736–41.
Wan Y, Zhou H, Zhang X. An interpretation architecture for deep learning models with the application of covid-19 diagnosis. Entropy. 2021;23(2):204.
Loh HW, Ooi CP, Seoni S, Barua PD, Molinari F, Acharya UR. Application of explainable artificial intelligence for healthcare: a systematic review of the last decade (2011–2022). Comput Methods Programs Biomed. 2022;20: 107161.
Duckworth C, Chmiel FP, Burns DK, Zlatev ZD, White NM, Daniels TW, Kiuber M, Boniface MJ. Using explainable machine learning to characterise data drift and detect emergent health risks for emergency department admissions during covid-19. Sci Rep. 2021;11(1):1–10.
Antoniadi AM, Galvin M, Heverin M, Hardiman O, Mooney C. Prediction of caregiver quality of life in amyotrophic lateral sclerosis using explainable machine learning. Sci Rep. 2021;11(1):1–13.
Zeng X, Hu Y, Shu L, Li J, Duan H, Shu Q, Li H. Explainable machine-learning predictions for complications after pediatric congenital heart surgery. Sci Rep. 2021;11(1):1–11.
Farhadloo M, Molnar C, Luo G, Li Y, Shekhar S, Maus RL, Markovic S, Leontovich A, Moore R. Samcnet: towards a spatially explainable AI approach for classifying MXIF oncology data. In: Proceedings of the 28th ACM SIGKDD conference on knowledge discovery and data mining; 2022. p. 2860–70.
Jiang Z, Bo L, Xu Z, Song Y, Wang J, Wen P, Wan X, Yang T, Deng X, Bian J. An explainable machine learning algorithm for risk factor analysis of in-hospital mortality in sepsis survivors with ICU readmission. Comput Methods Programs Biomed. 2021;204:106040.
Liu H, Zhong C, Alnusair A, Islam SR. Faixid: a framework for enhancing AI explainability of intrusion detection results using data cleaning techniques. J Netw Syst Manage. 2021;29(4):1–30.
Amarasinghe K, Manic M. Improving user trust on deep neural networks based intrusion detection systems. In: IECON 2018-44th annual conference of the IEEE Industrial electronics society; 2018. p. 3262–68. IEEE.
Amarasinghe K, Kenney K, Manic M. Toward explainable deep neural network based anomaly detection. In: 2018 11th international conference on human system interaction (HSI); 2018. IEEE. p. 311–7.
Chen S, Bateni S, Grandhi S, Li X, Liu C, Yang W. Denas: automated rule generation by knowledge extraction from neural networks. In: Proceedings of the 28th ACM joint meeting on European software engineering conference and symposium on the foundations of software engineering, 2020; p. 813–25.
Gulmezoglu B. Xai-based microarchitectural side-channel analysis for website fingerprinting attacks and defenses. IEEE Trans Depend Sec Comput. 2021;20:10.
Feichtner J, Gruber S. Understanding privacy awareness in android app descriptions using deep learning. In: Proceedings of the tenth ACM conference on data and application security and privacy; 2020. p. 203–14.
Iadarola G, Martinelli F, Mercaldo F, Santone A. Towards an interpretable deep learning model for mobile malware detection and family identification. Comput Secur. 2021;105:102198.
Guo W, Mu D, Xu J, Su P, Wang G, **ng X. Lemna: explaining deep learning based security applications. In: Proceedings of the 2018 ACM SIGSAC conference on computer and communications security; 2018. p. 364–79.
Yan A, Chen Z, Zhang H, Peng L, Yan Q, Hassan MU, Zhao C, Yang B. Effective detection of mobile malware behavior based on explainable deep neural network. Neurocomputing. 2021;453:482–92.
Bach S, Binder A, Montavon G, Klauschen F, Müller K-R, Samek W. On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. PLoS One. 2015;10(7):0130140.
Mane S, Rao D. Explaining network intrusion detection system using explainable AI framework. ar**v:2103.07110 (ar**v preprint); 2021.
Dash S, Gunluk O, Wei D. Boolean decision rules via column generation. Adv Neural Inf Process Syst. 2018;31:25.
Bose S, Barao T, Liu X. Explaining AI for malware detection: analysis of mechanisms of malconv. In: 2020 international joint conference on neural networks (IJCNN); 2020. IEEE. p. 1–8.
Al-Fawa’reh M, Saif A, Jafar MT, Elhassan A. Malware detection by eating a whole APK. In: 2020 15th international conference for internet technology and secured transactions (ICITST); 2020. IEEE. p. 1–7.
Ohana JJ, Ohana S, Benhamou E, Saltiel D, Guez B. Explainable AI (XAI) models applied to the multi-agent environment of financial markets. In: International workshop on explainable, transparent autonomous agents and multi-agent systems. Springer; 2021. p. 189–207.
Gramegna A, Giudici P. Shap and lime: an evaluation of discriminative power in credit risk. Front Artif Intell. 2021;140:25.
Wijnands M. Explaining black box decision-making: adopting explainable artificial intelligence in credit risk prediction for p2p lending. Master’s thesis, University of Twente; 2021.
El Qadi A, Trocan M, Diaz-Rodriguez N, Frossard T. Feature contribution alignment with expert knowledge for artificial intelligence credit scoring. Signal, Image and Video Processing; 2022. 1–8.
de Lange PE, Melsom B, Vennerød CB, Westgaard S. Explainable AI for credit assessment in banks. J Risk Financ Manage. 2022;15(12):556.
Górski Ł, Ramakrishna S. Explainable artificial intelligence, lawyer’s perspective. In: Proceedings of the eighteenth international conference on artificial intelligence and law; 2021. p. 60–8.
Berk RA, Bleich J. Statistical procedures for forecasting criminal behavior: a comparative assessment. Criminol Pub Pol’y. 2013;12:513.
Mardaoui D, Garreau D. An analysis of lime for text data. In: International conference on artificial intelligence and statistics; 2021. p. 3493–501. PMLR.
Khosravi H, Shum SB, Chen G, Conati C, Tsai Y-S, Kay J, Knight S, Martinez-Maldonado R, Sadiq S, Gašević D. Explainable artificial intelligence in education. Comput Educ Artif Intell. 2022;3:100074.
Alonso JM, Casalino G. Explainable artificial intelligence for human-centric data analysis in virtual learning environments. In: International workshop on higher education learning methodologies and technologies online. Springer; 2019. p. 125–38.
Ghai B, Liao QV, Zhang Y, Bellamy R, Mueller K. Explainable active learning (XAL) toward AI explanations as interfaces for machine teachers. Proc ACM Human Comput Interact. 2021;4(CSCW3):1–28.
Hu Y, Mello RF, Gašević D. Automatic analysis of cognitive presence in online discussions: an approach using deep learning and explainable artificial intelligence. Comput Educ Artif Intell. 2021;2:100037.
Hooshyar D, Yang Y. Neural-symbolic computing: a step toward interpretable AI in education. Bull Tech Committee Learn Technol (ISSN: 2306-0212) 2021;21(4), 2–6.
Melo E, Silva I, Costa DG, Viegas CM, Barros TM. On the use of explainable artificial intelligence to evaluate school dropout. Educ Sci. 2022;12(12):845.
Fernandez-Nieto GM, Echeverria V, Shum SB, Mangaroska K, Kitto K, Palominos E, Axisa C, Martinez-Maldonado R. Storytelling with learner data: guiding student reflection on multimodal team data. IEEE Trans Learn Technol. 2021;14(5):695–708.
Knight S, Shibani A, Abel S, Gibson A, Ryan P. Acawriter: a learning analytics tool for formative feedback on academic writing. J Writing Res. 2020;20:20.
Conati C, Barral O, Putnam V, Rieger L. Toward personalized XAI: a case study in intelligent tutoring systems. Artif Intell. 2021;298:103503.
Bojarski M, Choromanska A, Choromanski K, Firner B, Jackel L, Muller U, Zieba K. Visualbackprop: visualizing cnns for autonomous driving. ar**v:1611.05418 (ar**v preprint); 2016.
Kim J, Canny J. Interpretable learning for self-driving cars by visualizing causal attention. In: Proceedings of the IEEE international conference on computer vision; 2017. p. 2942–50.
Zeng W, Luo W, Suo S, Sadat A, Yang B, Casas S, Urtasun R. End-to-end interpretable neural motion planner. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition; 2019. p. 8660–9.
Omeiza D, Webb H, Jirotka M, Kunze L. Explanations in autonomous driving: a survey. IEEE Trans Intell Transport Syst. 2021;20:20.
Kothawade S, Khandelwal V, Basu K, Wang H, Gupta G. Auto-discern: autonomous driving using common sense reasoning. ar**v:2110.13606 (ar**v preprint); 2021.
Gao Y, Zhang S, Sun J, Yu S, Yamamoto T, Li Z, Li X. A joint framework based on accountable AI for driving behavior assessment and backtracking. In: 2022 IEEE 25th international conference on intelligent transportation systems (ITSC); 2022. IEEE. p. 268–74.
Machlev R, Heistrene L, Perl M, Levy K, Belikov J, Mannor S, Levron Y. Explainable artificial intelligence (XAI) techniques for energy and power systems: review, challenges and opportunities. Energy AI. 2022; 20:100169.
Zhang K, Xu P, Zhang J. Explainable AI in deep reinforcement learning models: a shap method applied in power system emergency control. In: 2020 IEEE 4th conference on energy internet and energy system integration (EI2); 2020. IEEE. p. 711–6.
Zhang K, Zhang J, Xu P-D, Gao T, Gao DW. Explainable AI in deep reinforcement learning models for power system emergency control. IEEE Trans Comput Soc Syst. 2021;9(2):419–27.
Shen D, Wu G, Suk H-I. Deep learning in medical image analysis. Annu Rev Biomed Eng. 2017;19:221–48.
Thompson AC, Jammal AA, Medeiros FA. A review of deep learning for screening, diagnosis, and detection of glaucoma progression. Transl Vis Sci Technol. 2020;9(2):42–42.
Moolayil J. An introduction to deep learning and Keras. In: Learn Keras for deep neural networks. Berlin: Springer; 2019. p. 1–16.
Zhang Z, Chen P, Sapkota M, Yang L. Tandemnet: Distilling knowledge from medical images using diagnostic reports as optional semantic references. In: International conference on medical image computing and computer-assisted intervention. Springer; 2017. p. 320–8.
Altinkaya E, Polat K, Barakli B. Detection of Alzheimer’s disease and dementia states based on deep learning from MRI images: a comprehensive review. J Inst Electron Comput. 2020;1(1):39–53.
Mathews SM. Explainable artificial intelligence applications in nlp, biomedical, and malware classification: a literature review. In: Intelligent computing-proceedings of the computing conference; Springer. 2019. p. 1269–92.
Madanu R, Abbod MF, Hsiao F-J, Chen W-T, Shieh J-S. Explainable AI (XAI) applied in machine learning for pain modeling: a review. Technologies. 2022;10(3):74.
Garvin MR, Prates ET, Pavicic M, Jones P, Amos BK, Geiger A, Shah MB, Streich J, Gazolla JGFM, Kainer D, et al. Potentially adaptive SARS-COV-2 mutations discovered with novel spatiotemporal and explainable AI models. Genome Biol. 2020;21(1):1–26.
Cliff A, Romero J, Kainer D, Walker A, Furches A, Jacobson D. A high-performance computing implementation of iterative random forest for the creation of predictive expression networks. Genes. 2019;10(12):996.
Shah RD, Meinshausen N. Random intersection trees. J Mach Learn Res. 2014;15(1):629–54.
Ikemura T, Wada K, Wada Y, Iwasaki Y, Abe T. Unsupervised explainable AI for simultaneous molecular evolutionary study of forty thousand sars-cov-2 genomes. bioRxiv; 2020.
Lundberg SM, Lee S-I. A unified approach to interpreting model predictions. Adv Neural Inf Process Syst. 2017;30:25.
Prajod P, Huber T, André E. Using explainable AI to identify differences between clinical and experimental pain detection models based on facial expressions. In: International conference on multimedia modeling. Springer; 2022. p. 311–22.
Dasgupta D, Akhtar Z, Sen S. Machine learning in cybersecurity: a comprehensive survey. J Defense Model Simul. 2022;19(1):57–106.
Ucci D, Aniello L, Baldoni R. Survey of machine learning techniques for malware analysis. Comput Secur. 2019;81:123–47.
Perarasi T, Vidhya S, Ramya P, et al. Malicious vehicles identifying and trust management algorithm for enhance the security in 5g-vanet. In: 2020 second international conference on inventive research in computing applications (ICIRCA); 2020. p. 269–75. IEEE.
Jaswal G, Kanhangad V, Ramachandra R. AI and deep learning in biometric security: trends, potential, and challenges. Boca Raton: CRC Press; 2021.
Rudin C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat Mach Intell. 2019;1(5):206–15.
Zhang Z, Hamadi HA, Damiani E, Yeun CY, Taher F. Explainable artificial intelligence applications in cyber security: state-of-the-art in research. ar**v:2208.14937 (ar**v preprint); 2022.
Capuano N, Fenza G, Loia V, Stanzione C. Explainable artificial intelligence in cybersecurity: a survey. IEEE Access. 2022;10:93575–600.
Buczak AL, Guven E. A survey of data mining and machine learning methods for cyber security intrusion detection. IEEE Commun Surv Tutor. 2015;18(2):1153–76.
Chalapathy R, Chawla S. Deep learning for anomaly detection: a survey. ar**v:1901.03407 (ar**v preprint); 2019.
Carta S, Podda AS, Reforgiato Recupero D, Stanciu MM. Explainable AI for financial forecasting. In: International conference on machine learning, optimization, and data science; Springer. 2021. p. 51–69.
Chromik M, Eiband M, Buchner F., Krüger A, Butz A. I think i get your point, AI! the illusion of explanatory depth in explainable AI. In: 26th international conference on intelligent user interfaces; 2021. p. 307–17.
Bussmann N, Giudici P, Marinelli D, Papenbrock J. Explainable machine learning in credit risk management. Comput Econ. 2021;57(1):203–16.
Agarwal A, Bhatia A, Malhi A, Kaler P, Pannu HS, et al. Machine learning based explainable financial forecasting. In: 2022 4th international conference on computer communication and the internet (ICCCI); 2022. p. 34–8. IEEE.
Eliot DLB. The need for explainable AI (XAI) is especially crucial in the law. Available at SSRN 3975778; 2021.
Williamson B. Digital policy sociology: software and science in data-intensive precision education. Crit Stud Educ. 2019;20:1–17.
Luan H, Tsai C-C. A review of using machine learning approaches for precision education. Educ Technol Soc. 2021;24(1):250–66.
Akgun S, Greenhow C. Artificial intelligence in education: addressing ethical challenges in k-12 settings. AI Ethics. 2021;20:1–10.
Gardner J, Brooks C, Baker R. Evaluating the fairness of predictive student models through slicing analysis. In: Proceedings of the 9th international conference on learning analytics and knowledge. p. 225–234; 2019.
Atakishiyev S, Salameh M, Yao H, Goebel R. Explainable artificial intelligence for autonomous driving: a comprehensive overview and field guide for future research directions. ar**v:2112.11561 (ar**v preprint); 2021.
Ni J, Chen Y, Chen Y, Zhu J, Ali D, Cao W. A survey on theories and applications for self-driving cars based on deep learning methods. Appl Sci. 2020;10(8):2749.
Yousuf H, Zainal AY, Alshurideh M, Salloum SA. Artificial intelligence models in power system analysis. In: Artificial intelligence for sustainable development: theory, practice and future applications; Springer. 2021. p. 231–42.
Lorente MPS, Lopez EM, Florez LA, Espino AL, Martínez JAI, de Miguel AS. Explaining deep learning-based driver models. Appl Sci. 2021;11(8):3321.
Konstantakopoulos IC, Das HP, Barkan AR, He S, Veeravalli T, Liu H, Manasawala AB, Lin Y-W, Spanos CJ. Design, benchmarking and explainability analysis of a game-theoretic framework towards energy efficiency in smart infrastructure. ar**v:1910.07899 (ar**v preprint); 2019.
Pannu HS, Malhi A, et al. Deep learning-based explainable target classification for synthetic aperture radar images. In: 2020 13th international conference on human system interaction (HSI); 2020. p. 34–9. IEEE.
Nahata R, Omeiza D, Howard R, Kunze L. Assessing and explaining collision risk in dynamic environments for autonomous driving safety. In: 2021 IEEE international intelligent transportation systems conference (ITSC); 2021. p. 223–30. IEEE.
Acknowledgements
This work was supported by the Institute of Information and Communications Technology Planning and Evaluation (IITP) grant funded by the Korean Government (MSIT) (2022-0-00078, Explainable Logical Reasoning for Medical Knowledge Generation).
Author information
Authors and Affiliations
Contributions
The authors confirm their contribution to the paper as follows: study conception and design: WY; draft manuscript preparation: YW, HW, YC, GH, XL, RL, NY, XW, and XG. Supervision: MBA and BK. All authors reviewed the results and approved the final version of the manuscript.
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Yang, W., Wei, Y., Wei, H. et al. Survey on Explainable AI: From Approaches, Limitations and Applications Aspects. Hum-Cent Intell Syst 3, 161–188 (2023). https://doi.org/10.1007/s44230-023-00038-y
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s44230-023-00038-y