[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 

Advances in Explainable Artificial Intelligence (XAI)

A special issue of Machine Learning and Knowledge Extraction (ISSN 2504-4990). This special issue belongs to the section "Learning".

Deadline for manuscript submissions: closed (15 July 2023) | Viewed by 91802

Special Issue Editor

School of Computer Science, Technological University Dublin, D08 X622 Dublin, Ireland
Interests: explainable artificial intelligence; defeasible argumentation; deep learning; human-centred design; mental workload modeling
Special Issues, Collections and Topics in MDPI journals

Special Issue Information

Dear Colleagues,

Recently, artificial intelligence has seen a shift in focus towards the design and deployment of intelligent systems that are interpretable and explainable, with the rise of a new field: explainable artificial intelligence (XAI). This has echoed both in the research literature and in the press, attracting scholars from all around the world as well as a lay audience. Initially devoted to the design of post-hoc methods for explainability, essentially wrapping machine- and deep-learning models with explanations, it is now expanding its boundaries to ante-hoc methods for the production of self-interpretable models. Along with this, neuro-symbolic approaches for reasoning have been employed in conjunction with machine learning in order to extend modelling accuracy and precision with self-explainability and justifiability. Scholars started also shifting the focus on the structure of explanations since the ultimate users of interactive technologies are humans, linking artificial intelligence and computer sciences to psychology, human–computer interaction, philosophy, and sociology.
 
It is certain that explainable artificial intelligence is gaining momentum, and this Special Issue calls for contributions in this new fascinating area of research, seeking articles that are devoted to the theoretical foundation of XAI, its historical perspectives, and the design of explanations and interactive human-centered intelligent systems with knowledge-representation principles and automated learning capabilities, not only for experts but for the lay audience as well.

Dr. Luca Longo
Guest Editor

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Machine Learning and Knowledge Extraction is an international peer-reviewed open access quarterly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 1800 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • explainable artificial intelligence (XAI)
  • neuro-symbolic reasoning for XAI
  • interpretable deep learning
  • argument-based models of explanations
  • graph neural networks for explainability
  • machine learning and knowledge-graphs
  • human-centric explainable AI
  • interpretation of black-box models
  • human-understandable machine learning
  • counterfactual explanations for machine learning
  • natural language processing in XAI
  • quantitative/qualitative evaluation metrics for XAI
  • ante and post-hoc XAI methods
  • rule-based systems for XAI
  • fuzzy systems and explainability
  • human-centered learning and explanations
  • model-dependent and model-agnostic explainability
  • case-based explanations for AI systems
  • interactive machine learning and explanations

Benefits of Publishing in a Special Issue

  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • e-Book format: Special Issues with more than 10 articles can be published as dedicated e-books, ensuring wide and rapid dissemination.

Further information on MDPI's Special Issue polices can be found here.

Related Special Issue

Published Papers (11 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

Jump to: Other

20 pages, 1544 KiB  
Article
Alternative Formulations of Decision Rule Learning from Neural Networks
by Litao Qiao, Weijia Wang and Bill Lin
Mach. Learn. Knowl. Extr. 2023, 5(3), 937-956; https://doi.org/10.3390/make5030049 - 3 Aug 2023
Viewed by 1932
Abstract
This paper extends recent work on decision rule learning from neural networks for tabular data classification. We propose alternative formulations to trainable Boolean logic operators as neurons with continuous weights, including trainable NAND neurons. These alternative formulations provide uniform treatments to different trainable [...] Read more.
This paper extends recent work on decision rule learning from neural networks for tabular data classification. We propose alternative formulations to trainable Boolean logic operators as neurons with continuous weights, including trainable NAND neurons. These alternative formulations provide uniform treatments to different trainable logic neurons so that they can be uniformly trained, which enables, for example, the direct application of existing sparsity-promoting neural net training techniques like reweighted L1 regularization to derive sparse networks that translate to simpler rules. In addition, we present an alternative network architecture based on trainable NAND neurons by applying De Morgan’s law to realize a NAND-NAND network instead of an AND-OR network, both of which can be readily mapped to decision rule sets. Our experimental results show that these alternative formulations can also generate accurate decision rule sets that achieve state-of-the-art performance in terms of accuracy in tabular learning applications. Full article
(This article belongs to the Special Issue Advances in Explainable Artificial Intelligence (XAI))
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) An example of the DR-Net architecture with 4 AND neurons is shown. The blue lines to the AND neurons represent positive weights, while red lines represent negative weights. The dashed line indicates the exclusion of the corresponding input feature. Please note that we represent “NOT (GPA ≥ 3.0)” as “GPA &lt; 3.0” in the third rule. Similarly, “NOT (SAT ≥ 1000)” is represented as “SAT &lt; 1000”. For the output OR neuron, the blue line indicates that the corresponding rule is included in the rule set, and the dashed line indicates that the corresponding rule is excluded. (<b>b</b>) The network maps directly to the corresponding decision rule set shown in the box on the right.</p>
Full article ">Figure 2
<p>(<b>a</b>) A variation of the example in <a href="#make-05-00049-f001" class="html-fig">Figure 1</a>, in which the red line to the output OR neuron indicates the negation of the corresponding rule “GPA &lt; 3.0 AND SAT &lt; 1000.” By De Morgan’s law, the negation of “GPA &lt; 3.0 AND SAT &lt; 1000” becomes “GPA ≥ 3.0 OR SAT ≥ 1000”, which results in the same decision rule set. (<b>b</b>) The corresponding decision rule set is shown on the right.</p>
Full article ">Figure 3
<p>Training statistics (training loss, training accuracy, number of rules, and rule complexity) as functions of the number of epochs in the training process.</p>
Full article ">Figure 4
<p>The relations between complexities (rule complexity and number of rules) and regularization parameters (<math display="inline"><semantics><msub><mi>λ</mi><mn>1</mn></msub></semantics></math> and <math display="inline"><semantics><msub><mi>λ</mi><mn>2</mn></msub></semantics></math>). All x-axes are on a log10 scale. All complexity values were averaged over five cross-validation partitions, and the vertical bars represent standard deviations.</p>
Full article ">Figure 5
<p>Accuracy–Complexity trade-offs on all datasets for DR-Net and NN-Net trained using <math display="inline"><semantics><msub><mi>L</mi><mn>0</mn></msub></semantics></math> regularization. Pareto efficient points are connected by line segments.</p>
Full article ">Figure 6
<p>Accuracy–Complexity trade-offs on all datasets for DR-Net and NN-Net trained using reweighted <math display="inline"><semantics><msub><mi>L</mi><mn>1</mn></msub></semantics></math> regularization. Pareto efficient points are connected by line segments.</p>
Full article ">
15 pages, 2157 KiB  
Article
Achievable Minimally-Contrastive Counterfactual Explanations
by Hosein Barzekar and Susan McRoy
Mach. Learn. Knowl. Extr. 2023, 5(3), 922-936; https://doi.org/10.3390/make5030048 - 3 Aug 2023
Viewed by 2165
Abstract
Decision support systems based on machine learning models should be able to help users identify opportunities and threats. Popular model-agnostic explanation models can identify factors that support various predictions, answering questions such as “What factors affect sales?” or “Why did sales decline?”, but [...] Read more.
Decision support systems based on machine learning models should be able to help users identify opportunities and threats. Popular model-agnostic explanation models can identify factors that support various predictions, answering questions such as “What factors affect sales?” or “Why did sales decline?”, but do not highlight what a person should or could do to get a more desirable outcome. Counterfactual explanation approaches address intervention, and some even consider feasibility, but none consider their suitability for real-time applications, such as question answering. Here, we address this gap by introducing a novel model-agnostic method that provides specific, feasible changes that would impact the outcomes of a complex Black Box AI model for a given instance and assess its real-world utility by measuring its real-time performance and ability to find achievable changes. The method uses the instance of concern to generate high-precision explanations and then applies a secondary method to find achievable minimally-contrastive counterfactual explanations (AMCC) while limiting the search to modifications that satisfy domain-specific constraints. Using a widely recognized dataset, we evaluated the classification task to ascertain the frequency and time required to identify successful counterfactuals. For a 90% accurate classifier, our algorithm identified AMCC explanations in 47% of cases (38 of 81), with an average discovery time of 80 ms. These findings verify the algorithm’s efficiency in swiftly producing AMCC explanations, suitable for real-time systems. The AMCC method enhances the transparency of Black Box AI models, aiding individuals in evaluating remedial strategies or assessing potential outcomes. Full article
(This article belongs to the Special Issue Advances in Explainable Artificial Intelligence (XAI))
Show Figures

Figure 1

Figure 1
<p>Distribution of success (1) and failure (0) of modifications.</p>
Full article ">Figure 2
<p>Distribution of time taken (in seconds) for modification.</p>
Full article ">Figure 3
<p>Comparison of time taken for successful (1) vs. unsuccessful (0) modifications.</p>
Full article ">Figure 4
<p>Variable modifications and their respective durations.</p>
Full article ">
21 pages, 17177 KiB  
Article
What about the Latent Space? The Need for Latent Feature Saliency Detection in Deep Time Series Classification
by Maresa Schröder, Alireza Zamanian and Narges Ahmidi
Mach. Learn. Knowl. Extr. 2023, 5(2), 539-559; https://doi.org/10.3390/make5020032 - 18 May 2023
Cited by 3 | Viewed by 2444
Abstract
Saliency methods are designed to provide explainability for deep image processing models by assigning feature-wise importance scores and thus detecting informative regions in the input images. Recently, these methods have been widely adapted to the time series domain, aiming to identify important temporal [...] Read more.
Saliency methods are designed to provide explainability for deep image processing models by assigning feature-wise importance scores and thus detecting informative regions in the input images. Recently, these methods have been widely adapted to the time series domain, aiming to identify important temporal regions in a time series. This paper extends our former work on identifying the systematic failure of such methods in the time series domain to produce relevant results when informative patterns are based on underlying latent information rather than temporal regions. First, we both visually and quantitatively assess the quality of explanations provided by multiple state-of-the-art saliency methods, including Integrated Gradients, Deep-Lift, Kernel SHAP, and Lime using univariate simulated time series data with temporal or latent patterns. In addition, to emphasize the severity of the latent feature saliency detection problem, we also run experiments on a real-world predictive maintenance dataset with known latent patterns. We identify Integrated Gradients, Deep-Lift, and the input-cell attention mechanism as potential candidates for refinement to yield latent saliency scores. Finally, we provide recommendations on using saliency methods for time series classification and suggest a guideline for developing latent saliency methods for time series. Full article
(This article belongs to the Special Issue Advances in Explainable Artificial Intelligence (XAI))
Show Figures

Figure 1

Figure 1
<p>Example of classification of vibration systems where shapelets (oscillation patterns) are proxies for latent parameters (damping ratio). Each plot contains the responses from three example systems of the same system type (illustrated with different shades of the same color). For these systems, a label, correlated with the damping ratio <math display="inline"><semantics> <mi>ζ</mi> </semantics></math>, can be potentially predicted by shapelet-based classifiers; however, a conventional saliency method applied to this problem will only highlight a proxy of the informative latent feature, namely, the existing fluctuations and oscillations of the time series. (<b>a</b>) Underdamped systems: <math display="inline"><semantics> <mrow> <mi>ζ</mi> <mo>&lt;</mo> <mn>1</mn> </mrow> </semantics></math>; (<b>b</b>) overdamped systems: <math display="inline"><semantics> <mrow> <mi>ζ</mi> <mo>&gt;</mo> <mn>1</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 2
<p>Explainability toy examples of multiple label-making scenarios in the time series domain. Per each column, example time series are colored blue for class 0, and gold for class 1. Influential time steps (regions with high saliency scores) are shaded in gray for frequency (the peaks), amplitude (highest peaks), trend (a window enough for inferring about the trend), and shapelet (presence of the informative pattern) [<a href="#B5-make-05-00032" class="html-bibr">5</a>].</p>
Full article ">Figure 3
<p>Distribution of engineered features of the CWRU dataset for normal (blue) and faulty (orange) bearings. Important features which allow for a simple distinctive line between the classes are circled.</p>
Full article ">Figure 4
<p>Comparison of importance heat maps from feature attribution methods IG, Deep-Lift, Lime, and SHAP for the CNN + SGT on a frequency, amplitude, phase shift, and shapelet experiment, respectively. Explanations provided by IG and Deep-Lift clearly focus on aspects related to the latent feature (peaks and valleys for amplitude and frequency, beginning of time sequence for phase shift) and the shapelet, respectively. Maps of Lime and SHAP are visually uninterpretable [<a href="#B5-make-05-00032" class="html-bibr">5</a>].</p>
Full article ">Figure 5
<p>Comparison of explanations provided by the gradient-based saliency method IG on the LSTM, LSTM + SGT, and input-cell attention LSTM if the latent feature amplitude (<b>top</b>) or a shapelet at a fixed position (<b>bottom</b>) is class-distinctive.</p>
Full article ">Figure 6
<p>Comparison of importance heat maps from attention scores (<b>left</b>) and IG (<b>right</b>) for the two classes of experiment 6. The attention mechanism clearly focuses on frequency-related patterns in the time series. The identified importance patterns of both saliency maps coincide.</p>
Full article ">Figure 7
<p>Average saliency performance of the tested post hoc methods evaluated through sanity (<b>left</b>) and faithfulness (<b>right</b>).</p>
Full article ">Figure 8
<p>Heat map of faithfulness results split by experiments.</p>
Full article ">Figure 9
<p>Faithfulness of saliency maps split by type of class-distinctive feature (latent feature vs. shapelet) for the two well-performing classifiers CNN + SGT (<b>left</b>) and TCN (<b>right</b>).</p>
Full article ">Figure 10
<p>Quantitative saliency evaluation for the CNN trained via saliency-guided training on the CWRU Bearing dataset depicted as a comparison of all four post hoc saliency methods based on sanity and faithfulness scores (<b>right</b>) and drop in similarity between saliency maps when randomizing different percentages of network weights (<b>right</b>).</p>
Full article ">Figure 11
<p>Explainability heat maps of Integrated Gradients (<b>left</b>) and Deep Lift (<b>right</b>) for each one sample of the positive (<b>top</b>) and the negative (<b>bottom</b>) class of the classification of the CNN trained via saliency-guided training on the CWRU Bearing dataset.</p>
Full article ">Figure 12
<p>Explanations provided by the counterfactual method Native Guide when the class label depends on one of the latent features amplitude (<b>left</b>) or phase shift (<b>right</b>).</p>
Full article ">
18 pages, 3481 KiB  
Article
Painting the Black Box White: Experimental Findings from Applying XAI to an ECG Reading Setting
by Federico Cabitza, Andrea Campagner, Chiara Natali, Enea Parimbelli, Luca Ronzio and Matteo Cameli
Mach. Learn. Knowl. Extr. 2023, 5(1), 269-286; https://doi.org/10.3390/make5010017 - 8 Mar 2023
Cited by 9 | Viewed by 3662
Abstract
The emergence of black-box, subsymbolic, and statistical AI systems has motivated a rapid increase in the interest regarding explainable AI (XAI), which encompasses both inherently explainable techniques, as well as approaches to make black-box AI systems explainable to human decision makers. Rather than [...] Read more.
The emergence of black-box, subsymbolic, and statistical AI systems has motivated a rapid increase in the interest regarding explainable AI (XAI), which encompasses both inherently explainable techniques, as well as approaches to make black-box AI systems explainable to human decision makers. Rather than always making black boxes transparent, these approaches are at risk of painting the black boxes white, thus failing to provide a level of transparency that would increase the system’s usability and comprehensibility, or even at risk of generating new errors (i.e., white-box paradox). To address these usability-related issues, in this work we focus on the cognitive dimension of users’ perception of explanations and XAI systems. We investigated these perceptions in light of their relationship with users’ characteristics (e.g., expertise) through a questionnaire-based user study involved 44 cardiology residents and specialists in an AI-supported ECG reading task. Our results point to the relevance and correlation of the dimensions of trust, perceived quality of explanations, and tendency to defer the decision process to automation (i.e., technology dominance). This contribution calls for the evaluation of AI-based support systems from a human–AI interaction-oriented perspective, laying the ground for further investigation of XAI and its effects on decision making and user experience. Full article
(This article belongs to the Special Issue Advances in Explainable Artificial Intelligence (XAI))
Show Figures

Figure 1

Figure 1
<p>Screenshot taken from one of the pages of the online questionnaire was used in the user study. The top of the image shows part of the ECG shown to participants for a clinical case and, particularly on the left, a magnified portion of the trace from the "magnifying glass" feature that could be activated by simply hovering the pointer over the image. Visible below is the advice given by the AI (in this case, “atrial fibrillation”) and in the area below the explanation provided (in this case, “absence of P waves and irregularities in the frequency of the QRS complexes”). At the bottom, one can see the question item by which the user could confirm the diagnosis given previously (on the previous page), or choose the diagnosis provided by the AI, or enter another diagnosis (different from the one entered previously).</p>
Full article ">Figure 2
<p>BPMN representation of the study design. Information collected is represented as data objects, coming from collection tasks. Its name is denoted by the name of the main actor. After the initial collection of the perceived “trust in AI” (Initial Trust, IT), the subprocess is repeated for each ECG case, where HD1, AI, HD2, XAI, and FHD items are collected, together with comprehensibility, appropriateness, and utility; these acronyms denote (see the abbreviations list at the end of the article) the first diagnosis humans provide (HD1) before receiving the AI advice (AI), the diagnosis recorded immediately after receiving this AI advice (HD2), the information regarding the XAI support (XAI), and the final and definitive diagnosis (FHD). Finally, a posttest “trust in AI” is collected again (Final Trust, FT).</p>
Full article ">Figure 3
<p>Violin plots of the distributions of the initial (pretest, IT in <a href="#make-05-00017-f002" class="html-fig">Figure 2</a>) and final (posttest, FT in <a href="#make-05-00017-f002" class="html-fig">Figure 2</a>) trust scores reported by the study participants, stratified by (<b>left</b>) readers’ expertise (novices vs. experts), and (<b>right</b>) interaction protocol (human-first vs. AI-first).</p>
Full article ">Figure 4
<p>Scatter plots of the correlations observed between the appropriateness, comprehensibility, and utility scores associated with each explanation provided by the XAI module of the AI support.</p>
Full article ">Figure 5
<p>Violin plots of the explanations’ quality, stratified by (<b>left</b>) readers’ expertise, (<b>center</b>) interaction protocol, and (<b>right</b>) readers’ baseline accuracy.</p>
Full article ">Figure 6
<p>Scatter plots of the correlations observed between the scores of explanations’ perceived quality and (<b>left</b>) initial (pretest) trust, (<b>right</b>) final (posttest) trust.</p>
Full article ">Figure 7
<p>Scatter plots of the correlations between dominance (both positive and negative) and explanations’ quality. Dominance is defined as the number of decision changes occurring between HD2 and FHD.</p>
Full article ">Figure 8
<p>Violin plots showing the effect of the classifications’ and explanations’ correctness on (<b>left</b>) perceived explanations’ quality and (<b>right</b>) dominance.</p>
Full article ">Figure 9
<p>Matrix of the pairwise comparisons of the effects of the classifications’ and explanations’ correctness on dominance. Cells under the diagonal denote the <span class="html-italic">p</span>-value (Nemenyi post-hoc test) for the given pair of configurations, whereas cells above the diagonal report the corresponding effect size (RBC). Brighter shades of red (resp. blue) denote significance (resp. strength of the effect).</p>
Full article ">
12 pages, 2247 KiB  
Article
An Explainable Deep Learning Framework for Detecting and Localising Smoke and Fire Incidents: Evaluation of Grad-CAM++ and LIME
by Ioannis D. Apostolopoulos, Ifigeneia Athanasoula, Mpesi Tzani and Peter P. Groumpos
Mach. Learn. Knowl. Extr. 2022, 4(4), 1124-1135; https://doi.org/10.3390/make4040057 - 6 Dec 2022
Cited by 11 | Viewed by 3487
Abstract
Climate change is expected to increase fire events and activity with multiple impacts on human lives. Large grids of forest and city monitoring devices can assist in incident detection, accelerating human intervention in extinguishing fires before they get out of control. Artificial Intelligence [...] Read more.
Climate change is expected to increase fire events and activity with multiple impacts on human lives. Large grids of forest and city monitoring devices can assist in incident detection, accelerating human intervention in extinguishing fires before they get out of control. Artificial Intelligence promises to automate the detection of fire-related incidents. This study enrols 53,585 fire/smoke and normal images and benchmarks seventeen state-of-the-art Convolutional Neural Networks for distinguishing between the two classes. The Xception network proves to be superior to the rest of the CNNs, obtaining very high accuracy. Grad-CAM++ and LIME algorithms improve the post hoc explainability of Xception and verify that it is learning features found in the critical locations of the image. Both methods agree on the suggested locations, strengthening the abovementioned outcome. Full article
(This article belongs to the Special Issue Advances in Explainable Artificial Intelligence (XAI))
Show Figures

Figure 1

Figure 1
<p>Dataset creation pipeline.</p>
Full article ">Figure 2
<p>Fire and smoke detection framework.</p>
Full article ">Figure 3
<p>Random samples from the Grad-CAM++ assisted output of Xception CNN. The red color implies areas of high significance according to the model. Green implies medium significance and blue minor sigificance.</p>
Full article ">Figure 4
<p>Random samples produced by LIME applied to Xception CNN. LIME draws a yellow segmentation area around the most significant location according to the model.</p>
Full article ">
23 pages, 4081 KiB  
Article
On the Dimensionality and Utility of Convolutional Autoencoder’s Latent Space Trained with Topology-Preserving Spectral EEG Head-Maps
by Arjun Vinayak Chikkankod and Luca Longo
Mach. Learn. Knowl. Extr. 2022, 4(4), 1042-1064; https://doi.org/10.3390/make4040053 - 18 Nov 2022
Cited by 13 | Viewed by 3126
Abstract
Electroencephalography (EEG) signals can be analyzed in the temporal, spatial, or frequency domains. Noise and artifacts during the data acquisition phase contaminate these signals adding difficulties in their analysis. Techniques such as Independent Component Analysis (ICA) require human intervention to remove noise and [...] Read more.
Electroencephalography (EEG) signals can be analyzed in the temporal, spatial, or frequency domains. Noise and artifacts during the data acquisition phase contaminate these signals adding difficulties in their analysis. Techniques such as Independent Component Analysis (ICA) require human intervention to remove noise and artifacts. Autoencoders have automatized artifact detection and removal by representing inputs in a lower dimensional latent space. However, little research is devoted to understanding the minimum dimension of such latent space that allows meaningful input reconstruction. Person-specific convolutional autoencoders are designed by manipulating the size of their latent space. A sliding window technique with overlapping is employed to segment varied-sized windows. Five topographic head-maps are formed in the frequency domain for each window. The latent space of autoencoders is assessed using the input reconstruction capacity and classification utility. Findings indicate that the minimal latent space dimension is 25% of the size of the topographic maps for achieving maximum reconstruction capacity and maximizing classification accuracy, which is achieved with a window length of at least 1 s and a shift of 125 ms, using the 128 Hz sampling rate. This research contributes to the body of knowledge with an architectural pipeline for eliminating redundant EEG data while preserving relevant features with deep autoencoders. Full article
(This article belongs to the Special Issue Advances in Explainable Artificial Intelligence (XAI))
Show Figures

Figure 1

Figure 1
<p>(<b>A</b>) Data pre-processing pipeline gives topographic head-maps (TPHM) as output from 32 channel EEG signals (<b>B</b>) ConvAE for obtaining the latent space (<b>C</b>) Reconstruction metrics to evaluate the latent space (LS) quality (<b>D</b>) DNN with LS and TPHM as inputs (<b>E</b>) Utility (U) to measure the model’s performance.</p>
Full article ">Figure 2
<p>Convolutional Autoencoder’s (ConvAE) architecture for learning optimal latent space (LS) from topology-preserving head-maps (TPHM) for delta, theta, alpha, beta, and gamma EEG bands.</p>
Full article ">Figure 3
<p>A fully connected Dense Neural Network (DNN) architecture for predicting Video ID in music video classification problems. ConvAE’s LS and TPHM are input to DNN at separate instances to examine optimal features.</p>
Full article ">Figure 4
<p>Reconstruction ability, as measured by Structural Similarity Index (SSIM), Mean Squared Error (MSE), Normalised Root Mean Squared Error (NRMSE), and Peak Signal-to-Noise Ratio (PSNR) for 48 different ConvAE models with varying window length (WL), window shift (WS), and latent space (LS).</p>
Full article ">Figure 5
<p>3D plot depicting mean reconstruction scores (SSIM, MSE, NRMSE, and PSNR) for 48 different ConvAE models with varying window length (WL), window shift (WS), and latent space (LS).</p>
Full article ">Figure 6
<p>Utility (U) scores (Accuracy, F1-score) for 60 different DNN models with varying window length (WL), window shift (WS), and two kinds of input viz latent space (LS) and Topology Preserved Head-Maps (TPHM).</p>
Full article ">Figure 7
<p>3D plot depicting mean Utility (U) scores (Accuracy, F1-score) for 60 different DNN models with varying window length (WL), window shift (WS), and two kinds of input viz latent space (LS) and Topology Preserved Head-Maps (TPHM).</p>
Full article ">Figure 8
<p>Aggregate mean accuracy percentage of the 60 models on window length (WL), window shift (WS), and latent space (LS).</p>
Full article ">Figure 9
<p>Accuracy distribution of the classification models for all the participants across window length, window shift, and latent space.</p>
Full article ">Figure 9 Cont.
<p>Accuracy distribution of the classification models for all the participants across window length, window shift, and latent space.</p>
Full article ">
22 pages, 923 KiB  
Article
A Multi-Component Framework for the Analysis and Design of Explainable Artificial Intelligence
by Mi-Young Kim, Shahin Atakishiyev, Housam Khalifa Bashier Babiker, Nawshad Farruque, Randy Goebel, Osmar R. Zaïane, Mohammad-Hossein Motallebi, Juliano Rabelo, Talat Syed, Hengshuai Yao and Peter Chun
Mach. Learn. Knowl. Extr. 2021, 3(4), 900-921; https://doi.org/10.3390/make3040045 - 18 Nov 2021
Cited by 29 | Viewed by 6880
Abstract
The rapid growth of research in explainable artificial intelligence (XAI) follows on two substantial developments. First, the enormous application success of modern machine learning methods, especially deep and reinforcement learning, have created high expectations for industrial, commercial, and social value. Second, the emerging [...] Read more.
The rapid growth of research in explainable artificial intelligence (XAI) follows on two substantial developments. First, the enormous application success of modern machine learning methods, especially deep and reinforcement learning, have created high expectations for industrial, commercial, and social value. Second, the emerging and growing concern for creating ethical and trusted AI systems, including compliance with regulatory principles to ensure transparency and trust. These two threads have created a kind of “perfect storm” of research activity, all motivated to create and deliver any set of tools and techniques to address the XAI demand. As some surveys of current XAI suggest, there is yet to appear a principled framework that respects the literature of explainability in the history of science and which provides a basis for the development of a framework for transparent XAI. We identify four foundational components, including the requirements for (1) explicit explanation knowledge representation, (2) delivery of alternative explanations, (3) adjusting explanations based on knowledge of the explainee, and (4) exploiting the advantage of interactive explanation. With those four components in mind, we intend to provide a strategic inventory of XAI requirements, demonstrate their connection to a basic history of XAI ideas, and then synthesize those ideas into a simple framework that can guide the design of AI systems that require XAI. Full article
(This article belongs to the Special Issue Advances in Explainable Artificial Intelligence (XAI))
Show Figures

Figure 1

Figure 1
<p>Interpretability of a model vs. explainability of a prediction.</p>
Full article ">Figure 2
<p>The process steps of the reasoning methods.</p>
Full article ">Figure 3
<p>Major explanatory components (stacked bar) and their potential role in a scale of explanation.</p>
Full article ">
31 pages, 4782 KiB  
Article
Explainable Artificial Intelligence for Human Decision Support System in the Medical Domain
by Samanta Knapič, Avleen Malhi, Rohit Saluja and Kary Främling
Mach. Learn. Knowl. Extr. 2021, 3(3), 740-770; https://doi.org/10.3390/make3030037 - 19 Sep 2021
Cited by 87 | Viewed by 11421
Abstract
In this paper, we present the potential of Explainable Artificial Intelligence methods for decision support in medical image analysis scenarios. Using three types of explainable methods applied to the same medical image data set, we aimed to improve the comprehensibility of the decisions [...] Read more.
In this paper, we present the potential of Explainable Artificial Intelligence methods for decision support in medical image analysis scenarios. Using three types of explainable methods applied to the same medical image data set, we aimed to improve the comprehensibility of the decisions provided by the Convolutional Neural Network (CNN). In vivo gastral images obtained by a video capsule endoscopy (VCE) were the subject of visual explanations, with the goal of increasing health professionals’ trust in black-box predictions. We implemented two post hoc interpretable machine learning methods, called Local Interpretable Model-Agnostic Explanations (LIME) and SHapley Additive exPlanations (SHAP), and an alternative explanation approach, the Contextual Importance and Utility (CIU) method. The produced explanations were assessed by human evaluation. We conducted three user studies based on explanations provided by LIME, SHAP and CIU. Users from different non-medical backgrounds carried out a series of tests in a web-based survey setting and stated their experience and understanding of the given explanations. Three user groups (n = 20, 20, 20) with three distinct forms of explanations were quantitatively analyzed. We found that, as hypothesized, the CIU-explainable method performed better than both LIME and SHAP methods in terms of improving support for human decision-making and being more transparent and thus understandable to users. Additionally, CIU outperformed LIME and SHAP by generating explanations more rapidly. Our findings suggest that there are notable differences in human decision-making between various explanation support settings. In line with that, we present three potential explainable methods that, with future improvements in implementation, can be generalized to different medical data sets and can provide effective decision support to medical experts. Full article
(This article belongs to the Special Issue Advances in Explainable Artificial Intelligence (XAI))
Show Figures

Figure 1

Figure 1
<p>XAI helping medical professionals in decision-making.</p>
Full article ">Figure 2
<p>Basic concepts of XAI.</p>
Full article ">Figure 3
<p>Pipeline of post hoc explainable tool LIME.</p>
Full article ">Figure 4
<p>Pipeline of post hoc explainable tool SHAP.</p>
Full article ">Figure 5
<p>CIU explainable method [<a href="#B5-make-03-00037" class="html-bibr">5</a>].</p>
Full article ">Figure 6
<p>Workflow of the proposed method.</p>
Full article ">Figure 7
<p>CNN model.</p>
Full article ">Figure 8
<p>Model’s accuracy and loss.</p>
Full article ">Figure 9
<p>Used image data set. Validation part of the bleeding and non-bleeding images.</p>
Full article ">Figure 10
<p>LIME explanations.</p>
Full article ">Figure 11
<p>A few examples of SHAP explanations produced from the same input data as in the cases of LIME and CIU.</p>
Full article ">Figure 12
<p>CIU explanations.</p>
Full article ">Figure 13
<p>LIME explanations in the user study.</p>
Full article ">Figure 14
<p>SHAP explanations in the user study.</p>
Full article ">Figure 15
<p>CIU explanations in the user study.</p>
Full article ">Figure 16
<p>User study design.</p>
Full article ">Figure 17
<p>Incorrect LIME explanations in the user study.</p>
Full article ">Figure 18
<p>Incorrect SHAP explanations in the user study.</p>
Full article ">Figure 19
<p>Incorrect CIU explanations in the user study.</p>
Full article ">
47 pages, 6520 KiB  
Article
Classification of Explainable Artificial Intelligence Methods through Their Output Formats
by Giulia Vilone and Luca Longo
Mach. Learn. Knowl. Extr. 2021, 3(3), 615-661; https://doi.org/10.3390/make3030032 - 4 Aug 2021
Cited by 92 | Viewed by 17535
Abstract
Machine and deep learning have proven their utility to generate data-driven models with high accuracy and precision. However, their non-linear, complex structures are often difficult to interpret. Consequently, many scholars have developed a plethora of methods to explain their functioning and the logic [...] Read more.
Machine and deep learning have proven their utility to generate data-driven models with high accuracy and precision. However, their non-linear, complex structures are often difficult to interpret. Consequently, many scholars have developed a plethora of methods to explain their functioning and the logic of their inferences. This systematic review aimed to organise these methods into a hierarchical classification system that builds upon and extends existing taxonomies by adding a significant dimension—the output formats. The reviewed scientific papers were retrieved by conducting an initial search on Google Scholar with the keywords “explainable artificial intelligence”; “explainable machine learning”; and “interpretable machine learning”. A subsequent iterative search was carried out by checking the bibliography of these articles. The addition of the dimension of the explanation format makes the proposed classification system a practical tool for scholars, supporting them to select the most suitable type of explanation format for the problem at hand. Given the wide variety of challenges faced by researchers, the existing XAI methods provide several solutions to meet the requirements that differ considerably between the users, problems and application fields of artificial intelligence (AI). The task of identifying the most appropriate explanation can be daunting, thus the need for a classification system that helps with the selection of methods. This work concludes by critically identifying the limitations of the formats of explanations and by providing recommendations and possible future research directions on how to build a more generally applicable XAI method. Future work should be flexible enough to meet the many requirements posed by the widespread use of AI in several fields, and the new regulations. Full article
(This article belongs to the Special Issue Advances in Explainable Artificial Intelligence (XAI))
Show Figures

Figure 1

Figure 1
<p>Diagrammatic view of how an explainable artificial intelligence (XAI) solution is typically constructed.</p>
Full article ">Figure 2
<p>Classification of XAI methods into a hierarchical system.</p>
Full article ">Figure 3
<p>Distribution of XAI methods by output format across scope, stage, input type and problem type categories.</p>
Full article ">Figure 4
<p>Distribution of the scientific articles of the XAI literature split by output format and learning approach.</p>
Full article ">Figure 5
<p>Examples of numerical explanations generated by three model-agnostic XAI methods that highlight the contribution of the input features to the model’s prediction and can be presented to the users as (<b>a</b>) surface charts (GSA [<a href="#B44-make-03-00032" class="html-bibr">44</a>]); or (<b>b</b>) bar plots (Explain and Ime [<a href="#B42-make-03-00032" class="html-bibr">42</a>]).</p>
Full article ">Figure 6
<p>Examples of numerical explanation generated by a method for the explainability of neural networks showing the contribution of the most relevant features to the network’s predictions: (<b>a</b>) Concept Activation Vectors [<a href="#B34-make-03-00032" class="html-bibr">34</a>]; and (<b>b</b>) contextual importance and utility [<a href="#B59-make-03-00032" class="html-bibr">59</a>].</p>
Full article ">Figure 7
<p>Examples of rule-based explanations generated by model-agnostic methods which can be visualised as: (<b>a</b>) G-REX [<a href="#B69-make-03-00032" class="html-bibr">69</a>], a decision tree, or (<b>b</b>) Anchor [<a href="#B28-make-03-00032" class="html-bibr">28</a>], a list of rules accompanied by textual and visual examples.</p>
Full article ">Figure 8
<p>Examples of rule-based explanations generated by XAI methods for neural networks and visualised as (<b>a</b>) decision trees (Decision Tree Extraction [<a href="#B86-make-03-00032" class="html-bibr">86</a>]), or (<b>b</b>) by showing the most relevant input (Word Importance Scores [<a href="#B94-make-03-00032" class="html-bibr">94</a>]).</p>
Full article ">Figure 9
<p>Examples of rule-based explanations generated by ante hoc XAI methods aiming to make rule-based inference systems transparent by construction: (<b>a</b>) Bayesian rule lists [<a href="#B102-make-03-00032" class="html-bibr">102</a>]; and (<b>b</b>) fuzzy inference systems [<a href="#B117-make-03-00032" class="html-bibr">117</a>].</p>
Full article ">Figure 10
<p>Examples of textual explanation generated by InterpNET [<a href="#B125-make-03-00032" class="html-bibr">125</a>], a XAI method for neural networks that utilises their activation values to extract the most significant input features and translates them into a statement.</p>
Full article ">Figure 11
<p>Examples of visual explanations generated by model-agnostic methods such as (<b>a</b>) Explanation Graph [<a href="#B143-make-03-00032" class="html-bibr">143</a>] graphs; or (<b>b</b>) RSRS [<a href="#B137-make-03-00032" class="html-bibr">137</a>] restricted support regions and heat-maps.</p>
Full article ">Figure 12
<p>Examples of visual explanations generated by XAI methods for neural networks as salient masks (<b>a</b>,<b>b</b>) (Guided BP [<a href="#B157-make-03-00032" class="html-bibr">157</a>], Twin-systems [<a href="#B171-make-03-00032" class="html-bibr">171</a>]) and scatter-plots (<b>c</b>,<b>d</b>) (PCA [<a href="#B177-make-03-00032" class="html-bibr">177</a>], t-SNE maps [<a href="#B178-make-03-00032" class="html-bibr">178</a>]).</p>
Full article ">Figure 13
<p>Examples of miscellaneous visual explanations generated by XAI methods for neural networks. Some methods modify the input images by removing parts to check the network’s reaction—(<b>a</b>) GAN Dissection [<a href="#B182-make-03-00032" class="html-bibr">182</a>] or by maximising the activation of a given hidden neuron with respect to each pixel—(<b>b</b>) Activation Max [<a href="#B187-make-03-00032" class="html-bibr">187</a>]. Another alternative is to highlight the most relevant words of the input text—(<b>c</b>) Cell Activation [<a href="#B202-make-03-00032" class="html-bibr">202</a>] or to display the network’s structure as a graph—(<b>d</b>) Data-Flow graphs [<a href="#B24-make-03-00032" class="html-bibr">24</a>].</p>
Full article ">Figure 14
<p>Examples of mixed explanations generated by model-agnostic XAI methods which consists of a combination of visual and textual explanations in (<b>a</b>) Rivelo [<a href="#B215-make-03-00032" class="html-bibr">215</a>] interactive interfaces; or (<b>b</b>) MMD-critic [<a href="#B223-make-03-00032" class="html-bibr">223</a>] a selection of prototypes from the input data.</p>
Full article ">Figure 15
<p>Examples of mixed explanations, consisting of combinations of heatmaps and textual captions, generated by XAI methods for neural networks, which highlight the most relevant parts of the input images. (<b>a</b>) Attention Alignment [<a href="#B2-make-03-00032" class="html-bibr">2</a>]; (<b>b</b>) PJ-X [<a href="#B225-make-03-00032" class="html-bibr">225</a>]; (<b>c</b>) Attention Mechanism [<a href="#B35-make-03-00032" class="html-bibr">35</a>].</p>
Full article ">Figure 16
<p>Summary of the pros and cons associated to each explanation format, namely numeric, rules, textual, visual and mixed explanations.</p>
Full article ">Figure 17
<p>Diagram of the factors affecting the selection of XAI methods.</p>
Full article ">
17 pages, 1553 KiB  
Article
Deterministic Local Interpretable Model-Agnostic Explanations for Stable Explainability
by Muhammad Rehman Zafar and Naimul Khan
Mach. Learn. Knowl. Extr. 2021, 3(3), 525-541; https://doi.org/10.3390/make3030027 - 30 Jun 2021
Cited by 137 | Viewed by 12444
Abstract
Local Interpretable Model-Agnostic Explanations (LIME) is a popular technique used to increase the interpretability and explainability of black box Machine Learning (ML) algorithms. LIME typically creates an explanation for a single prediction by any ML model by learning a simpler interpretable model (e.g., [...] Read more.
Local Interpretable Model-Agnostic Explanations (LIME) is a popular technique used to increase the interpretability and explainability of black box Machine Learning (ML) algorithms. LIME typically creates an explanation for a single prediction by any ML model by learning a simpler interpretable model (e.g., linear classifier) around the prediction through generating simulated data around the instance by random perturbation, and obtaining feature importance through applying some form of feature selection. While LIME and similar local algorithms have gained popularity due to their simplicity, the random perturbation methods result in shifts in data and instability in the generated explanations, where for the same prediction, different explanations can be generated. These are critical issues that can prevent deployment of LIME in sensitive domains. We propose a deterministic version of LIME. Instead of random perturbation, we utilize Agglomerative Hierarchical Clustering (AHC) to group the training data together and K-Nearest Neighbour (KNN) to select the relevant cluster of the new instance that is being explained. After finding the relevant cluster, a simple model (i.e., linear model or decision tree) is trained over the selected cluster to generate the explanations. Experimental results on six public (three binary and three multi-class) and six synthetic datasets show the superiority for Deterministic Local Interpretable Model-Agnostic Explanations (DLIME), where we quantitatively determine the stability and faithfulness of DLIME compared to LIME. Full article
(This article belongs to the Special Issue Advances in Explainable Artificial Intelligence (XAI))
Show Figures

Figure 1

Figure 1
<p>A block diagram of the LIME framework.</p>
Full article ">Figure 2
<p>A high level block diagram of the DLIME framework.</p>
Full article ">Figure 3
<p>Dendrograms of binary and multi-class datasets.</p>
Full article ">Figure 4
<p>Explanations for neural network generated by DLIME (Linear) and LIME, and respective Jaccard distances over 10 iterations. Highlighted features with yellow color in (<b>b</b>,<b>d</b>) represents the difference in selected features for the same instance over 2 iterations. The order of features in (<b>a</b>–<b>d</b>) is higher to lower importance. (<b>e</b>,<b>f</b>) shows the Jaccard distance matrix among the features selected over 10 iterations.</p>
Full article ">Figure 5
<p>Explanations generated for neural network by DLIME-Tree and LIME, and respective Jaccard distances over 10 iterations. Features outlined with red color in (<b>a</b>,<b>c</b>) represents insignificant features with 0 contribution. Highlighted features with yellow color in (<b>b</b>,<b>d</b>) represents the difference in selected features for the same instance over 2 iterations. The order of features in (<b>a</b>–<b>d</b>) is higher to lower importance. (<b>e</b>,<b>f</b>) shows the Jaccard distance matrix among the features selected over 10 iterations.</p>
Full article ">

Other

Jump to: Research

31 pages, 1054 KiB  
Systematic Review
XAIR: A Systematic Metareview of Explainable AI (XAI) Aligned to the Software Development Process
by Tobias Clement, Nils Kemmerzell, Mohamed Abdelaal and Michael Amberg
Mach. Learn. Knowl. Extr. 2023, 5(1), 78-108; https://doi.org/10.3390/make5010006 - 11 Jan 2023
Cited by 51 | Viewed by 21014
Abstract
Currently, explainability represents a major barrier that Artificial Intelligence (AI) is facing in regard to its practical implementation in various application domains. To combat the lack of understanding of AI-based systems, Explainable AI (XAI) aims to make black-box AI models more transparent and [...] Read more.
Currently, explainability represents a major barrier that Artificial Intelligence (AI) is facing in regard to its practical implementation in various application domains. To combat the lack of understanding of AI-based systems, Explainable AI (XAI) aims to make black-box AI models more transparent and comprehensible for humans. Fortunately, plenty of XAI methods have been introduced to tackle the explainability problem from different perspectives. However, due to the vast search space, it is challenging for ML practitioners and data scientists to start with the development of XAI software and to optimally select the most suitable XAI methods. To tackle this challenge, we introduce XAIR, a novel systematic metareview of the most promising XAI methods and tools. XAIR differentiates itself from existing reviews by aligning its results to the five steps of the software development process, including requirement analysis, design, implementation, evaluation, and deployment. Through this mapping, we aim to create a better understanding of the individual steps of developing XAI software and to foster the creation of real-world AI applications that incorporate explainability. Finally, we conclude with highlighting new directions for future research. Full article
(This article belongs to the Special Issue Advances in Explainable Artificial Intelligence (XAI))
Show Figures

Figure 1

Figure 1
<p>Classification of AI models according to their level of complexity, explainability, and their potential in modern AI applications.</p>
Full article ">Figure 2
<p>Addressed research questions (RQ) aligned to the XAI software development process.</p>
Full article ">Figure 3
<p>Explainable AI (XAI) and possible stakeholders.</p>
Full article ">Figure 4
<p>Results of the quantitative analysis. (<b>a</b>) Research focus. (<b>b</b>) Application domains. (<b>c</b>) Time distribution.</p>
Full article ">Figure 5
<p>Classification of the reported XAI design methods.</p>
Full article ">Figure 6
<p>Examples of the visual-based XAI methods. (<b>a</b>) PDP &amp; ICE curves. (<b>b</b>) ALE curves.</p>
Full article ">
Back to TopTop