[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
Addressing Semantic Variability in Clinical Outcome Reporting Using Large Language Models
Previous Article in Journal
Improvement of Statistical Models by Considering Correlations among Parameters: Local Anesthetic Agent Simulator for Pharmacological Education
Previous Article in Special Issue
Abdominal MRI Unconditional Synthesis with Medical Assessment
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Part-Prototype Models in Medical Imaging: Applications and Current Challenges

by
Lisa Anita De Santi
1,2,
Franco Italo Piparo
1,
Filippo Bargagna
1,2,
Maria Filomena Santarelli
3,
Simona Celi
2 and
Vincenzo Positano
2,*
1
Department of Information Engineering, University of Pisa, 56122 Pisa, Italy
2
Bioengineering Unit, Fondazione Toscana G. Monasterio, 56124 Pisa, Italy
3
CNR Institute of Clinical Physiology, 56124 Pisa, Italy
*
Author to whom correspondence should be addressed.
BioMedInformatics 2024, 4(4), 2149-2172; https://doi.org/10.3390/biomedinformatics4040115
Submission received: 6 September 2024 / Revised: 21 October 2024 / Accepted: 23 October 2024 / Published: 28 October 2024
(This article belongs to the Special Issue Advances in Quantitative Imaging Analysis: From Theory to Practice)

Abstract

:
Recent developments in Artificial Intelligence have increasingly focused on explainability research. The potential of Explainable Artificial Intelligence (XAI) in producing trustworthy computer-aided diagnosis systems and its usage for knowledge discovery are gaining interest in the medical imaging (MI) community to support the diagnostic process and the discovery of image biomarkers. Most of the existing XAI applications in MI are focused on interpreting the predictions made using deep neural networks, typically including attribution techniques with saliency map approaches and other feature visualization methods. However, these are often criticized for providing incorrect and incomplete representations of the black-box models’ behaviour. This highlights the importance of proposing models intentionally designed to be self-explanatory. In particular, part-prototype (PP) models are interpretable-by-design computer vision (CV) models that base their decision process on learning and identifying representative prototypical parts from input images, and they are gaining increasing interest and results in MI applications. However, the medical field has unique characteristics that could benefit from more advanced implementations of these types of architectures. This narrative review summarizes existing PP networks, their application in MI analysis, and current challenges.

1. Introduction

Medical imaging (MI) represents the main tool in the diagnosis, treatment planning, and monitoring of almost all medical diseases. In clinical practice, different imaging modalities capture different aspects of anatomical structures and physiological functions. Structural imaging, mainly performed through radiography, computed tomography (CT), structural magnetic resonance imaging (sMRI), and ultrasound (US), focuses on visualizing the anatomical structures of the body, providing detailed pictures of organs, bones, and tissues. Functional imaging measures changes in metabolism, blood flow, tissue composition, and tracer absorption. Positron emission tomography (PET) and single-photon emission computed tomography (SPECT) use isotopes as probes to reflect the spatial distribution of chemical compounds within the body. Functional magnetic resonance imaging (fMRI) can monitor physiological processes using several techniques, such as blood-oxygen-level-dependent imaging (BOLD), diffusion, and assessment of blood perfusion. For almost all image modalities, diagnoses are supported by image analysis procedures carried out by software tools that automatically or semi-automatically extract information patterns from images. More recently, Artificial Intelligence (AI) algorithms were developed to analyze medical images to identify patterns that could suggest the presence of a specific disease, helping in early disease diagnosis and treatment planning. Machine and Deep Learning (ML and DL) models have demonstrated interesting performances in supporting MI analysis, but their black-box nature raises technical, ethical, and legal concerns in this high-stakes domain. In this scenario, Explainable Artificial Intelligence (XAI) offers many opportunities to foster the development of responsible and trustworthy AI systems for healthcare applications [1,2,3,4,5].
The last few decades of ML explainability research have mainly been focused on developing methods for explaining complex black-box models (post hoc interpretability), which typically include saliency-based visual methods [2,3,4]. Saliency approaches exploit the spatial information, preserved through the convolutional layers of DL models, to analyze which parts of an image lead to a resulting decision. However, these methods are often criticized for providing incorrect and incomplete representations of the black-box models’ behaviour, and explanations that are not fully interpretable for a human user [5,6,7,8].
To address the limitations of post hoc XAI, a recent trend has emerged, proposing models intentionally designed to be self-explanatory, known as ante hoc interpretability or interpretability-by-design. The interpretability-by-design approach focuses on developing and training transparent predictors: models that are simple enough to be easily understood, yet sophisticated enough to accurately capture relationships between inputs and outputs. This is achieved by ensuring that the AI model itself is interpretable, either by limiting the number of input features (referred to as sparsity or by designing decision-making processes that can be replicated by human users [9]. This includes the use of low-complexity models, such as Linear Models, decision trees, rule-based learners, generalized additive models, and Bayesian models. The primary advantage of simple models is that they are fully understandable without needing additional interpretability techniques. However, the trade-off between transparency and performance is well documented, making it limiting to rely solely on interpretable models for AI applications [9,10,11]. Recently, authors have started proposing black-box methods augmented with explainability methods as alternatives to low-level complexity ML models, overcoming the well-known trade-off between performance and interpretability [1,4,7]. Part-prototype (PP) neural networks (also known as self-explaining neural networks) constitute a promising research direction in computer vision (CV). PP networks perform image classification based on the identification of class-representative prototypical parts in the input image, a reasoning process based on the recognition-by-components theory [8,12]. These self-explaining models reproduce the human case-based reasoning process, and their explanations are faithful representations of the computations performed by the model to make its decisions [6,8,13].
A PP neural network is generally constituted by a deep neural network backbone trained to learn prototypical parts. This feature extractor is trained only using image-level labels (no additional image sub-annotations are required). Once trained, the PP model classifies an input image by detecting the image patches that are more similar to the learned prototypical parts. The model’s interpretability is provided thanks to the transparent reasoning process implemented in the form of “this looks like that”, and the direct relation between the prototypes and the classification prevents unfaithful explanation [8,13,14,15]. A representative example of the case-based reasoning process implemented by the PP networks architectures is reported in Figure 1.
Following the XAI taxonomy [5], a PP network can provide both global (the model’s overall behaviour) and local (relative to a certain prediction) explanations by, respectively, showing all the learned prototypes and the detected prototypes for the given prediction, as shown in Figure 2.
In addition to designing explainable models, properly evaluating XAI is crucial for promoting their application in real-world scenarios [5]. The evaluation process of XAI methods is currently an open challenge, and several studies have started proposing systematic evaluation frameworks to objectively assess the system developed [1,16,17]. Aspects such as the multidisciplinarity of XAI and the absence of a gold standard for a “good explanation” make the interpretability assessment a non-trivial task, and it remains unclear how XAI methods, including PP networks, should be evaluated [5].
This narrative review summarizes existing PP networks, their application in MI analysis, and their current open challenges.

2. Foundational Part-Prototype Models

To date, there have been different PP model variants proposed in the literature [8,18], including ProtoPNet [13], XProtoNet [19], ProtoTree [20], ProtoPShare [21], ProtoPool [22], and PIPNet [8]. This section aims to summarize the main theoretical characteristics of the existing PP nets. For each network, we will describe the architecture, training process, and visualization of the prototypical parts as patches of the input images.

2.1. Part-Prototype Network (ProtoPNet)

ProtoPNet [13] was the first PP network proposed in the CV domain. The ProtoPNet architecture is composed of a (1) CNN backbone f with parameters w f ; a (2) prototype layer g p of M prototypes P = { p j = 1 M } ; and a final (3) fully connected layer h with weight w h .
We show the ProtoPNet architecture in Figure 3.
The CNN backbone takes an input image x and returns as output the extracted features z = f ( x | w f ) with dimensions H × W × D . This could be any available CNN’s feature extractor, such as pre-trained architectures like VGG-16, VGG-19, ResNet-34, ResNet-152, DenseNet-121, or DenseNet-161.
The network learns M prototypes P = { p j = 1 M } of dimensions H 1 × W 1 × D , where H 1 H , W 1 W . Every prototype p j represents an activation pattern of the convolutional output, corresponding to a certain prototypical image patch in the original pixel space. In this way, a prototype p j is the latent representation of a certain image prototypical part. The prototype layer g p performs the computation reported in Equation (1). It evaluates the squared L 2 distances between the j - th prototype p j and patches of z with the same shape as p j . These distances are inverted to produce an activation map of the similarity score, which indicates how strong a prototypical part is in the image. Finally, global max-pooling of every prototype’s activation map returns a similarity score, indicating how strongly a prototypical part is present in a patch of the input image (Equation (1)). The model requires the allocation of a pre-determined number of prototypes M k for each class k { 1 , . . . , K } , and the subset of prototypes associated with class k   P k P should capture the most relevant parts for identifying images of that class.
g p j ( z ) = m a x z ¯ p a t c h e s ( z ) l o g | | z ¯ p j | | 2 2 + 1 | | z ¯ p j | | 2 2 + ϵ
The last fully connected layer h multiplies the m similarity scores by the weight matrix w h to produce the output logits, which are finally normalized with a softmax function to obtain the class probabilities. The weight w h ( j , k ) corresponds to the connection between the output of a class prototype unit g p j and the logit of class k. Here, a positive (and, respectively, a negative) connection between prototype j and class k means that the similarity to a class prototype j increases (decreases) the probability that the image belongs to class k.
The ProtoPNet training process requires three main steps: (1) optimization of the convolutional and prototype layers; (2) prototypes’ projection; (3) convex optimization of the fully connected layer.
The backbone and prototype layers’ optimization (1) aims to learn a meaningful latent space, where the most important patches for classifying images are clustered (in L2-distance) around semantically similar prototypes of the images’ true classes and the clusters of different classes are well separated. The weights of the CNN backbone w f and the prototype layer P = { p j = 1 M } are jointly optimized using stochastic gradient descent (SGD), while the weights of the fully connected layer w h are fixed and set to
w h ( j , k ) = 1 j if p j P k 0.5 j if p j P k
In this way, the similarity to a prototype (not) of class k (decreases) increases the predicted probability that the image belongs to class k. SDG optimizes the following loss function:
m i n P , w f 1 n i = 1 n C r s E n t ( h g p f ( x i ) , y i ) + λ 1 C l s t + λ 2 S e p
where the elements are defined as follows:
  • C r s E n t ( . ) : Cross-entropy loss, which penalizes misclassification on the training data.
  • C l s t : Cluster cost, which promotes every training image to have some latent patch that is close to at least one prototype of its own class.
    C l s t = 1 n i = 1 n m i n j : p j P y i m i n z p a t c h e s ( f ( x i ) ) | | z p j | | 2 2
  • S e p : Separation cost, which promotes every latent patch of a training image to stay away from the prototypes not of its class.
    S e p = 1 n i = 1 n m i n j : p j P y i m i n z p a t c h e s ( f ( x i ) ) | | z p j | | 2 2
With the prototypes’ projection (2), every prototype p j is projected onto the nearest latent training patch of a training image of the same class. After projection, a prototype p j corresponds to some patch of the latent representation f ( x ) of an image in the training set x .
p j a r g m i n z Z j | | z p j | | 2
where Z j = { z ¯ : z ¯ p a t c h e s ( f ( x i ) ) i s . t . y i = k } .
The fully connected layer optimization (3) adjusts the connections from the similarity score and the logit of the class to enhance the sparsity:
m i n w h 1 n i = 1 n C r s E n t ( h g p f ( x i ) , y i ) + λ k = 1 K j : p j P k | w h ( k , j ) |
that is, for k and j with p j P k , we have w h ( j , k ) 0 . In doing so, we foster the model in implementing a positive reasoning process (use prototypes which add evidence for a class, not the ones whose presence decreases the evidence for the class), and further improve the accuracy without changing the learned latent space or prototypes.
Once the ProtoPNet is trained, we can visualize every prototype p j as a patch of an image x in the training set. This visualization process involves the following steps:
  • Forwarding x through a ProtoPNet to produce the activation map associated with the prototypes p j .
  • Upsampling the activation map to the dimension of the input image.
  • Localizing the smallest rectangular patch whose corresponding activation is at least as large as the 95th percentile of all the activation values in that same map.
We schematically show the visualization process in Figure 4.

2.2. XProtoNet

XProtoNet differs from ProtoPNet due to its ability to learn representative features within a dynamic area.
XProtoNet is composed of a (1) feature extractor with two main modules: a (1.a) feature module F and a (1.b) occurrence module M p k c ; a (2) similarity score measurement; and a final (3) fully connected layer.
XProtoNet takes an input image x and extracts the feature vector f p k c ( x ) for each one of the learned prototypes per class k , P k = { p j } j = 1 M k with k = { 1 , . . . , K } .
f p k c ( x ) = u M p k c ( x ) , u F u ( x )
where u [ H × W ) denotes the spatial location of F ( x ) and M p k c ( x ) . The feature module extracts the image latent representation F ( x ) R H × W × D while the occurrence module predicts an occurrence map for each prototype M p k c ( x ) R H × W , which indicates where every prototype is likely to appear. The feature vector f p k c ( x ) then represents a certain feature in the highly activated area of the occurrence map.
Then, XProtNet uses the cosine similarity to compute a similarity score between the feature of the input image to every prototype:
s ( x , p k c ) = f p k c ( x ) · p k c ( x ) | | f p k c ( x ) | | | | p k c ( x ) | |
Finally, every prototype p k c contributes to the prediction score of the related class with an importance determined by the weights of the linear layer w p k c .
p ( y c | x ) = σ p k c P c w p k c s ( x , p k c )
where σ is a sigmoid activation function.
XProtoNet follows the ProtoPNet training scheme (backbone and prototype optimization; prototype projection; fully connected layer optimization). Its cost function is composed of three different terms:
L t o t a l = L c l s + λ c l s t L c l s t + λ s e p L s e p + λ o c c u r L o c c u r
which is defined as follows:
  • Classification weighted loss, L c l s c , is used to address the imbalance in the dataset:
    L c l s c = i 1 | N p o s c | ( 1 p i c ) γ y i c l o g ( p i c ) i 1 | N n e g c | ( p i c ) γ ( 1 y i c ) l o g ( 1 p i c )
    where p i c = p ( y c | x i ) is the prediction score of the i - th sample x i , γ is a parameter for class balance, | N n e g c | and | N p o s c | are the numbers of negative and positive labels on disease c, and y i c { 0 , 1 } is the target label of x i on disease c.
  • Regularization for interpretability which, similarly to [13], includes two different terms, L c l s t c and L s e p c , to, respectively, maximize the similarity between x and p k c for positive samples and minimize it for negative samples:
    L c l s t c = y c m a x p k c P c s ( x , p k c )
    L s e p c = ( 1 y c ) m a x p k c P c s ( x , p k c )
  • Regularization for the occurrence map:
    L o c c u r c = L t r a n s c + p k c P c | | M p k c ( x ) | | 1
    which is defined as follows:
    -
    The term L t r a n s c considers that an affine transformation A ( . ) of an image does not change the relative location, so it should not affect the occurrence map either.
    L t r a n s c = p k c P c | | A ( M p k c ( x ) ) M p k c ( A ( x ) ) | | 1
    -
    The term p k c P c | | M p k c ( x ) | | 1 regularizes to have an occurrence area as small as possible to not include unnecessary regions.
Once the model is trained, the prototype p k c is replaced with the most similar feature vector f p k c from the training images.
The learned prototypes are visualized as follows:
  • Upsampling the occurrence maps to the input image size.
  • Normalizing with the maximum value of the upsampled mask.
  • Marking with contour the occurrence values that are greater than a factor of 0.3 of the maximum intensity.

2.3. Neural Prototype Tree (ProtoTree)

Nauta et al. [20] developed Neural Protype Tree (ProtoTree), which integrates the prototype learning approach into a hierarchical decision tree structure. The model includes a CNN followed by a binary tree structure, where each node corresponds to a prototypical part. The prototypical parts are tensors that can be visualized as a patch of a training sample learned with backpropagation during the training process without requiring additional annotations. Similarly to the PP network, the ProtoTree architecture itself explains the global reasoning process implemented (here, as a hierarchical sequence of steps), and the local explanations are constituted by the route through the tree for every single prediction.
ProtoTree consists of a (1) CNN backbone f with parameters w f ; a (2) binary soft decision tree constituted of N nodes, L leaves, and E edges, which takes the image latent representation and returns the class probability distribution over the K classes.
The CNN backbone takes the input image x and extracts the latent features z = f ( x ; w f ) consisting of D 2-dim ( H × W ) feature maps, ( H × W × D ) .
The binary soft decision tree takes as input the image latent representation z and returns the class probability distribution over K classes y ^ . Each internal node n N has two children, n . r i g h t and n . l e f t , connected, respectively, with e ( n , n . r i g h t ) E e ( n , n . l e f t ) E and corresponds to a prototype p n P . Here, a prototype is a trainable tensor of dimensions H 1 × W 1 × D , with the same depth as the convolutional output z and H 1 H , W 1 W . The prototype p n acts as a kernel, which slides over z , computes the Euclidean distance between p n and its receptive field z ˜ , and applies a minimum pooling to select the patch H 1 × W 1 × D closest to p n :
z ˜ * = a r g m i n z ˜ | | z ˜ p n | |
The selected closest patch is then routed through both edges (soft routing) of the child nodes with
p e ( n , n . r i g h t ) ( z ) = e x p ( | | z ˜ * p n | | )
p e ( n , n . l e f t ) = 1 p e ( n , n . r i g h t )
So, the similarity between p n and z ˜ * determines to which extent z is routed to the right node child n . r i g h t . Each leaf node l L receives the convolutional output z with a probability given by the product of the edge probabilities p e in the path P l followed by z from the root node to leaf l:
π l ( z ) = e P l p e ( z )
Each leaf has a trainable parameter c l , which denotes the class probability distribution over K classes and computes the class probability distribution of leaf l applying the softmax normalization σ ( c l ) .
The class probability distribution over K classes y ^ is given by the weighted contribution of all the leaves.
y ^ ( x ) = l L σ ( c l ) · π l ( f ( x ; w ) )
The ProtoTree’s architecture is initialized by selecting a pre-trained CNN backbone f ( x ; w ) and defining the maximum height h of the binary tree, which determines the number of prototypes | P | = 2 h 1 . Then, the training process is constituted by the (1) optimization of the convolutional and prototype layers; (2) convex optimization of the decision tree’s leaves; and (3) prototypes’ projection.
The CNN backbone w f and prototype layers’ P are optimized (1) with backpropagation gradient descent by minimizing the cross-entropy loss between the predicted class probability distribution y ^ and ground-truth y .
The convex optimization of c (2) learns the leaves’ distribution using a derivative-free approach:
c l ( t + 1 ) = 1 B · c l ( t ) + x , y T ( σ ( c l ( t ) ) y π l ) y ^
where t indexes a training epoch.
Finally, with the prototypes’ projection (3) p n z ˜ n * , every prototype p n P is replaced with its nearest latent patch present in the training data z n * :
z n * = a r g m i n z { f ( x ) } , x T | | z * p n | |
Once trained, ProtoTree has a pruning step to remove ineffective prototypes; this reduces the explanation’s sizes, improving the model’s interpretability. This consists of removing leaves with nearly uniform distributions, signs of little discriminative power between classes, by defining a threshold τ slightly greater than 1 K and pruning the ones where m a x ( σ ( c l ) ) τ .
Considering x n * as the training image corresponding to the latent patch z ˜ * , the prototype p n is visualized as a patch of x n * as follows:
  • Forwarding x n * through f : z = f ( x n * ) .
  • Creating a 2-dim similarity map:
    S n ( i , j ) = e x p ( | | | z ˜ ( i , j ) p n | | )
    where ( i , j ) denotes the location of patch z ˜ in patches of z.
  • Upsampling S n with bicubic interpolation to the shape of x n * .
  • Visualizing p n as a rectangular patch at the same location nearest to the latent patch z ˜ n * .z

2.4. Protypical Part Shared Network (ProtoPShare)

Rymarczyk et al. [21] developed ProtoPShare, a part-prototype network that extended ProtoPNet, addressing two of its limitations: the (1) ability to share prototypes between the classes; the (2) ability to identify semantically similar prototypes, even with a distant representation in the latent space. The authors achieved this by implementing a data-dependent pruning algorithm based on the feature maps. This results in a model with a reduced explanation size, fostering its interpretability.
ProtoPShare shared the same architecture as that of ProtoPNet, constituted of a CNN backbone f extracting the latent representation of the image z = f ( x ) ; a prototype layer g of m k prototypes per classes P k = { p j j = 1 m k } ; and a fully connected layer of weight w h , which predicts the output class.
ProtoPShared adopted the ProtoPNet training scheme followed by a data-dependent merge-pruning to obtain a network with a smaller number of shared prototypes. The pruning process consists of the following steps:
  • Computing the data-dependent similarity for pair of prototypes ( p , p ˜ ) P 2 , given by the compliance on the similarity scores for all the training input image x X . This considers two prototypes similar if they activate alike on the training images, even if far in the latent space:
    d D D ( p , p ˜ ) = 1 x X ( g p i ( z ) g p ˜ i ( z ) ) 2
  • Selecting a percentage ζ of the most similar pairs of prototypes to merge per step ( p , p ˜ ) .
  • For each pair, removing prototype p and its weights w h ( p ) and reusing prototype p ˜ , aggregating weights w h ( p ˜ ) and w h ( p ) .
The results obtained compared to other self-explained prototypical part models suggested an increase in interpretability and a novel ability to discover semantic similarity discovery while maintaining high accuracy.
As with ProtoPNet [13], the visualization process of the prototypes p j is based on the localization of the smallest and most activated rectangular patch in the upsampled image activation map.

2.5. ProtoPool

Rymarczyk et al. [22] developed ProtoPool, integrating into the PP network (1) the concept of prototype soft assignment, which optimizes the model’s compactness without requiring a pruning stage in the training process, and (2) the definition of a focal similarity function to focus the model on salient features. As in ProtoPShare [21], the prototypes can be shared between classes.
ProtoPool shared the same architecture as that of ProtoPNet, constituting a CNN backbone f; a prototype layer g of m prototypes P = { p j = 1 M R D } ; and a fully connected layer of weight w h , which predicts the output class.
The image latent representation z = f ( x ) with the dimensions H × W × D is considered a set of H W ˙ vectors, each one corresponding to a location in the image and with dimensions D: Z x = { z i f ( x ) : z i R D , i = 1 , . . . , H W ˙ } .
The prototype layer contains K slots for each class, and each slot is implemented as a distribution of the prototypes in the pool q k R M , where q k is the probability of assigning successive prototypes to the slot. The layer computes on each k slot the aggregated similarity between Z x and all the prototypes, considering their slot distribution q k :
g k = i = 1 M q k i g p i
Here, ProtoPool computes the activation of prototype p with respect to image x g p using the novel introduced focal similarity function:
g p = m a x z Z x g p ( z ) m e a n z Z x g p ( z )
where
g p ( z ) = l o g 1 + 1 | | z p | | 2
The focal similarity has the advantage of preventing high activation due to all the elements in Z x being similar to a prototype, which might result in the obtainment of prototypes focused on the background regions, and the gradient is passed only through the most active part of the image. ProtoPool also uses the concept of soft assignment on prototype distributions, applying the Gumbel-softmax estimator to avoid many prototypes being assigned to one slot (which might decrease the interpretability):
Gumbel - softmax ( q , τ ) = ( y 1 , . . . , y M ) R M
where
y i = e x p ( q i + η i ) / τ m = 1 M e x p ( ( q m + η m ) / τ )
where e t a m for m 1 , . . . , M are sampled from the standard Gumbel distribution.
ProtoPool adopted the ProtoPNet training scheme (backbone and prototype training; prototype projection; backbone, prototype and fully connected layer fine-tuning) by extending the loss function, avoiding the same prototype being assigned to many slots of one class:
L o r t h = i < j K q i , q j | | q i | | 2 · | | q j | | 2
As in ProtoPNet, the prototypes are projected to replace every learned abstract prototype with the representation of the nearest training patch:
p a r g m i n z Z C | | z p | | 2
where Z C = { z : z Z x ( x , y ) : y C } ; here, differently from ProtoPNet, C is the set of all the classes assigned to prototype p.
Even for ProtoPool, the visualization process of the prototypes is based on the upsampling of the activation maps, as in ProtoPNet [13].

2.6. Patch-Based Intuitive Prototype Network (PIPNet)

Nauta et al. [8] developed a part-prototype architecture that presents different advantages compared to the existing models: (1) the absence of a semantic gap between the learned prototypes and human vision; (2) compactness; and (3) the ability to handle Out-of-Distribution (OoD) data.
The lack of a semantic gap means that the prototypes learned correlate with human concepts, ground-truth object parts, and human visual perception. This was covered by implementing an extra regularization of the prototypes using Self-supervised Representation Learning. PIPNet overcomes the assumption that some of the images from the same class have the same prototypes (the regularization of interpretability at the class level).
The weights in the fully connected layer connecting the prototypes to classes are trained with a sparsity regularization, which promotes the model’s compactness. PIPNet only needs an upper bound on the number of prototypes, selecting as few prototypes as possible for good classification accuracy and allowing class-sharable prototypes. This regularization was performed by introducing a novel function that optimizes the classification performance and compactness.
Finally, PIPNet manages OoD data by abstaining from decisions when no relevant prototypes are present in the image. This was implemented enabling the assignment of near-zero scores for all classes using normalization (necessary for scale-invariance issues) in logits, where a score of zero stays the same and the prototype presence scores are allowed to behave independently of each other.
PIPNet consists of a (1) CNN backbone f with parameters w f followed by a softmax activation; a (2) global max-pooling operation extracting the prototype layer p ; and a final positive sparse (3) fully connected layer w c , which acts as a scoring sheet system to predict the output classes. PIPNet classifies images based on the presence of prototypical parts in the input image. The relevantly present prototypical parts add up scores (evidence) of the model’s classes with a proportional contribution determined by the linear weights, and the model can abstain from making decisions when there is not enough evidence for any classes.
The CNN backbone takes the input image x and extracts the latent features z = f ( x ; w f ) consisting of D 2-dim ( H × W ) feature maps, ( H × W × D ) . The softmax activation normalizes so that z = d = 1 D z h , w , d = 1 . Here, z h , w , d represents the probability that a patch in position h , w H × W corresponds to the prototype d, that is, the one-hot encoding of patch h , w to the prototype d.
The global max-pooling 2D operation extracts D prototypes and calculates the prototype’s presence scores p , where p [ 0 . , 1 . ] D , and where p j measures the presence of the prototype d in the input image.
Finally, a linear sparse classification layer with positive weights w c R 0 D × K connects the prototypes to the classes acting as scoring sheet systems: o = p · w c where o is 1 × K , and o k = d = 1 D p d w c d , k , where w c d , k represents the contribution of prototype d for class k.
The PIPNet training process is composed of two main steps: the (1) Self-Supervised Pre-Training of Prototypes and the (2) PIPNet training.
The prototypes’ pre-training aims to learn an image encoding p with semantic similarity independently from the classification task (the last linear layer is kept frozen). We optimize for patch alignment under a contrastive learning approach by training PIPNet to assign the same prototypes to two views of an augmented image patch.
This involves a positive pair creation x , x step by applying data augmentation transformation to the input image x selected so that humans consider the two views to be similar. The Adam optimizer optimizes the following loss function:
λ A L A + λ T L T
where
  • Alignment Loss, L A , optimizes for near-binary encodings, where an image patch corresponds to exactly one prototype:
    L A = 1 H W ( h , w ) H × W l o g ( z h , w , : · z h , w , : )
    where the dot product z h , w , : · z h , w , : assesses the similarity between the latent patches of two views of an image patch if z h , w , : = z h , w , : L A = 0 .
  • Tanh Loss, L T prevents the trivial solution in which one prototype node is activated on all image patches in each image in the dataset forcing every prototype to be present at least once in a mini-batch:
    L T ( p ) = 1 D d = 1 D l o g ( ( t a n h ( b = 1 B ) p b ) + ϵ )
The PIPNet training optimizes the classification performances and fine-tunes the prototypes for the downstream classification task.
λ A L A + λ T L T + λ C L C
where the Classification Loss, L C , is the log-likelihood loss between the ground-truth labels and the predictions. During training, the output scores are computed as o = log ( ( p w c ) 2 + 1 ) , acting as regularization for sparsity.
The visualization process of the prototypes p j is based on the upsampling of the latent output based on the single most activated latent patch.

3. Application and Advances in Medical Imaging

There is increasing interest in applying the PP model in medical imaging.
Most of the existing PP nets were originally designed for general CV purposes, so they typically work by taking RGB images as the input. However, medical imaging diagnosis is often performed using the entire volumetric human anatomy, so being able to efficiently process 3D data volumes might be particularly relevant in this context. In addition, clinicians often use multiple data sources in their decision-making process [23,24], and some authors have started developing explainable models that use multiple data sources to predict outcomes.
We searched major databases (Google Scholar, Scopus, Web of Science) for articles including original research papers and preprints on PP networks applied in medical image classification published from 2019 to August 2024. We described all the collected works, grouped according to the input data type (2D images, 3D images, multimodal data), in the following sections and summarize them in Table 1.

3.1. Two-Dimensional Image Models

Singh et al. proposed three ProtoPNet variants for classifying chest images as COVID-19 positive, pneumonia-positive, or normal: Negative-Positive ProtoPNet (NP-ProtoPNet) [25], Generalized ProtoPNet (Gen-ProtoPNet) [26], and Pseudo-ProtoPNet (Ps-ProtoPNet) [27].
In NP-ProtoPNet [25], the authors modified the original fully connected layer by connecting the similarity scores to correct/incorrect classes’ logits with weights fixed, respectively, to 1 or −1. This allows the network to reason both in a “positive” way (“this looks like that”) and in a “negative” way (“this does not look like those”), i.e., by exclusion. In Gen-ProtoPNet [26], the authors introduced a generalized distance function to select prototypes of varying dimensions. They achieved accuracies of 88.99% in the first work and 87.27% in the second. They also trained black-box CNNs for the same task, showing that their explainable model performances are comparable with state-of-the-art (SOTA) black-box models. For both studies, they used a combination of two datasets, one containing frontal chest radiographs of healthy people and those of pneumonia patients (Chest X-ray [28]), and the COVID-19 Image Data Collection [29], from which they took radiographs of COVID-19 patients. The Ps-ProtoPNet [27] integrates NP-ProtoPNet and Gen-ProtoPNet. For this implementation, they used CT images from the COVIDx CT-2 Dataset available on Kaggle, achieving an accuracy of 99.24%.
Kim et al. [19] implemented X-ProtoPNet, which introduces the prediction of occurrence maps, which indicate the area where a sign of the disease (i.e., a prototype) is likely to appear, so it compares the features in the predicted area with the prototypes. This novelty can help to identify whether the model is focusing on the correct parts of the image or if it is being misled by irrelevant features. They used the NIH chest X-ray dataset [30] for a multilabel classification task (atelectasis, cardiomegaly, effusion, infiltration, mass, nodule, pneumonia, pneumothorax consolidation, edema, emphysema, fibrosis, pleural thickening, and hernia), achieving a mean Area Under the Receiver Operating Characteristic (AUROC) of 0.822.
Mohammadjafari et al. [31] applied ProtoPNet to brain MRI scans to detect Alzheimer’s disease, achieving an accuracy of 87.17% in the OASIS dataset [32] and of 91.02% in the ADNI dataset. They used different CNN architectures as feature extractors, and ProtoPNet’s performances were demonstrated to be comparable to or slightly less than those of its black-box counterparts.
Barnett et al. [33] designed a novel PP net modifying the training process of ProtoPNet with a loss function including fine-grade expert image annotation and using top-k average pooling instead of max-pooling. They trained their model on an internal dataset of digital mammograms for mass margin classification and mass margin malignancy prediction and reported an equal or higher accuracy compared to that of PrototPNet and a black-box model.
Carloni et al. [34] used ProtoPNet to classify benign/malignant breast masses in mammogram images from a publicly available dataset (CBIS-DDSM [35]), achieving an accuracy of 68.5%. While this performance may not yet be ideal for clinical practice, the authors suggest three tasks for a qualitative evaluation of the explanations by a radiologist.
Amorim et al. [36] implemented ProtoPNet to classify histologic patches from the PatchCamelyon dataset [37] into benign and malignant using top-k average pooling instead of max-pooling to extract similarity scores from the activation maps. The authors achieved an accuracy of 98.14% by using Densenet-121 as a feature extractor. This performance was comparable with the one obtained with the black-box CNN.
Flores-Araiza et al. [38] used ProtoPNet to identify types of kidney stones (whewellite, weddellite, anhydrous uric acid, struvite, brushite, and cystine) using a simulated in vivo dataset of endoscopic images. They achieved an accuracy of 88.21% and further evaluated the explanations by perturbing the global visual characteristics of the images (hue, texture, contrast, shape, brightness, and saturation) to describe their relevance to the prototypes and the sensitivity of the model.
Kong et al. [39] designed a dual-path PP network, DP-ProtoNet, to increase the generalization performances of single networks. They applied their model to the public ISIC-HAM10000 skin disease dermatoscopic dataset [40]. Compared to ProtoPNet, DP-ProtoNet achieved a better performance while maintaining the model’s interpretability.
Santiago et al. [41] integrated ProtoPNet with content-based image retrieval to provide explanations in terms of image-level prototypes and patch-level prototypes. They applied their approach to skin lesion diagnoses in the public ISIC dermoscopic images [40], outperforming both black-box models and SOTA explainable approaches.
Cui et al. [42] proposed MBC-ProtoTree, an interpretable fine-grained image classification network based on ProtoTree. They improved ProtoTree by designing a multi-grained feature extraction network, a new background prototype removal mechanism, and a novel loss function. The improved model achieved a higher classification accuracy on the chest X-ray dataset.
Nauta et al. [14] applied their PIPNet to two open benchmark datasets, respectively, for skin cancer diagnosis (ISIC), bone X-ray abnormality detection (MURA), and two real-world datasets for hip and ankle fracture detection, respectively. From this study, the authors obtained prototypes that were generally in line with medical knowledge and demonstrated the possibility of correcting the undesired model’s reasoning process with a human-in-the-loop configuration.
Santos et al. [43] integrated ProtoPNet into a Deep Active Learning framework to predict diabetic retinopathy on the Messidor dataset [44]. The framework allows ProtoPNet to be trained on a training set of instances selected using a search strategy, and this may offer benefits in scenarios where the datasets are expensive to label. The reported performance demonstrated the success of applying interpretable models with reduced training data.
Wang et al. [45] proposed InterNRL, integrating ProtoPNet into a student–teacher framework, together with an accurate global image classifier. They applied their model for breast cancer and retinal disease diagnosis and obtained the SOTA performances, demonstrating the success of the reciprocal learning training paradigm in the medical imaging domain.
Xu et al. [46] proposed a prototype-based vision transformer applied to COVID-19 classification. They replaced the last two layers of a transformer encoder with a prototype block similar to ProtoPNet and obtained good performances on three different public datasets.
Sinhamahapatra et al. [47] proposed ProtoVerse, a PP model with a novel objective function applied to vertebral compression fracture classification. Expert radiologists evaluated the model interpretability, and the predictive performances outperformed those of ProtoPNet and other SOTA PP architectures.
Pathak et al. [18] further applied PIPNet to three different public datasets for breast cancer classification, obtaining competitive performances with respect to other SOTA black-box and prototype-based models and assessed the coherence of the model with quantitative metrics.
Gallée et al. [24] applied the Proto-Caps PP net on lung and chest CT nodule malignancy prediction and performed a human evaluation of the model with a user study.
In summary, PP networks are gaining increasing interest in the MI domain. Existing applications are based on the original or implementation variants of the identified PP foundational models (see Section 2). Most of the studies obtained comparable or higher performances than their black-box counterparts in a large variety of public and private real-world hospital datasets. These interesting results strongly support their usage to automatically classify medical images in an interpretable way.

3.2. Three-Dimensional and Multimodal Models

We observed increasing interest in extending PP-NN model applications from general computer vision tasks to 3D medical images. Wei et al. [48] proposed MProtoNet as the first medical prototype network extending PP models to 3D brain tumour classification from multiparametric MRI (mpMRI). This is based on an implementation of XProtoNet with a 3D ResNet backbone and a novel proposed soft masking and online-CAM loss function to enhance the localization of attention regions. The authors further proposed novel metrics to assess the correctness and localization coherence of the prototypes (following the Co-12 taxonomy [49]). The MProtoNet obtained a classification accuracy in line with that of the baseline CNN counterpart and ProtoPNet (re-implemented for 3D images).
Vaseli et al. [50] proposed ProtoASNet, a PP model that extracts spatio-temporal feature vectors to detect aortic stenosis from B-mode echocardiography videos. They assessed their model on both a private and a public TMED-2 dataset [51], and it outperformed the baseline ProtoPNet and XProtoNet. De Santi et al. [52] extended PIPNet to classify 3D MRI for Alzheimer’s disease. They tested the model using different CNN backbones and obtained the best result with ResNet-18 3D, and comparable results were obtained with the corresponding black-box counterpart. They further proposed two novel quantitative metrics of explainability: the prototype brain entropy to assess the covariate complexity and localization consistency to assess the consistency of prototype localization (under the Co-12 properties). They also evaluated prototypes with domain experts through a survey and assessed the coherency of prototypes in terms of the localization, pattern, and classification.
Physicians perform medical diagnoses by considering a wide range of patient information, not just image data. This includes clinical, instrumental, and other relevant examinations [23,24]. In this context, developing multimodal PP models could help create systems that implement a decision-making process closer to a real-life application scenario [23,24].
In the context of multimodality-based prototype learning, we can distinguish two main approaches in producing the prototype representation according to the way of introducing different modalities: (1) deterministic prototypes and (2) shifted prototypes [53]. The first approach typically trains different encoders for all the different modalities, and then a connection layer (cross-modal encoder) concatenates the output feature vector extracted, generating a unimodal vector into a joint space used by the PP net to perform a similarity calculation. Alternatively, some studies have introduced multimodal feature vectors to shift the prototype representations without prior fusion. Here, existing approaches leverage auxiliary modalities to enrich the embedding representation of a main modality. The shifted prototype can be explicitly modeled, or their influence can be indirectly added to the objective function used to train an embedded network [53]. Most of the multimodal PP networks proposed in general CV applications integrate visual information with text data [53]. Despite still being considered a relatively unexplored field, we can also find some applications in medical imaging.
Table 1. Part -prototype network applied in medical imaging. Here, we report predictive accuracy (Acc), balanced accuracy (Bal Acc), and Area Under the Curve (AUC) according to the provided predictive performances.
Table 1. Part -prototype network applied in medical imaging. Here, we report predictive accuracy (Acc), balanced accuracy (Bal Acc), and Area Under the Curve (AUC) according to the provided predictive performances.
PaperModalityDatasetClassesResults
Two-dimensional image models
Singh et al. [25]X-rayChest X-ray, COVID-19 ImageNormal, pneumonia, COVID−19Acc = 88.99%
Singh et al. [26]X-rayChest X-ray, COVID-19 ImageNormal, pneumonia, COVID-19Acc = 87.27%
Singh et al. [27]CTCOVIDx CT-2Normal, pneumonia, COVID−19Acc = 99.24%
Kim et al. [19]X-rayNIH chest X-rayAtelectasis, cardiomegaly, effusion, infiltration, mass, nodule, pneumonia, pneumathorax, consolidation, edema, emphysema, fibrosis, pleural thickening, herniaMean AUC = 0.822
Mohammadjafari et al. [31]MRIOASIS, ADNINormal vs. Alzheimer’s diseaseAcc: 87.17% (OASIS), 91.02% (ADNI)
Barnett et al. [33]MammographyInternal datasetMass margin classification Malignancy predictionAUC: 0.951 (Mass-margin) 0.84 (Malignancy)
Carloni et al. [34]MammographyCBIS-DDSMBenign, malignantAcc = 68.5%
Amorim et al. [36]HistologyPatchCamelyonBenign, malignantAcc = 98.14%
Flores-Araiza et al. [38]EndoscopySimulated in vivo datasetWhewellite, weddellite, anhydrous uric acid, struvite, brushite, cystineAcc = 88.21%
Kong et al. [39]Dermatoscopic imagesISIC-HAM10000Actinic keratosis, intraepithelial carcinoma, nevi, basal cell carcinoma, benign keratosis-like lesions, dermatofibroma, melanoma, vascular lesionsF1 = 74.6
Santiago et al. [41]Dermatoscopic imagesISIC-HAM10000Actinic keratosis, intraepithelial carcinoma, nevi, basal cell carcinoma, benign keratosis-like lesions, dermatofibroma, melanoma, vascular lesionsBal Acc = 75.0% (Highest, achieved with DenseNet)
Cui et al. [42]X-rayChest X-rayNormal, pneumoniaAcc = 91.4%
Nauta et al. [14]Dermoscopic images, X-rayISIC, MURA, Hip and ankle fraction internal datasetBenignant, malignant, normal, abnormal, fracture, no fractureAcc: 94.1% (ISIC) 82.1% (MURA) 94.0% (Hip) 77.3% (Ankle)
Santos et al. [43]RetinographMessidorHealthy vs. diseased retinopathyAUC = 0.79
Wang et al. [45]Mammography, retinal OCTMammography internal dataset, CMMD, NEH OCTCancer vs. non-cancer Benignant vs. malignant Normal, drusen, and choroidal neovascularizationAUC = 91.49 (Internal) AUC = 89.02 (CMMD) Acc = 91.9 (NEH OCT)
Xu et al. [46]X-ray Lung CTCOVIDx CXR-3 COVID-QU-Ex Lung CT scanCOVID-10, normal, pneumoniaF1: 99.2 96.8 98.5
Sinhamahapatra et al. [47]CTVerSe’19 datasetFracture vs. healthyF1 = 75.97
Pathak et al. [18]MammographyCBIS, VinDir, CMMDBenignant vs. malignantF1 (PIP-Net model): 63 ± 3% (CBIS) 63 ± 3% (VinDir) 70 ± 1% (CMMD)
Gallée et al. [24]Thorax CTLIDC-IDRIBenignant vs. malignantAcc = 93.0%
Three-dimensional image models
Wei et al. [48]3D mpMRI: T1, T1CE, T2, FLAIRBraTS 2020High-grade vs. low-grade gliomaBal Acc = 85.8%
Vaseli et al. [50]EchocardiographyPrivate dataset TMED-2Normal vs. mild vs. severe aortic stenosisAcc: 80.0% (Private) 79.7 (TMED-2)%
De Santi et al. [52]3D MRIADNINormal vs. Alzheimer’s diseaseBal Acc = 82.02%
Multimodal models
Wolf et al. [54]3D 18F-FDG PET and Tabular dataADNINormal vs. Alzheimer’s diseaseBal Acc = 60.7%
Wang et al. [55]Chest X-ray and reportsMIMIC-CXRAtelectasis, cardiomegaly, consolidation, edema, enlarged cardiomediastinum, fracture, lung lesion, lung opacity, pleural effusion, pleural other, pneumonia, pneumothorax, support deviceMean AUC = 0.828
De Santi et al. [56]3D MRI and AgesADNINormal vs. Alzheimer’s diseaseBal Acc = 83.04%
Wolf et al. [54] developed PANIC, a Prototypical Additive Neural Network for Interpretable Classification of Alzheimer’s Disease, which performs the classification of 3D 18F-FDG PET images integrated with tabular data as Alzheimer’s Disease or healthy. PANIC integrated the image prototypes extracted by a 3D ProtoPNet with N tabular features using an interpretable Generative Additive Model. PANIC obtained a higher balanced accuracy compared to the performances obtained with a black-box model for heterogeneous data. Wang et al. [55] proposed MProtoNet, an interpretable multimodal network that performs diagnosis using images and textual information from medical reports. The architecture combines a position embedding and a multimodal attention module applied to the chest MIMIC-CXR dataset [57], and reported improvements in predictive performances with respect to ProtoPNet without sacrificing the interpretable nature. De Santi et al. [56] proposed a Patch-based Intuitive Multimodal Prototype Network (PIMPNet), a multimodal prototype classifier that learns 3D image part-prototypes and prototypical values from structured data, to predict a patient’s cognitive level in AD from sMRI and age values. They introduce the concept of age-prototypical layer to directly learn the relevant age values. PIMPNet performs classification by concatenating the image prototypes extracted by a 3D PIPNet with age prototypes using an interpretable scoring sheet system; however, the age prototypes do not improve the predictive performance of the model trained with images only.
The unique features of the medical domain may have benefits for extending general-purpose PP models so that they can process volumetric data and data coming from different modalities. Although there are emerging proposals of this kind, the literature remains limited compared to that related to 2D image classification. In particular, it seems that an optimal solution for effectively managing data from various modalities has not yet been identified.

4. Evaluation of Prototypes

Evaluating the XAI method is crucial to verify that the explanations produced are robust, sensitive to the model and data, and consistent, but this is still considered an open and developing field [5,58].
Assessing the XAI outputs is considered challenging: there is still no agreement on what interpretability and explainability really are, and a general ground truth to compare explanations is missing. Researchers have proposed different evaluation frameworks to assess XAI methods, and despite there not being a unique solution yet, most approaches agree concerning the need to consider the multidisciplinary and multidomain nature of the explanation [1]. In this scenario, including quantitative metrics to assess the technical appropriateness and reliability of explanations is crucial. To effectively capture the whole impact of XAI on supporting decision systems in real-world scenarios, it is also fundamental to include the human aspect of explanations in the evaluation process [17,59]. The human evaluation should assess how explanations are perceived, their relationship with any expert domain background knowledge, how they affect the classification performances, and how all those aspects vary according to the presentation format of the explanations.
However, while the effectiveness of post hoc explainability methods has been investigated extensively, research on the quantitative analysis of self-explainable approaches is still in its early stage [60]. Particularly relevant is the general evaluation framework proposed by Nauta et al. [49], the so-called “Co-12 properties”, which have been further applied to PP models, providing a method for systematically evaluating this family of interpretable models [15]. This overview might be particularly relevant as most XAI evaluation methods were designed for posthoc XAI [15,60]. Here, the “explainability” is considered a non-binary characteristic, and authors have developed a categorization scheme based on three main dimensions further divided into 12 conceptual properties. This “evaluation cheat-sheet” suggested the implementation of all content, presentation and user-related experiments, involving experts from different relevant disciplines, including computer science, human–computer interaction (HCI), and user experience design. We summarized the evaluation framework, with the properties grouped by their most prominent dimension and contextualized to the family of PP networks, in Table 2.
The availability of synthetic quantitative metrics is fundamental for objectively assessing the explainability quality offered by XAI systems, and authors have started proposing different solutions in this direction.
Wei et al. [48] proposed two novel interpretability metrics to assess the correctness and localization coherence of prototypes (following the Co-12 taxonomy [49]). The correctness was evaluated by proposing the incremental deletion score (IDS), which is based on an incremental deletion process. The localization coherence was evaluated with activation precision (AP) metrics, which assessed the intersection between the activation map and the human-annotated label.
Pathak et al. [18] further developed a general framework for coherence (one of the Co-12 properties), the so-called PEF-Coh, to quantitatively measure the quality of prototypes with respect to the domain knowledge for breast cancer prediction from mammography. The PEF-Coh presents six metrics whose assessment requires a dataset annotated with regions of interest: Relevance, Specialization, Uniqueness, Coverage, Class-specific, and localization.
Gallée et al. [24] evaluated the Proto-Caps PP net, which predicts malignancy scores in lung chest CT nodules, and provides explanations in terms of predefined image feature attributes and image prototypes. They evaluated their model by performing a user study with six radiologists based on a survey to test how explanations affect the (1) user’s performance, (2) trust in the model, and if they are (3) helpful or not. They structured a questionnaire differentiating by radiologists’ experience; assessing for the malignancy; their confidence in the prediction; and the model’s output. Here, test cases were provided to the users with different levels of explainability. Finally, the questioner asked for an overall assessment of the explanations’ helpfulness. The results confirmed the findings in the existing literature in observing that (1) explanations improve performances when the model is correct, but might also convince the user of incorrect predictions; (2) the trust in the model is both influenced by the model’s accuracy and the extent of the model’s reasoning; and (3) explanations are generally perceived as useful for the user.
Recent work aims to evaluate explanations using synthetic datasets. In particular, the FunnyBirds framework introduces a synthetically generated dataset of birds with concept annotation, and a set of metrics to evaluate explanations following the Co-12 taxonomy.
Opłatek et al. [61] applied the FunnyBirds framework to evaluate prototypes produced by ProtoPNet. Their results suggested that explanations provided in the form of similarity maps instead of bounding boxes provide a more faithful representation of the model, resulting in higher-quality explanations. This study further highlights the importance of considering proper visualization techniques in PP models.
The prototype evaluation process has also recently highlighted some concerns about the correctness of the spatial localization of the prototypes [60]. The visualization process of the prototypical part extracted by PP models is generally based on the upsampling of the extracted similarity maps. Despite PP nets being claimed to implement an inherently self-explainable case-based reasoning process, recent studies have reported concerns about the faithfulness of the model-agnostic upsampling process for part-prototype visualization [60,62]. Guatam et al. [60] proposed Prototype Relevance Propagation (PRP) as a model-aware visualization strategy for prototypical parts to address the main drawbacks of PP net upsampling visualization: the low-resolution activation maps and spatially imprecise prototype explanations. Xu et al. observed that the upsampling process may incorrectly locate parts of the images, suggesting the use of what they reported as more faithful saliency methods like SmoothGrads or PRP [62].

5. Discussion

Part-prototype networks are self-explainable methods that combine the computational power of DL into an interpretable-by-design decision-making process. PP nets constitute a prominent solution in medical imaging analysis, where explainability is crucial to ensure the trustworthiness of the AI system, but post hoc explainable methods have reported several drawbacks in terms of reliability and completeness [5,18].
ProtoPNet was the first part-prototype network proposed to perform an interpretable imaging classification; then, other architectural variants have followed it [13]. These still implement a classification using a case-based reasoning process, but differ from the first one in terms of how they combine prototypical parts with base decisions [20], introduce the concept of class-shared prototypes [21], and address some of the reported ProtoPNet limitations, like the semantic gap between latent and image prototypes representations [8].
Most PP nets were originally designed for general CV applications and then shifted to the medical imaging domain. Integrating a deep feature extractor into a PP network often led to competitive performances compared to the fully black-box model counterparts [26,31,36], with the potential to also correct undesired reasoning processes by performing a prototype suppression [14]. Benefits have already been reported for medical images with the use of the general CV PP net; however, this domain might also require particular adaptations such as the ability to process 3D image scans and the multimodal data domain [48,54,56]. In particular, an optimal solution for the management of multiple data types still seems to be missing.
Evaluating XAI methods, including PP nets, with standard quantitative frameworks is still considered an open challenge, but is crucial for developing reliable and trustworthy systems. Authors generally agree that the evaluation process should be multidisciplinary and multidomain to be able to capture the multiple aspects that characterize the explanations, such as the evaluation method proposed by Nauta et al. [15]. Most MI applications of the PP net involve an evaluation process that includes domain expert user studies and quantitative metrics. Although the training of the PP net does not require further annotation with class-level labels, most of the prototype evaluation metrics proposed are based on the usage of further image-level annotations, such as the segmentation of regions of interest in the images. So, the availability of sub-annotated image datasets might also be beneficial for the development of PP models. Finally, despite the general belief in considering PP nets to be reliable XAI methods due to their self-explainable nature, recent studies have highlighted drawbacks in prototype visualization based on a model-agnostic upsampling process, and proposed alternative solutions to produce more spatially accurate prototypes [60,62].

Author Contributions

Writing—review and editing, L.A.D.S., F.I.P., F.B., M.F.S. and V.P.; supervision, S.C. and V.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing is not applicable.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
APactivation precision
AUCArea Under the Curve
AUROCArea Under the Receiver Operating Characteristic
CNNConvolutional Neural Network
CTcomputed tomography
CVComputer Vision
DLDeep Learning
IDSincremental deletion score
MIMedical Imaging
APactivation precision
AUCArea Under the Curve
AUROCArea Under the Receiver Operating Characteristic
CNNConvolutional Neural Network
CTComputed Tomography
CVComputer Vision
DLDeep Learning
IDSincremental deletion score
MIMedical Imaging
MLMachine Learning
MRmagnetic resonance
MRImagnetic resonance imaging
OoDOut-of-Distribution
PIPNetPatch-based Intuitive Prototype Network
PPpart-prototype
RXRadiography
SGDstochastic gradient descent
SOTAState-of-the-Art
XAIExplainable Artificial Intelligence

References

  1. Salahuddin, Z.; Woodruff, H.C.; Chatterjee, A.; Lambin, P. Transparency of deep neural networks for medical image analysis: A review of interpretability methods. Comput. Biol. Med. 2022, 140, 105111. [Google Scholar] [CrossRef] [PubMed]
  2. Borys, K.; Schmitt, Y.A.; Nauta, M.; Seifert, C.; Krämer, N.; Friedrich, C.M.; Nensa, F. Explainable AI in medical imaging: An overview for clinical practitioners—Saliency-based XAI approaches. Eur. J. Radiol. 2023, 162, 110787. [Google Scholar] [CrossRef] [PubMed]
  3. Borys, K.; Schmitt, Y.A.; Nauta, M.; Seifert, C.; Krämer, N.; Friedrich, C.M.; Nensa, F. Explainable AI in medical imaging: An overview for clinical practitioners—Beyond saliency-based XAI approaches. Eur. J. Radiol. 2023, 162, 110786. [Google Scholar] [CrossRef]
  4. Allgaier, J.; Mulansky, L.; Draelos, R.L.; Pryss, R. How does the model make predictions? A systematic literature review on the explainability power of machine learning in healthcare. Artif. Intell. Med. 2023, 143, 102616. [Google Scholar] [CrossRef]
  5. Longo, L.; Brcic, M.; Cabitza, F.; Choi, J.; Confalonieri, R.; Ser, J.D.; Guidotti, R.; Hayashi, Y.; Herrera, F.; Holzinger, A.; et al. Explainable Artificial Intelligence (XAI) 2.0: A manifesto of open challenges and interdisciplinary research directions. Inf. Fusion 2024, 106, 102301. [Google Scholar] [CrossRef]
  6. Li, O.; Liu, H.; Chen, C.; Rudin, C. Deep Learning for Case-Based Reasoning through Prototypes: A Neural Network that Explains Its Predictions. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2017. [Google Scholar]
  7. Cynthia, R. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 2019, 1, 206–215. [Google Scholar] [CrossRef]
  8. Nauta, M.; Schlötterer, J.; van Keulen, M.; Seifert, C. PIP-Net: Patch-Based Intuitive Prototypes for Interpretable Image Classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 2744–2753. [Google Scholar]
  9. Barredo Arrieta, A.; Díaz-Rodríguez, N.; Del Ser, J.; Bennetot, A.; Tabik, S.; Barbado, A.; Garcia, S.; Gil-Lopez, S.; Molina, D.; Benjamins, R.; et al. Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Inf. Fusion 2020, 58, 82–115. [Google Scholar] [CrossRef]
  10. Montavon, G.; Samek, W.; Müller, K.R. Methods for interpreting and understanding deep neural networks. Digit. Signal Process. 2018, 73, 1–15. [Google Scholar] [CrossRef]
  11. David, G.; Eric, V.; Yunyan, W.J.; Matt, T. DARPA’s explainable AI (XAI) program: A retrospective. Appl. Lett. 2021, 2, e61. [Google Scholar] [CrossRef]
  12. Biederman, I. Recognition-by-Components: A Theory of Human Image Understanding. Psychol. Rev. 1987, 94, 115–147. [Google Scholar] [CrossRef]
  13. Chen, C.; Li, O.; Tao, C.; Barnett, A.J.; Su, J.; Rudin, C. This Looks Like That: Deep Learning for Interpretable Image Recognition. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  14. Nauta, M.; Hegeman, J.H.; Geerdink, J.; Schlötterer, J.; Keulen, M.v.; Seifert, C. Interpreting and Correcting Medical Image Classification with PIP-Net. In Artificial Intelligence, ECAI 2023 International Workshops, Proceedings of the XAI^3, TACTIFUL, XI-ML, SEDAMI, RAAIT, AI4S, HYDRA, AI4AI, Kraków, Poland, 30 September–4 October 2023; Springer: Cham, Switzerland, 2024; pp. 198–215. [Google Scholar]
  15. Nauta, M.; Seifert, C. The Co-12 Recipe for Evaluating Interpretable Part-Prototype Image Classifiers. In Explainable Artificial Intelligence, Proceedings of the First World Conference, xAI 2023, Lisbon, Portugal, 26–28 July 2023; Longo, L., Ed.; Springer: Cham, Switzerland, 2023; pp. 397–420. [Google Scholar]
  16. Doshi-Velez, F.; Kim, B. Towards A Rigorous Science of Interpretable Machine Learning. arXiv 2017, arXiv:1702.08608. [Google Scholar]
  17. Jin, W.; Li, X.; Fatehi, M.; Hamarneh, G. Guidelines and evaluation of clinical explainable AI in medical image analysis. Med. Image Anal. 2023, 84, 102684. [Google Scholar] [CrossRef] [PubMed]
  18. Pathak, S.; Schlötterer, J.; Veltman, J.; Geerdink, J.; Keulen, M.V.; Seifert, C.; Pathak, S. Prototype-Based Interpretable Breast Cancer Prediction Models: Analysis and Challenges. In Explainable Artificial Intelligence, Proceedings of the Second World Conference, xAI 2024, Valletta, Malta, 17–19 July 2024; Springer: Cham, Switzerland, 2024; pp. 21–42. [Google Scholar] [CrossRef]
  19. Kim, E.; Kim, S.; Seo, M.; Yoon, S. XProtoNet: Diagnosis in chest radiography with global and local explanations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 15719–15728. [Google Scholar]
  20. Nauta, M.; van Bree, R.; Seifert, C. Neural Prototype Trees for Interpretable Fine-Grained Image Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 14933–14943. [Google Scholar]
  21. Rymarczyk, D.; Struski, Ł.; Tabor, J.; Zieliński, B. ProtoPShare: Prototypical Parts Sharing for Similarity Discovery in In-terpretable Image Classification. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, Singapore, 14–18 August 2021; Volume 11. [Google Scholar] [CrossRef]
  22. Rymarczyk, D.; Struski, Ł.; Górszczak, M.; Lewandowska, K.; Tabor, J.; Zieliński, B. Interpretable Image Classification with Differentiable Prototypes Assignment. In Computer Vision—ECCV 2022, Proceedings of the 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2022; Volume 13672, pp. 351–368. [Google Scholar] [CrossRef]
  23. van de Beld, J.J.; Pathak, S.; Geerdink, J.; Hegeman, J.H.; Seifert, C. Feature Importance to Explain Multimodal Prediction Models. a Clinical Use Case. In Explainable Artificial Intelligence, Proceedings of the Second World Conference, xAI 2024, Valletta, Malta, 17–19 July 2024; Springer: Cham, Switzerland, 2024; pp. 84–101. [Google Scholar] [CrossRef]
  24. Gallée, L.; Lisson, C.S.; Lisson, C.G.; Drees, D.; Weig, F.; Vogele, D.; Beer, M.; Götz, M. Evaluating the Explainability of Attributes and Prototypes for a Medical Classification Model. In Explainable Artificial Intelligence, Proceedings of the Second World Conference, xAI 2024, Valletta, Malta, 17–19 July 2024; Springer: Cham, Switzerland, 2024; pp. 43–56. [Google Scholar] [CrossRef]
  25. Singh, G.; Yow, K.C. These do not look like those: An interpretable deep learning model for image recognition. IEEE Access 2021, 9, 41482–41493. [Google Scholar] [CrossRef]
  26. Singh, G.; Yow, K.C. An Interpretable Deep Learning Model for COVID-19 Detection with Chest X-Ray Images. IEEE Access 2021, 9, 85198–85208. [Google Scholar] [CrossRef]
  27. Singh, G.; Yow, K.C. Object or background: An interpretable deep learning model for COVID-19 detection from CT-scan images. Diagnostics 2021, 11, 1732. [Google Scholar] [CrossRef]
  28. Kermany, D.; Zhang, K.; Goldbaum, M. Large dataset of labeled optical coherence tomography (oct) and chest X-ray images. Mendeley Data 2018, 3. [Google Scholar] [CrossRef]
  29. Cohen, J.P.; Morrison, P.; Dao, L. COVID-19 image data collection. arXiv 2020, arXiv:2003.11597. [Google Scholar]
  30. Wang, X.; Peng, Y.; Lu, L.; Lu, Z.; Bagheri, M.; Summers, R.M. Chestx-ray8: Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2097–2106. [Google Scholar]
  31. Mohammadjafari, S.; Cevik, M.; Thanabalasingam, M.; Basar, A.; Initiative, A.D.N. Using ProtoPNet for Interpretable Alzheimer’s Disease Classification. In Proceedings of the Canadian AI 2021, Canadian Artificial Intelligence Association (CAIAC), Vancouver, BC, Canada, 25–28 May 2021; Available online: https://caiac.pubpub.org/pub/klwhoig4 (accessed on 5 September 2024). [CrossRef]
  32. Marcus, D.S.; Wang, T.H.; Parker, J.; Csernansky, J.G.; Morris, J.C.; Buckner, R.L. Open Access Series of Imaging Studies (OASIS): Cross-sectional MRI data in young, middle aged, nondemented, and demented older adults. J. Cogn. Neurosci. 2007, 19, 1498–1507. [Google Scholar] [CrossRef]
  33. Barnett, A.J.; Schwartz, F.R.; Tao, C.; Chen, C.; Ren, Y.; Lo, J.Y.; Rudin, C. A case-based interpretable deep learning model for classification of mass lesions in digital mammography. Nat. Mach. Intell. 2021, 3, 1061–1070. [Google Scholar] [CrossRef]
  34. Carloni, G.; Berti, A.; Iacconi, C.; Pascali, M.A.; Colantonio, S. On the applicability of prototypical part learning in medical images: Breast masses classification using ProtoPNet. In Pattern Recognition, Computer Vision, and Image Processing, Proceedings of the ICPR 2022 International Workshops and Challenges, Montreal, QC, Canada, 21–25 August 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 539–557. [Google Scholar]
  35. Lee, R.S.; Gimenez, F.; Hoogi, A.; Miyake, K.K.; Gorovoy, M.; Rubin, D.L. A curated mammography data set for use in computer-aided detection and diagnosis research. Sci. Data 2017, 4, 170177. [Google Scholar] [CrossRef]
  36. Amorim, J.P.; Abreu, P.H.; Santos, J.; Müller, H. Evaluating Post-hoc Interpretability with Intrinsic Interpretability. arXiv 2023, arXiv:2305.03002. [Google Scholar]
  37. Bejnordi, B.E.; Veta, M.; Van Diest, P.J.; Van Ginneken, B.; Karssemeijer, N.; Litjens, G.; Van Der Laak, J.A.; Hermsen, M.; Manson, Q.F.; Balkenhol, M.; et al. Diagnostic assessment of deep learning algorithms for detection of lymph node metastases in women with breast cancer. Jama 2017, 318, 2199–2210. [Google Scholar] [CrossRef] [PubMed]
  38. Flores-Araiza, D.; Lopez-Tiro, F.; El-Beze, J.; Hubert, J.; Gonzalez-Mendoza, M.; Ochoa-Ruiz, G.; Daul, C. Deep prototypical-parts ease morphological kidney stone identification and are competitively robust to photometric perturbations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 295–304. [Google Scholar]
  39. Kong, L.; Gong, L.; Wang, G.; Liu, S. DP-ProtoNet: An interpretable dual path prototype network for medical image diagnosis. In Proceedings of the 2023 IEEE 22nd International Conference on Trust, Security and Privacy in Computing and Communications, TrustCom/BigDataSE/CSE/EUC/iSCI 2023, Exeter, UK, 1–3 November 2023; pp. 2797–2804. [Google Scholar] [CrossRef]
  40. Tschandl, P.; Rosendahl, C.; Kittler, H. The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions. Sci. Data 2018, 5, 180161. [Google Scholar] [CrossRef] [PubMed]
  41. Santiago, C.; Correia, M.; Verdelho, M.R.; Bissoto, A.; Barata, C. Global and Local Explanations for Skin Cancer Diagnosis Using Prototypes. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2023 Workshops, Proceedings of the ISIC 2023, Care-AI 2023, MedAGI 2023, DeCaF 2023, Held in Conjunction with MICCAI 2023, Vancouver, BC, Canada, 8–12 October 2023; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2023; Volume 14393, pp. 47–56. [Google Scholar] [CrossRef]
  42. Cui, J.; Gong, J.; Wang, G.; Li, J.; Liu, X.; Liu, S. An Novel Interpretable Fine-grained Image Classification Model Based on Improved Neural Prototype Tree. In Proceedings of the IEEE International Symposium on Circuits and Systems, Monterey, CA, USA, 21–25 May 2023. [Google Scholar] [CrossRef]
  43. de A. Santos, I.B.; de Carvalho, A.C.P.L.F. ProtoAL: Interpretable Deep Active Learning with prototypes for medical imaging. arxiv 2024, arXiv:cs.CV/2404.04736. [Google Scholar]
  44. Gu, Z.; Cheng, J.; Fu, H.; Zhou, K.; Hao, H.; Zhao, Y.; Zhang, T.; Gao, S.; Liu, J. CE-Net: Context Encoder Network for 2D Medical Image Segmentation. IEEE Trans. Med. Imaging 2019, 38, 2281–2292. [Google Scholar] [CrossRef] [PubMed]
  45. Wang, C.; Chen, Y.; Liu, F.; Elliott, M.; Kwok, C.F.; Pena-Solorzano, C.; Frazer, H.; Mccarthy, D.J.; Carneiro, G. An Interpretable and Accurate Deep-Learning Diagnosis Framework Modeled With Fully and Semi-Supervised Reciprocal Learning. IEEE Trans. Med. Imaging 2024, 43, 392–404. [Google Scholar] [CrossRef]
  46. Xu, Y.; Meng, Z. Interpretable vision transformer based on prototype parts for COVID-19 detection. IET Image Process. 2024, 18, 1927–1937. [Google Scholar] [CrossRef]
  47. Sinhamahapatra, P.; Shit, S.; Sekuboyina, A.; Husseini, M.; Schinz, D.; Lenhart, N.; Menze, J.; Kirschke, J.; Roscher, K.; Guennemann, S. Enhancing Interpretability of Vertebrae Fracture Grading using Human-interpretable Prototypes. J. Mach. Learn. Biomed. Imaging 2024, 2024, 977–1002. [Google Scholar] [CrossRef]
  48. Wei, Y.; Tam, R.; Tang, X. MProtoNet: A Case-Based Interpretable Model for Brain Tumor Classification with 3D Multi-parametric Magnetic Resonance Imaging. In Proceedings of the Medical Imaging with Deep Learning, Nashville, TN, USA, 10–12 July 2023. [Google Scholar]
  49. Nauta, M.; Trienes, J.; Pathak, S.; Nguyen, E.; Peters, M.; Schmitt, Y.; Schlötterer, J.; van Keulen, M.; Seifert, C. From Anecdotal Evidence to Quantitative Evaluation Methods: A Systematic Review on Evaluating Explainable AI. ACM Comput. Surv. 2023, 55, 1–42. [Google Scholar] [CrossRef]
  50. Vaseli, H.; Gu, A.N.; Amiri, S.N.A.; Tsang, M.Y.; Fung, A.; Kondori, N.; Saadat, A.; Abolmaesumi, P.; Tsang, T.S. ProtoASNet: Dynamic Prototypes for Inherently Interpretable and Uncertainty-Aware Aortic Stenosis Classification in Echocardiography. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2023, Proceedings of the 26th International Conference, Vancouver, BC, Canada, 8–12 October 2023; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2023; Volume 14225, pp. 368–378. [Google Scholar] [CrossRef]
  51. Huang, Z.; Long, G.; Wessler, B.; Hughes, M. TMED 2: A Dataset for Semi-Supervised Classification of Echocardiograms. 2022. Available online: https://www.michaelchughes.com/papers/HuangEtAl_TMED2_DataPerf_2022.pdf (accessed on 5 September 2024).
  52. De Santi, L.A.; Schlötterer, J.; Scheschenja, M.; Wessendorf, J.; Nauta, M.; Positano, V.; Seifert, C. PIPNet3D: Interpretable Detection of Alzheimer in MRI Scans. arXiv 2024, arXiv:2403.18328. [Google Scholar]
  53. Ma, Y.; Zhao, S.; Wang, W.; Li, Y.; King, I. Multimodality in meta-learning: A comprehensive survey. Know.-Based Syst. 2022, 250, 108976. [Google Scholar] [CrossRef]
  54. Wolf, T.N.; Pölsterl, S.; Wachinger, C. Don’t PANIC: Prototypical Additive Neural Network for Interpretable Classification of Alzheimer’s Disease. In Proceedings of the Information Processing in Medical Imaging: 28th International Conference, IPMI 2023, San Carlos de Bariloche, Argentina, 18–23 June 2023; Proceedings. Springer: Berlin/Heidelberg, Germany, 2023; pp. 82–94. [Google Scholar] [CrossRef]
  55. Wang, G.; Li, J.; Tian, C.; Ma, X.; Liu, S. A Novel Multimodal Prototype Network for Interpretable Medical Image Classification. In Proceedings of the Conference Proceedings—IEEE International Conference on Systems, Man and Cybernetics, Honolulu, HI, USA, 1–4 October 2023; pp. 2577–2583. [Google Scholar] [CrossRef]
  56. De Santi, L.A.; Schlötterer, J.; Nauta, M.; Positano, V.; Seifert, C. Patch-based Intuitive Multimodal Prototypes Network (PIMPNet) for Alzheimer’s Disease classification. In Proceedings of the xAI 2024 Late-breaking Work, Demos and Doctoral Consortium co-located with the 2nd World Conference on eXplainable Artificial Intelligence (xAI 2024), Valletta, Malta, 17–19 July 2024; pp. 73–80. Available online: https://ceur-ws.org/Vol-3793/paper_10.pdf (accessed on 5 September 2024).
  57. Johnson, A.E.W.; Pollard, T.J.; Greenbaum, N.R.; Lungren, M.P.; ying Deng, C.; Peng, Y.; Lu, Z.; Mark, R.G.; Berkowitz, S.J.; Horng, S. MIMIC-CXR-JPG, a large publicly available database of labeled chest radiographs. arXiv 2019, arXiv:1901.07042. [Google Scholar]
  58. van der Velden, B.H.; Kuijf, H.J.; Gilhuijs, K.G.; Viergever, M.A. Explainable artificial intelligence (XAI) in deep learning-based medical image analysis. Med. Image Anal. 2022, 79, 102470. [Google Scholar] [CrossRef] [PubMed]
  59. Cabitza, F.; Campagner, A.; Ronzio, L.; Cameli, M.; Mandoli, G.E.; Pastore, M.C.; Sconfienza, L.M.; Folgado, D.; Barandas, M.; Gamboa, H. Rams, hounds and white boxes: Investigating human–AI collaboration protocols in medical diagnosis. Artif. Intell. Med. 2023, 138, 102506. [Google Scholar] [CrossRef]
  60. Gautam, S.; Höhne, M.M.C.; Hansen, S.; Jenssen, R.; Kampffmeyer, M. This looks More Like that: Enhancing Self-Explaining Models by Prototypical Relevance Propagation. Pattern Recogn. 2023, 136, 109172. [Google Scholar] [CrossRef]
  61. Opłatek, S.; Rymarczyk, D.; Zieliński, B. Revisiting FunnyBirds Evaluation Framework for Prototypical Parts Networks. In Explainable Artificial Intelligence, Proceedings of the Second World Conference, xAI 2024, Valletta, Malta, 17–19 July 2024; Springer: Cham, Switzerland, 2024; pp. 57–68. [Google Scholar] [CrossRef]
  62. Xu-Darme, R.; Quénot, G.; Chihani, Z.; Rousset, M.C. Sanity checks for patch visualisation in prototype-based image classification. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Vancouver, BC, Canada, 17–24 June 2023; pp. 3691–3696. [Google Scholar] [CrossRef]
Figure 1. Part−prototype network reasoning process during prediction (normal vs. pneumonia classification task from RX images). These models learn prototypes in terms of representative image regions for the predicted class from the training set and perform the classification based on their detection of new images (prototypical regions marked with yellow boxes).
Figure 1. Part−prototype network reasoning process during prediction (normal vs. pneumonia classification task from RX images). These models learn prototypes in terms of representative image regions for the predicted class from the training set and perform the classification based on their detection of new images (prototypical regions marked with yellow boxes).
Biomedinformatics 04 00115 g001
Figure 2. Global and local explanations of part-prototype network (classification of Alzheimer’s disease from MR images). The global explanation shows all the learned prototypes. The local explanation shows the model’s reasoning for a specific instance.
Figure 2. Global and local explanations of part-prototype network (classification of Alzheimer’s disease from MR images). The global explanation shows all the learned prototypes. The local explanation shows the model’s reasoning for a specific instance.
Biomedinformatics 04 00115 g002
Figure 3. ProtoPNet architecture.
Figure 3. ProtoPNet architecture.
Biomedinformatics 04 00115 g003
Figure 4. Prototypical part visualization in a normal vs. pneumonia classification task for a normal test image (marked with yellow box). Radiological images were displayed in the standard grayscale (windowing over the entire signal range) while activation maps and heatmaps were visualized using the same color map.
Figure 4. Prototypical part visualization in a normal vs. pneumonia classification task for a normal test image (marked with yellow box). Radiological images were displayed in the standard grayscale (windowing over the entire signal range) while activation maps and heatmaps were visualized using the same color map.
Biomedinformatics 04 00115 g004
Table 2. Conceptual properties (Co-12 Properties) for quantitatively assessing quality of part-prototype network, extracted from Nauta et al. [15].
Table 2. Conceptual properties (Co-12 Properties) for quantitatively assessing quality of part-prototype network, extracted from Nauta et al. [15].
Co-12 PropertyDescription
Content
Correctness:Since PP models are interpretable by design, the explanations are generated together with the prediction, and the reasoning process is correctly represented by design. Also, the faithfulness of the prototype visualization (from the latent representation to the input image patches), originally performed by carrying out bicubic upsampling, is not guaranteed by design and should be evaluated.
Completeness:The relation between the prototypes and classes is transparently shown, so the output-completeness is fulfilled by design, but the computation performed by the CNN backbone is not taken into consideration.
Consistency:PP models should not have random components in their designs, but nondeterminism may occur from the backbones’ initialisation and random seeds. It might be assessed by comparing explanations from models trained with different initializations or with different shuffling of the training data.
Continuity:It should be evaluated whether slightly perturbed inputs lead to the same explanation, given that the model makes the same classification.
Contrastivity:The incorporated interpretability of PP models results in a contrastivity incorporated by design; such a different classification corresponds to a different reasoning and, hence, to a different explanation. This evaluation might also include a target sensitivity analysis by inspecting where prototypes are detected in the test image.
Covariate complexity:The complexity of the features present in the prototypes is assessed with the ground truth, such as predefined concepts provided by human judgements (perceived homogeneity) or with object part annotations.
Presentation
Compactness:The number of prototypes which constitute the full classification model (global explanation size) in every input image (local explanation sizes) and the redundancy in the information content presented in different prototypes should be evaluated. The size of the explanation should be appropriate to not overwhelm the user.
Composition:How PP can be best presented to the user, and how these prototypes can be best structured and included in the reasoning process by comparing different explanation formats or by asking users about their preferences regarding the presentation and structure of the explanation should be assessed.
Confidence:Estimate the confidence of the explanation generation method, including measurements such as the prototype similarity scores.
User
Context:PP models should be evaluated with application-grounded user studies, similarly to evaluations with heatmaps, to understand their needs.
Coherence:Prototypes are often evaluated based on anecdotal evidence, with automated evaluation with an annotated dataset, or with manual evaluation. User studies might include the assessment of satisfaction, preference, and trust for part-prototypes.
Controllability:The ability to directly manipulate the explanation and the model’s reasoning e.g., enable users to suppress or modify learned prototypes, eventually with the aid of a graphical user interface.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

De Santi, L.A.; Piparo, F.I.; Bargagna, F.; Santarelli, M.F.; Celi, S.; Positano, V. Part-Prototype Models in Medical Imaging: Applications and Current Challenges. BioMedInformatics 2024, 4, 2149-2172. https://doi.org/10.3390/biomedinformatics4040115

AMA Style

De Santi LA, Piparo FI, Bargagna F, Santarelli MF, Celi S, Positano V. Part-Prototype Models in Medical Imaging: Applications and Current Challenges. BioMedInformatics. 2024; 4(4):2149-2172. https://doi.org/10.3390/biomedinformatics4040115

Chicago/Turabian Style

De Santi, Lisa Anita, Franco Italo Piparo, Filippo Bargagna, Maria Filomena Santarelli, Simona Celi, and Vincenzo Positano. 2024. "Part-Prototype Models in Medical Imaging: Applications and Current Challenges" BioMedInformatics 4, no. 4: 2149-2172. https://doi.org/10.3390/biomedinformatics4040115

APA Style

De Santi, L. A., Piparo, F. I., Bargagna, F., Santarelli, M. F., Celi, S., & Positano, V. (2024). Part-Prototype Models in Medical Imaging: Applications and Current Challenges. BioMedInformatics, 4(4), 2149-2172. https://doi.org/10.3390/biomedinformatics4040115

Article Metrics

Back to TopTop