This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Free Lunch in Pathology Foundation Model: Task-specific Model Adaptation with Concept-Guided Feature Enhancement

Yanyan Huang
The University of Hong Kong
yanyanh@connect.hku.hk
&Weiqin Zhao
The University of Hong Kong
wqzhao98@connect.hku.hk
&Yihang Chen
The University of Hong Kong
yihangc@connect.hku.hk
&Yu Fu
Lanzhou University
fuyu@lzu.edu.cn
&Lequan Yu
The University of Hong Kong
lqyu@hku.hk
Corresponding Auther.
Abstract

Whole slide image (WSI) analysis is gaining prominence within the medical imaging field. Recent advances in pathology foundation models have shown the potential to extract powerful feature representations from WSIs for downstream tasks. However, these foundation models are usually designed for general-purpose pathology image analysis and may not be optimal for specific downstream tasks or cancer types. In this work, we present Concept Anchor-guided Task-specific Feature Enhancement (CATE), an adaptable paradigm that can boost the expressivity and discriminativeness of pathology foundation models for specific downstream tasks. Based on a set of task-specific concepts derived from the pathology vision-language model with expert-designed prompts, we introduce two interconnected modules to dynamically calibrate the generic image features extracted by foundation models for certain tasks or cancer types. Specifically, we design a Concept-guided Information Bottleneck module to enhance task-relevant characteristics by maximizing the mutual information between image features and concept anchors while suppressing superfluous information. Moreover, a Concept-Feature Interference module is proposed to utilize the similarity between calibrated features and concept anchors to further generate discriminative task-specific features. The extensive experiments on public WSI datasets demonstrate that CATE significantly enhances the performance and generalizability of MIL models. Additionally, heatmap and umap visualization results also reveal the effectiveness and interpretability of CATE. The source code is available at https://github.com/HKU-MedAI/CATE.

1 Introduction

Multiple Instance Learning (MIL) lu2021data ; shmatko2022artificial ; lipkova2022artificial ; chen2022scaling is widely adopted for weakly supervised analysis in computational pathology, where the input of MIL is typically a set of patch features generated by a pre-trained feature extractor (i.e., image encoder). Although promising progress has been achieved, the effectiveness of MIL models heavily relies on the quality of the extracted features. A robust feature extractor can discern more distinctive pathological features, thereby improving the predictive capabilities of MIL models. Recently, several studies have explored using pretrained foundation models on large-scale pathology datasets with self-supervised learning as the feature extractors for WSI analysis wang2022transformer ; filiot2023scaling ; chen2024towards ; vorontsov2023virchow . Additionally, drawing inspiration from the success of Contrastive Language-Image Pretraining (CLIP) radford2021learning ; lai2023clipath in bridging visual and linguistic modalities, some works have aimed to develop a pathology vision-language foundation model (VLM) to simultaneously learn representations of pathology images and their corresponding captions ikezogwo2024quilt ; lu2024visual . The intrinsic consistency between the image feature space and caption embedding space in the pathology VLM enables the image encoder to extract more meaningful and discriminative features for downstream WSI analysis applications lu2024visual .

Although the development of these pathology foundation models has significantly advanced computational pathology, these models are designed for general-purpose pathology image analysis and may not be optimal for specific downstream tasks or cancer types, as the features extracted by the image encoder may contain generic yet task-irrelevant information that will harm the performance of specific downstream tasks. For example, as illustrated in Figure 1(a), the features extracted by the image encoder of a pathology VLM can include both task-relevant information (e.g., arrangement or morphology of tumor cells) and task-irrelevant elements(such as background information, stain styles, etc.). The latter information may act as "noise", distracting the learning process of MIL models tailored to specific tasks, and potentially impairing the generalization performance of these models across different data sources. Consequently, it is crucial to undertake task-specific adaptation to enhance feature extraction of generic foundation models and enable MIL models to concentrate on task-relevant information and thus improve analysis performance and generalization robey2021model ; wu2021collaborative .

Refer to caption
Figure 1: (a) Illustration of the key idea of concept-guided information bottleneck to enhance the task-relevant information and discard the task-irrelevant information. (b) Task-specific model adaptation with CATE to enhance the generalization across different data sources.

In this paper, we propose a novel paradigm, named Concept Anchor-guided Task-specific Feature Enhancement (CATE), to enhance the generic features extracted by the pathology VLM for specific downstream tasks (e.g., cancer subtyping). Without requiring additional supervision or significant computational resources, CATE offers an approximately "free lunch" in the context of pathology VLM. Specifically, we first derive a set of task-specific concept anchors from the pathology VLM with task-specific prompts, and these prompts rely on human expert design or are generated through querying large language models (LLMs), necessitating a certain level of pathological background knowledge. Based on these concept anchors, we design two concept-driven modules, i.e., the Concept-guided Information Bottleneck (CIB) module and the Concept-Feature Interference (CFI) module, to calibrate and generate task-specific features for downstream analysis. Particularly, with the task-specific concepts as the guidance, the CIB module enhances task-relevant features by maximizing the mutual information between the image features and the concept anchors and also eliminates task-irrelevant information by minimizing the superfluous information, as shown in Figure 1(a). Moreover, the CFI module further generates discriminative task-specific features by utilizing the similarities between the calibrated image features and concept anchors (i.e., concept scores). By incorporating the CATE into existing MIL frameworks, we not only obtain more discriminative features but also improve generalization regarding domain shift by eliminating task-irrelevant features and concentrating on pertinent information, as shown in Figure 1(b).

In summary, the main contributions of this work are threefold:

  • We introduce a novel method, named CATE, for model adaptation in computational pathology. To the best of our knowledge, this is the first initiative to conduct task-specific feature enhancement based on the pathology foundation model for MIL tasks.

  • We design a new CIB module to enhance the task-relevant information and discard irrelevant information with the guidance of task-specific concepts, and a new CFI module to generate task-specific features by exploiting the similarities between image features and concept anchors.

  • Extensive experiments on Whole Slide Image (WSI) analysis tasks demonstrate that CATE significantly enhances the performance and generalization capabilities of MIL models.

2 Related Work

Multiple Instance Learning (MIL) for WSI Analysis. MIL is the predominant paradigm for WSI analysis, treating each WSI as a bag of patch instances and classifying the entire WSI based on aggregated patch-level features. Attention-based methods ilse2018attention ; lu2021data ; javed2022additive ; yufei2022bayes are highly regarded for their ability to determine the significance of each instance within the bag. For instance, Ilse et al. ilse2018attention introduced an attention-based MIL model, while Lu et al. lu2021data proposed clustering-constrained-attention to refine this mechanism further. To model the relationships among instances, graph-based and Transformer-based methods have been developed hou2022h2 ; chen2022scaling ; huang2023conslide . For example, Chen et al. chen2022scaling introduced a Transformer-based hierarchical network to capitalize on the inherent hierarchical structure of WSIs.

Pathology Foundation Model. With the advancement of foundation models in computer vision, several pathology foundation models have been developed to serve as robust image encoders for WSI analysis. Riasatian et al. riasatian2021fine proposed fine-tuning the DenseNet huang2017densely on the TCGA dataset, while Filiot et al. filiot2023scaling utilized iBOT zhou2021ibot to pretrain a vision Transformer using the Masked Image Modeling framework. Recently, Chen et al. chen2024towards pre-trained a general-purpose foundation model on large-scale pathology datasets using DINOv2 oquab2023dinov2 , which has demonstrated strong and readily usable representations for WSI analysis. Inspired by CLIP radford2021learning , Ikezogwo et al. ikezogwo2024quilt , Huang et al. huang2023visual , and Lu et al. lu2024visual developed vision-language foundation models by training on large-scale pathology datasets with image-caption pairs. These foundation models have demonstrated superior performance in downstream tasks due to their ability to extract more discriminative features for WSI analysis.

Feature Enhancement in Computational Pathology. Several methods have been developed to obtain more discriminative features for WSI analysis by adapting pathology foundation models zhang2023text ; lu2024pathotune or designing new plug-and-play modules tang2024feature . For instance, Zhang et al. zhang2023text suggested aligning the image features with text features extracted from a pre-trained natural language model to enhance the feature representation of WSI patch images, while it operates solely at the patch level, without considering the informational relationship between image and text features. Recently, Tang et al. tang2024feature introduced Re-embedded Regional Transformer for feature re-embedding, aimed at enhancing WSI analysis when integrated with existing MIL methods. However, while this method considers the spatial information of WSIs and adds flexibility to MIL models, it falls short in extracting task-specific discriminative information for WSI analysis.

3 Method

3.1 Overview

The proposed CATE can be seamlessly integrated with any MIL framework to adapt the existing pathology foundation model (Pathology VLM) for performance-improved WSI analysis via task-specific enhancement. Specifically, consider a training set 𝒟={(𝐱,𝐲)}\mathcal{D}=\{\left(\boldsymbol{\rm{x}},\boldsymbol{\rm{y}}\right)\} of WSI-label pairs, where 𝐱={𝒙1,𝒙2,,𝒙N}\boldsymbol{\rm{x}}=\{\boldsymbol{x}_{1},\boldsymbol{x}_{2},...,\boldsymbol{x}_{N}\} is a set of patch features with dimension of CC (i.e., 𝒙iC\boldsymbol{x}_{i}\in\mathbb{R}^{C}) extracted by the image encoder of pathology VLM, NN denotes the number of patches, and 𝐲\boldsymbol{\rm{y}} is the corresponding label. The objective of CATE is to obtain the corresponding enhanced task-specific feature set 𝐳\boldsymbol{\rm{z}} from the original feature 𝐱\boldsymbol{\rm{x}} with the guidance of pre-extracted concepts anchors 𝐜\boldsymbol{\rm{c}} (see description below) for downstream usage:

𝐳=CATE(𝐱,𝐜),𝐲^=MIL(𝐳).\boldsymbol{\rm{z}}=\operatorname{\textbf{CATE}}\left(\boldsymbol{\rm{x}},\boldsymbol{\rm{c}}\right),\hat{\boldsymbol{\rm{y}}}=\operatorname{MIL}\left(\boldsymbol{\rm{z}}\right). (1)

As illustrated in Figure 2, we design two different modules to enhance the extracted features from foundation models: (1) Concept-guided Information Bottleneck (CIB) module calibrates original image features with the guidance of concept anchors with information bottleneck principle; and (2) Concept-Feature Interference (CFI) module generates discriminative task-specific features by leveraging the similarities between the calibrated image features and concept anchors. Specifically, the enhanced patch features can be represented as 𝐳={𝒛1,𝒛2,,𝒛N}\boldsymbol{\rm{z}}=\{\boldsymbol{z}_{1},\boldsymbol{z}_{2},...,\boldsymbol{z}_{N}\}, where 𝒛i\boldsymbol{z}_{i} is the concatenation of the calibrated feature 𝜶i\boldsymbol{\alpha}_{i} and the interference feature 𝜷i\boldsymbol{\beta}_{i} generated by CIB and CFI module:

𝒛i=Concat[𝜶i,𝜷i]=Concat[CIB(𝒙i,𝐜),CFI(𝒙i,𝐜)].\boldsymbol{z}_{i}=\operatorname{Concat}\left[\boldsymbol{\alpha}_{i},\boldsymbol{\beta}_{i}\right]=\operatorname{Concat}\left[\operatorname{\textbf{CIB}}\left(\boldsymbol{x}_{i},\boldsymbol{\rm{c}}\right),\operatorname{\textbf{CFI}}\left(\boldsymbol{x}_{i},\boldsymbol{\rm{c}}\right)\right]. (2)
Refer to caption
Figure 2: (a) Overview of CATE: the outputs of the CIB and CFI modules are concatenated to form the enhanced feature for downstream MIL models. (b) Task-relevant concept generation. (c) Concept-guided Information Bottleneck (CIB) module. (c) Concept-Feature Interference (CFI) module.

Concept Extraction. We extract two kinds of task-specific concept anchors, 𝐜={𝐜cs,𝐜ca}\boldsymbol{\rm{c}}=\{\boldsymbol{\rm{c}}^{\text{cs}},\boldsymbol{\rm{c}}^{\text{ca}}\}, comprising class-specific concepts 𝐜cs={𝒄ics}i=1m\boldsymbol{\rm{c}}^{\text{cs}}=\{\boldsymbol{c}^{\text{cs}}_{i}\}_{i=1}^{m} (e.g., subtyping classes) and class-agnostic concepts 𝐜ca={𝒄ica}i=1n\boldsymbol{\rm{c}}^{\text{ca}}=\{\boldsymbol{c}^{\text{ca}}_{i}\}_{i=1}^{n} (e.g., adipose, connective, and normal tissues), with mm and nn representing the numbers of class-specific and class-agnostic concepts, respectively. These concepts are generated by the text encoder of pathology VLM with prompt 𝐩\boldsymbol{\rm{p}}. Each prompt consists of a class name (e.g., "invasive ductal carcinoma") and a template (e.g., "An image of <CLASSNAME>"). To obtain more robust concepts, we use multiple prompts for each class and the final concept anchor is the average of the embeddings generated by different prompts. Details of class names and templates for various tasks are provided in Appendix G. Note that due to the inherent consistency between the image and text embedding space in VLM, these extracted concepts can also be regarded as image concept vectors.

3.2 Concept-guided Information Bottleneck

The objective of this module is to find a distribution p(𝜶|𝐱)p(\boldsymbol{\rm{\alpha}}|\boldsymbol{\rm{x}}) that maps the original image feature 𝐱\boldsymbol{\rm{x}} into a representation 𝜶\boldsymbol{\rm{\alpha}}, which contains enhanced task-discriminative characteristics and suppressed task-irrelevant information. WSIs typically contain various cell types or tissues (e.g., tumor cells, normal cells, adipose tissue, connective tissue), while only a subset of patches (e.g., with tumor cells) is crucial for certain tasks such as tumor subtyping. We thus define 𝐱^={𝒙^i}i=1k𝐱\hat{\boldsymbol{\rm{x}}}=\{\hat{\boldsymbol{x}}_{i}\}_{i=1}^{k}\subseteq\boldsymbol{\rm{x}} as the representative subset of the original feature set (e.g., tumor tissue patches), where kk denotes the number of representative patches. Note that this selection can be conducted with a simple comparison of image features with class-specific concepts (see discussion below). To this end, the corresponding enhanced feature set is 𝜶^={𝜶^i}i=1k𝜶\hat{\boldsymbol{\rm{\alpha}}}=\{\hat{\boldsymbol{\alpha}}_{i}\}_{i=1}^{k}\subseteq\boldsymbol{\rm{\alpha}} and we want to find the conditional distribution p(𝜶^i|𝒙^i)p(\hat{\boldsymbol{\alpha}}_{i}|\hat{\boldsymbol{x}}_{i}) to map the selected patch feature 𝒙^iC\hat{\boldsymbol{x}}_{i}\in\mathbb{R}^{C} into a more discriminative enhanced feature 𝜶^iC\hat{\boldsymbol{\alpha}}_{i}\in\mathbb{R}^{C}, which is discriminative enough to identify the label 𝐲\boldsymbol{\rm{y}}.

Sufficiency and Consistency Requirements. To quantify the informativeness requirement of the calibrated feature 𝜶^\hat{\boldsymbol{\alpha}}, we consider the sufficiency of 𝜶^\hat{\boldsymbol{\alpha}} for 𝐲\boldsymbol{\rm{y}}. As defined in Appendix E.1, the encoded feature 𝜶^\hat{\boldsymbol{\alpha}} derived from the original feature 𝒙^\hat{\boldsymbol{x}} is sufficient for determining the label 𝐲\boldsymbol{\rm{y}} if and only if the amount of task-specific information remains unchanged after calibration, i.e., I(𝒙^;𝐲)=I(𝜶^;𝐲)I(\hat{\boldsymbol{x}};\boldsymbol{\rm{y}})=I(\hat{\boldsymbol{\alpha}};\boldsymbol{\rm{y}}). However, the label 𝐲\boldsymbol{\rm{y}} pertains to the slide level and specific labels cannot be assigned to each instance due to the absence of patch-level annotations.

To address this challenge, we propose using the task-specific concept anchor as the guidance for each single 𝜶^\hat{\boldsymbol{\alpha}}. Specifically, we posit that the concept anchor 𝐜\boldsymbol{\rm{c}} is distinguishable for the task and contains task-relevant information for label 𝐲\boldsymbol{\rm{y}}. Given the consistency between image and text features in pathology VLM, any representation 𝜶^\hat{\boldsymbol{\alpha}} containing all information accessible from both image feature 𝒙^\hat{\boldsymbol{x}} and concept 𝐜\boldsymbol{\rm{c}} will also encapsulate the discriminative information required for the label. This consistency requirement is detailed in Appendix E.1. Thus, if 𝜶^\hat{\boldsymbol{\alpha}} is sufficient for 𝐜\boldsymbol{\rm{c}} (i.e., I(𝒙^;𝐜|𝜶^)=0I(\hat{\boldsymbol{x}};\boldsymbol{\rm{c}}|\hat{\boldsymbol{\alpha}})=0), then 𝜶^\hat{\boldsymbol{\alpha}} is as predictive for label 𝐲\boldsymbol{\rm{y}} as the joint of original feature 𝒙^\hat{\boldsymbol{x}} and concept anchor 𝐜\boldsymbol{\rm{c}}. Applying the chain rule of mutual information, we derive:

I(𝒙^;𝜶^)=I(𝜶^;𝐜)Predictive Information+I(𝒙^;𝜶^|𝐜)Superfluous Information.I(\hat{\boldsymbol{x}};\hat{\boldsymbol{\alpha}})=\underbrace{I(\hat{\boldsymbol{\alpha}};\boldsymbol{\rm{c}})}_{{\color[rgb]{.75,0,.25}\text{{Predictive Information}}}}+\underbrace{I(\hat{\boldsymbol{x}};\hat{\boldsymbol{\alpha}}|\boldsymbol{\rm{c}})}_{\color[rgb]{0.0600000000000001,0.46,1}{\text{{Superfluous Information}}}}. (3)

According to the consistency between the concept anchor and original feature, the mutual information term I(𝜶^;𝐜)I(\hat{\boldsymbol{\alpha}};\boldsymbol{\rm{c}}) represents the predictive information for the task, while the conditional information term I(𝒙^;𝜶^|𝐜)I(\hat{\boldsymbol{x}};\hat{\boldsymbol{\alpha}}|\boldsymbol{\rm{c}}) denotes task-irrelevant information (i.e., superfluous information) in original patch feature 𝒙^\hat{\boldsymbol{x}}, which can be minimized to enhance the robustness and generalization ability of downstream MIL models. As a result, the main objective of the feature calibration in this module can be formalized as maximize predictive information I(𝜶^;𝐜)I(\hat{\boldsymbol{\alpha}};\boldsymbol{\rm{c}}) while minimize the superfluous information I(𝒙^;𝜶^|𝐜)I(\hat{\boldsymbol{x}};\hat{\boldsymbol{\alpha}}|\boldsymbol{\rm{c}}).

Predictive Information Maximization (PIM). The predictive information in Equ (3) equals to the mutual information between the calibrated feature and concept anchors. To maximize this, we choose the InfoNCE oord2018representation to estimate the lower bound of the mutual information, which can be obtained by comparing positive pairs sampled from the joint distribution 𝜶^,𝒄poscsp(𝜶^,𝒄cs)\hat{\boldsymbol{\rm{\alpha}}},\boldsymbol{c}^{\text{cs}}_{\text{pos}}\sim p(\hat{\boldsymbol{\rm{\alpha}}},\boldsymbol{c}^{\text{cs}}) to pairs 𝜶^,𝒄jcs\hat{\boldsymbol{\rm{\alpha}}},\boldsymbol{c}^{\text{cs}}_{j} and 𝜶^,𝒄jca\hat{\boldsymbol{\rm{\alpha}}},\boldsymbol{c}^{\text{ca}}_{j} built using a set of negative class concepts 𝒄jcsp(𝒄cs)\boldsymbol{c}^{\text{cs}}_{j}\sim p\left(\boldsymbol{c}^{\text{cs}}\right) and class-agnostic concepts 𝒄jcap(𝒄ca)\boldsymbol{c}^{\text{ca}}_{j}\sim p\left(\boldsymbol{c}^{\text{ca}}\right):

INCE(𝜶^;𝐜)=𝔼𝜶^,𝒄poscs[i=1klogf(𝒄poscs,𝜶^i)j=1mf(𝒄jcs,𝜶^i)+j=1nf(𝒄jca,𝜶^i)].I_{NCE}\left(\hat{\boldsymbol{\rm{\alpha}}};\boldsymbol{\rm{c}}\right)=\operatornamewithlimits{\mathbb{E}}_{\hat{\boldsymbol{\rm{\alpha}}},\boldsymbol{c}^{\text{cs}}_{\text{pos}}}\left[\sum_{i=1}^{k}\log\frac{f\left(\boldsymbol{c}^{\text{cs}}_{\text{pos}},\hat{\boldsymbol{\alpha}}_{i}\right)}{\sum_{j=1}^{m}f\left(\boldsymbol{c}^{\text{cs}}_{j},\hat{\boldsymbol{\alpha}}_{i}\right)+\sum_{j=1}^{n}f\left(\boldsymbol{c}^{\text{ca}}_{j},\hat{\boldsymbol{\alpha}}_{i}\right)}\right]. (4)

We set f(𝒄,𝜶^i)=exp(𝜶^i𝖳𝒄/τ)f\left(\boldsymbol{c},\hat{\boldsymbol{\alpha}}_{i}\right)=\exp{\left({\hat{\boldsymbol{\alpha}}_{i}^{\mathsf{T}}\boldsymbol{c}}/{\tau}\right)} with τ>0\tau>0 in practice following oord2018representation . By maximizing this mutual information lower bound, f(𝒄,𝜶^i)f\left(\boldsymbol{c},\hat{\boldsymbol{\alpha}}_{i}\right) will be proportional to the density ratio p(𝒄,𝜶^i)/p(𝒄)p(𝜶^i){p\left(\boldsymbol{c},\hat{\boldsymbol{\alpha}}_{i}\right)}/{p\left(\boldsymbol{c}\right)p\left(\hat{\boldsymbol{\alpha}}_{i}\right)} as proved in oord2018representation . Hence, f(𝒄,𝜶^i)f\left(\boldsymbol{c},\hat{\boldsymbol{\alpha}}_{i}\right) preserves the mutual information between the calibrated feature and concept anchor. The detailed derivation can be found in Appendix E.2. The loss function for PIM can be denoted as:

PIM=𝔼𝜶^,𝒄poscs[i=1k𝜶^i𝖳𝒄poscsτ]+𝔼𝜶^,𝒄poscs[i=1klog(j=1mexp𝜶^i𝖳𝒄jcsτ+j=1nexp𝜶^i𝖳𝒄jcaτ)].\mathcal{L}_{PIM}=\operatornamewithlimits{\mathbb{E}}_{\hat{\boldsymbol{\rm{\alpha}}},\boldsymbol{c}^{\text{cs}}_{\text{pos}}}\left[-\sum_{i=1}^{k}\frac{\hat{\boldsymbol{\alpha}}_{i}^{\mathsf{T}}\boldsymbol{c}^{\text{cs}}_{\text{pos}}}{\tau}\right]+\operatornamewithlimits{\mathbb{E}}_{\hat{\boldsymbol{\rm{\alpha}}},\boldsymbol{c}^{\text{cs}}_{\text{pos}}}\left[\sum_{i=1}^{k}\log\left(\sum_{j=1}^{m}\exp{\frac{\hat{\boldsymbol{\alpha}}_{i}^{\mathsf{T}}\boldsymbol{c}^{\text{cs}}_{j}}{\tau}}+\sum_{j=1}^{n}\exp{\frac{\hat{\boldsymbol{\alpha}}_{i}^{\mathsf{T}}\boldsymbol{c}^{\text{ca}}_{j}}{\tau}}\right)\right]. (5)

Superfluous Information Minimization (SIM). To compress task-irrelevant information, we aim to minimize the superfluous information term as defined in Equ (3). This objective can be achieved by minimizing the mutual information I(𝒙^;𝜶^)I(\hat{\boldsymbol{x}};\hat{\boldsymbol{\alpha}}). In practice, we conduct SIM for all patches in the subset 𝐱\boldsymbol{\rm{x}}, as each patch may contain task-irrelevant information. Following alemi2016deep , it can be represented as:

I(𝒙;𝜶)=p(𝒙,𝜶)logp(𝜶|𝒙)𝑑𝒙𝑑𝜶p(𝜶)logp(𝜶)𝑑𝜶.I({\boldsymbol{x}};{\boldsymbol{\alpha}})=\int p\left({\boldsymbol{x}},{\boldsymbol{\alpha}}\right)\log p\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right)d{\boldsymbol{x}}d{\boldsymbol{\alpha}}-\int p\left({\boldsymbol{\alpha}}\right)\log p\left({\boldsymbol{\alpha}}\right)d{\boldsymbol{\alpha}}. (6)

After that, we let the distribution of 𝜶{\boldsymbol{\alpha}}: r(𝜶)r\left({\boldsymbol{\alpha}}\right) (e.g., Gaussian distribution in this work), be a variational approximation to the marginal distribution p(𝜶)p\left({\boldsymbol{\alpha}}\right), and we can obtain the upper bound for I(𝒙;𝜶)I({\boldsymbol{x}};{\boldsymbol{\alpha}}):

I(𝒙;𝜶)p(𝒙)p(𝜶|𝒙)logp(𝜶|𝒙)r(𝜶)d𝒙d𝜶.I({\boldsymbol{x}};{\boldsymbol{\alpha}})\leq\int p\left({\boldsymbol{x}}\right)p\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right)\log\frac{p\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right)}{r\left({\boldsymbol{\alpha}}\right)}d{\boldsymbol{x}}d{\boldsymbol{\alpha}}. (7)

Furthermore, we use a variational distribution qθ(𝜶|𝒙)q_{\theta}\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right) with parameter θ\theta to approximate p(𝜶|𝒙)p\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right) and we implement the parameterization of the variational distribution with MLP by predicting the mean and variance of the Gaussian distribution and sample the calibrated feature 𝜶{\boldsymbol{\alpha}} from this distribution:

𝜶𝒩(MLPμ(𝒙),MLPΣ(𝒙)).{\boldsymbol{\alpha}}\sim\mathcal{N}\left(MLP^{\mu}\left({\boldsymbol{x}}\right),MLP^{\Sigma}\left({\boldsymbol{x}}\right)\right). (8)

In practice, we implement this by utilizing the reparameterization trick kingma2013auto to obtain an unbiased estimate of the gradient and further optimize the variational distribution. The detailed derivation can be found in Appendix E.3. The minimization of the upper bound of I(𝒙;𝜶)I({\boldsymbol{x}};{\boldsymbol{\alpha}}) equals to the minimization of the Kullback-Leibler divergence between qθ(𝜶|𝒙)q_{\theta}\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right) and r(𝜶)r\left({\boldsymbol{\alpha}}\right). Therefore, the loss function can be represented as:

SIM=𝔼[i=1kDKL(qθ(𝜶i|𝒙i)||r(𝜶i))].\mathcal{L}_{SIM}=\mathbb{E}\left[\sum_{i=1}^{k}D_{KL}\left(q_{\theta}\left({\boldsymbol{\alpha}}_{i}|{\boldsymbol{x}}_{i}\right)||r\left({\boldsymbol{\alpha}}_{i}\right)\right)\right]. (9)

Discussion. We further provide explanation of CIB module with the information plane geiger2021information ; federici2020learning in Appendix F. It should be noted that the PIM supervises only the representative subset 𝐱^\hat{\boldsymbol{\rm{x}}} containing task-relevant information (selected by the similarity between image features and corresponding class-specific concepts). Meanwhile, the SIM is applied to all patches in 𝐱\boldsymbol{\rm{x}}, as any patch may carry information irrelevant to the task (e.g., background information and stain styles). Besides, SIM cannot be directly optimized without the guidance of concept anchors (i.e., PIM) due to the absence of patch-level labels. As demonstrated in the ablation study in Section 4.4, the absence of concept anchor guidance leads to the collapse of discriminative information in the calibrated feature, adversely affecting downstream task performance. By maximizing predictive information and minimizing superfluous details, the CIB module effectively enhances the discriminative capacity of the original features and aligns them with the task-specific concept anchors for improved prediction.

3.3 Concept-Feature Interference

We also propose the Concept-Feature Interference (CFI) module to utilize the similarity characteristic between calibrated features and concept anchors to further obtain robust and discriminative information for the downstream tasks. Our primary focus is on the class-specific concept anchors 𝐜cs\boldsymbol{\rm{c}}^{\text{cs}}. Specifically, for each CIB encoded feature 𝜶i\boldsymbol{\alpha}_{i}, we calculate the cosine similarity between 𝜶i\boldsymbol{\alpha}_{i} and each class-specific concept 𝒄ics\boldsymbol{c}^{\text{cs}}_{i}. It is important to note that the number of class-specific concepts mm is larger than the number of classes, as we use multiple <CLASSNAME> and templates to generate the concept anchor for each class, as shown in the Appendix G. Thus, we can obtain the similarity vector by concatenating the similarity scores between 𝜶i\boldsymbol{\alpha}_{i} and each class-specific concept 𝒄ics\boldsymbol{c}^{\text{cs}}_{i}. To integrate the interference information (similarity relationship) into the enhanced feature, we align the similarity vector with the calibrated feature 𝜶i\boldsymbol{\alpha}_{i} using a Self-Normalizing Network (SNN) layer klambauer2017self . This allows us to obtain the final interference vector 𝜷i\boldsymbol{\beta}_{i} of CFI:

𝜷i=SNN(Concat[{Sim(𝜶i,𝒄ics)}i=1m]).\boldsymbol{\beta}_{i}=\operatorname{SNN}\left(\operatorname{Concat}\left[\{\operatorname{Sim}\left(\boldsymbol{\alpha}_{i},\boldsymbol{c}^{\text{cs}}_{i}\right)\}_{i=1}^{m}\right]\right). (10)

The interference vector contains superficial information that indicates the similarity between the calibrated feature and concept anchor directly. This is completely different from the calibrated feature of the CIB module, which contains discriminative latent information for the downstream tasks. Therefore, integrating the interference feature can further provide robust and discriminative information for the downstream tasks.

Discussion. The CFI module is designed to utilize the similarity characteristic between calibrated feature and concept anchor as a discriminative feature, which can be further integrated into the calibrated feature for downstream tasks. This is different from other studies that directly compare the similarity between visual features and textual concept features of different classes to perform zero-shot classification lu2024visual .

3.4 Training Objective

The overall training objective of the CATE framework can be represented as the combination of the cross entropy loss CE\mathcal{L}_{CE} for the downstream tasks, the predictive information maximization loss PIM\mathcal{L}_{PIM}, and the superfluous information minimization loss SIM\mathcal{L}_{SIM}:

=CE+λPPIM+λSSIM,\mathcal{L}=\mathcal{L}_{CE}+\lambda_{P}\mathcal{L}_{PIM}+\lambda_{S}\mathcal{L}_{SIM}, (11)

where λP\lambda_{P} and λS\lambda_{S} are hyperparameters and influence of them is discussed in Appendix C.

4 Experiments

Table 1: Cancer Subtyping Results on BRCA of MIL Models Incorporated with CATE.
Method CATE BRCA (NINDN_{\text{IND}}=1)
OOD-AUC Gain OOD-ACC Gain IND-AUC Gain IND-ACC Gain
ABMIL 0.914±\pm0.015 N/A 0.852±\pm0.014 N/A 0.963±\pm0.044 N/A 0.888±\pm0.053 N/A
CLAM 0.907±\pm0.017 N/A 0.802±\pm0.053 N/A 0.965±\pm0.049 N/A 0.888±\pm0.068 N/A
DSMIL 0.925±\pm0.020 N/A 0.836±\pm0.048 N/A 0.969±\pm0.040 N/A 0.900±\pm0.080 N/A
DTFD-MIL 0.912±\pm0.012 N/A 0.858±\pm0.020 N/A 0.944±\pm0.058 N/A 0.894±\pm0.070 N/A
TransMIL 0.918±\pm0.015 N/A 0.832±\pm0.046 N/A 0.969±\pm0.036 N/A 0.918±\pm0.067 N/A
R2T-MIL 0.901±\pm0.027 N/A 0.816±\pm0.051 N/A 0.965±\pm0.033 N/A 0.894±\pm0.022 N/A
ABMIL 0.951±\pm0.003 \uparrow4.05% 0.897±\pm0.026 \uparrow5.28% 0.998±\pm0.006 \uparrow3.63% 0.965±\pm0.045 \uparrow8.67%
CLAM 0.951±\pm0.005 \uparrow4.85% 0.906±\pm0.020 \uparrow12.97% 0.998±\pm0.006 \uparrow3.42% 0.965±\pm0.037 \uparrow8.67%
DSMIL 0.936±\pm0.007 \uparrow1.19% 0.866±\pm0.036 \uparrow3.59% 0.990±\pm0.022 \uparrow2.17% 0.959±\pm0.044 \uparrow6.56%
DTFD-MIL 0.947±\pm0.004 \uparrow3.84% 0.906±\pm0.009 \uparrow5.59% 0.985±\pm0.028 \uparrow4.34% 0.953±\pm0.042 \uparrow6.60%
TransMIL 0.938±\pm0.005 \uparrow2.18% 0.880±\pm0.023 \uparrow5.77% 0.998±\pm0.006 \uparrow2.99% 0.965±\pm0.027 \uparrow5.12%
Method CATE BRCA (NINDN_{\text{IND}}=2)
OOD-AUC Gain OOD-ACC Gain IND-AUC Gain IND-ACC Gain
ABMIL 0.899±\pm0.035 N/A 0.892±\pm0.019 N/A 0.967±\pm0.019 N/A 0.941±\pm0.024 N/A
CLAM 0.893±\pm0.030 N/A 0.862±\pm0.019 N/A 0.960±\pm0.042 N/A 0.935±\pm0.027 N/A
DSMIL 0.881±\pm0.032 N/A 0.852±\pm0.028 N/A 0.946±\pm0.057 N/A 0.940±\pm0.020 N/A
DTFD-MIL 0.909±\pm0.019 N/A 0.878±\pm0.014 N/A 0.973±\pm0.023 N/A 0.945±\pm0.041 N/A
TransMIL 0.904±\pm0.023 N/A 0.852±\pm0.090 N/A 0.966±\pm0.031 N/A 0.936±\pm0.052 N/A
R2T-MIL 0.902±\pm0.028 N/A 0.873±\pm0.027 N/A 0.946±\pm0.060 N/A 0.929±\pm0.048 N/A
ABMIL 0.943±\pm0.006 \uparrow4.89% 0.907±\pm0.018 \uparrow1.68% 0.981±\pm0.018 \uparrow1.45% 0.948±\pm0.030 \uparrow0.74%
CLAM 0.945±\pm0.008 \uparrow5.82% 0.896±\pm0.030 \uparrow3.94% 0.976±\pm0.023 \uparrow1.67% 0.938±\pm0.043 \uparrow0.32%
DSMIL 0.919±\pm0.015 \uparrow4.31% 0.869±\pm0.036 \uparrow2.00% 0.958±\pm0.051 \uparrow1.27% 0.949±\pm0.024 \uparrow0.96%
DTFD-MIL 0.946±\pm0.005 \uparrow4.07% 0.887±\pm0.027 \uparrow1.03% 0.977±\pm0.023 \uparrow0.41% 0.946±\pm0.036 \uparrow0.11%
TransMIL 0.920±\pm0.011 \uparrow1.77% 0.867±\pm0.046 \uparrow1.76% 0.968±\pm0.045 \uparrow0.21% 0.940±\pm0.026 \uparrow0.43%
  • *

    The best results are highlighted in bold, and the second-best results are underlined.

  • R2T-MIL is designed for feature re-embedding that utilize ABMIL as base MIL model.

4.1 Experimental Settings

Tasks and Datasets. We conducted cancer subtyping tasks on three public WSI datasets from The Cancer Genome Atlas (TCGA) project: Invasive Breast Carcinoma (BRCA), Non-Small Cell Lung Cancer (NSCLC), and Renal Cell Carcinoma (RCC). Detailed dataset information is available in Appendix D.

IND and OOD Settings. The datasets in the TCGA contains samples from different source sites (i.e., different hospitals or laboratories), which are indicated in the sample barcodes111https://docs.gdc.cancer.gov/Encyclopedia/pages/TCGA_Barcode/. And different source sites have different staining protocols and imaging characteristics, causing feature domain shifts between different sites cheng2021robust ; de2021deep . Therefore, MIL models trained on several sites may not generalize well to others. To better evaluate the true performance of the models, we selected several sites as IND data (in-domain, the testing and training data are from the same sites), and used data from other sites as OOD data (out-of-domain, the testing and training data are from different sites), and reported the testing performance on both IND and OOD data. Specifically, we designated NINDN_{\text{IND}} sites as IND and the remaining as OOD. Each experiment involved splitting the IND data into training, validation, and testing sets, training the models on IND data, and evaluating them on both IND and OOD testing data. For the BRCA dataset, we randomly selected one or two sites as IND data and used the remaining sites as OOD data. However, for NSCLC (2 categories) and RCC (3 categories) datasets, each site contains samples from only one subtype. Therefore, we cannot select only one site as IND data, as it will include one category/subtype in the training data. Instead, we randomly selected one or two corresponding sites for each category as IND data for NSCLC and RCC, and used the other sites as OOD data. Finally, we obtained 1 or 2 IND sites for BRCA, 2 or 4 for NSCLC, and 3 or 6 for RCC.

Evaluation. We report the area under the receiver operating characteristic curve (AUC) and accuracy for the OOD and IND test sets, respectively, with means and standard deviations over 10 runs of Monte-Carlo Cross Validation. Notably, the OOD performance is emphasized for NSCLC and RCC, where each site contains samples from only one cancer subtype. Traditional MIL models tend to recognize site-specific patterns (e.g., staining) as shortcuts and excel in in-domain evaluations, rather than identifying useful class-specific features, making performance less reflective of the models’ actual capability. Therefore, OOD performance more accurately reflects the models’ discriminative and generalization capabilities.

Comparisons. Given that CATE is an adaptable method, we evaluated the performance variations across various MIL models both with and without the integration of CATE. We specifically focused on the following state-of-the-art MIL models: the original ABMIL ilse2018attention , CLAM lu2021data , DSMIL li2021dual , TransMIL shao2021transmil , DTFD-MIL zhang2022dtfd , and R2T-MIL tang2024feature . The R2T-MIL tang2024feature is a feature re-embedding method that utilizes ABMIL as the base MIL model.

Implementation Details. This study begins the image feature extraction process by segmenting the foreground tissue and then splitting the WSI into 512×\times512 pixels patches at 20×\times magnification. Subsequently, these patches are processed through a pre-trained image encoder from CONCH lu2024visual to extract image features. For concept anchors, we utilize CONCH’s text encoder to derive task-relevant concepts from predefined text prompts, with detailed prompt information available in Appendix G. Model parameters are optimized using the Adam optimizer with a learning rate of 10510^{-5}. The batch size is set to 1, and all the experiments are conducted on a single NVIDIA RTX 3090 GPU.

Table 2: Cancer Subtyping Results on NSCLC and RCC.
Method NSCLC (NINDN_{\text{IND}}=2) NSCLC (NINDN_{\text{IND}}=4)
OOD-AUC OOD-ACC IND-AUC# IND-ACC# OOD-AUC OOD-ACC IND-AUC# IND-ACC#
ABMIL 0.874±\pm0.021 0.803±\pm0.021 0.997±\pm0.004 0.954±\pm0.028 0.951±\pm0.023 0.883±\pm0.029 0.974±\pm0.018 0.910±\pm0.036
CLAM 0.875±\pm0.020 0.801±\pm0.021 0.997±\pm0.007 0.963±\pm0.042 0.931±\pm0.037 0.870±\pm0.036 0.977±\pm0.023 0.926±\pm0.048
DSMIL 0.839±\pm0.046 0.764±\pm0.043 0.993±\pm0.004 0.963±\pm0.028 0.934±\pm0.019 0.864±\pm0.026 0.974±\pm0.013 0.913±\pm0.042
DTFD-MIL 0.903±\pm0.023 0.836±\pm0.026 0.990±\pm0.009 0.958±\pm0.049 0.949±\pm0.010 0.893±\pm0.012 0.981±\pm0.012 0.918±\pm0.040
TransMIL 0.790±\pm0.028 0.712±\pm0.024 0.997±\pm0.004 0.954±\pm0.033 0.917±\pm0.022 0.832±\pm0.031 0.977±\pm0.014 0.923±\pm0.029
R2T-MIL 0.739±\pm0.088 0.690±\pm0.075 0.999±\pm0.002 0.971±\pm0.036 0.892±\pm0.041 0.800±\pm0.059 0.977±\pm0.018 0.916±\pm0.045
CATE-MIL 0.945±\pm0.016 0.840±\pm0.043 0.985±\pm0.011 0.938±\pm0.037 0.969±\pm0.003 0.906±\pm0.011 0.967±\pm0.019 0.905±\pm0.054
Method RCC (NINDN_{\text{IND}}=3) RCC (NINDN_{\text{IND}}=6)
OOD-AUC OOD-ACC IND-AUC# IND-ACC# OOD-AUC OOD-ACC IND-AUC# IND-ACC#
ABMIL 0.973±\pm0.005 0.891±\pm0.017 0.997±\pm0.004 0.961±\pm0.032 0.971±\pm0.007 0.885±\pm0.010 0.973±\pm0.010 0.897±\pm0.023
CLAM 0.972±\pm0.004 0.893±\pm0.012 0.991±\pm0.005 0.961±\pm0.032 0.969±\pm0.009 0.888±\pm0.015 0.975±\pm0.011 0.896±\pm0.031
DSMIL 0.977±\pm0.002 0.893±\pm0.010 0.996±\pm0.006 0.965±\pm0.026 0.969±\pm0.008 0.883±\pm0.016 0.980±\pm0.012 0.901±\pm0.022
DTFD-MIL 0.975±\pm0.003 0.897±\pm0.012 0.996±\pm0.004 0.943±\pm0.046 0.971±\pm0.007 0.893±\pm0.017 0.974±\pm0.012 0.878±\pm0.022
TransMIL 0.961±\pm0.010 0.864±\pm0.022 0.994±\pm0.004 0.930±\pm0.030 0.947±\pm0.017 0.828±\pm0.037 0.975±\pm0.013 0.894±\pm0.027
R2T-MIL 0.956±\pm0.018 0.847±\pm0.022 0.991±\pm0.008 0.936±\pm0.030 0.932±\pm0.020 0.803±\pm0.048 0.974±\pm0.012 0.897±\pm0.029
CATE-MIL 0.983±\pm0.002 0.911±\pm0.018 0.989±\pm0.009 0.944±\pm0.031 0.979±\pm0.007 0.905±\pm0.017 0.963±\pm0.011 0.882±\pm0.032
  • *

    The best results are highlighted in bold, and the second-best results are underlined.

  • R2T-MIL is designed for feature re-embedding that utilize ABMIL as base MIL model.

  • #

    The in-domain performance of NSCLC and RCC does not represent the true ability of the models, as each site contains only samples from one cancer subtype. We primarily focus on the OOD performance for these two datasets.

4.2 Experimental Results

Quantitative Results on BRCA Dataset. To fully evaluate the effectiveness of CATE, we assessed its impact on several state-of-the-art MIL models using the BRCA dataset. The results are shown in Table 1, where the MIL models integrated with CATE outperform their original counterparts in both in-domain (IND) and out-of-domain (OOD) testing, which demonstrates the effectiveness and generalization capabilities of CATE. Comparing with R2T-MIL, which is a feature re-embedding method that utilizes ABMIL as the base MIL model, CATE incorporated with ABMIL consistently achieves better performance in terms of both OOD and IND testing. To further investigate the effectiveness of CATE, we conducted experiments by altering the in-domain sites and applying traditional settings. Detailed results are available in Appendix B.

Qualitative Analysis. To qualitatively investigate the effectiveness of CATE, we visualized attention heatmaps, UMAP, and the similarities between original features and corresponding class concept features, as well as calibrated features and class concept features, as shown in Figure 3. Additional visualization results are provided in Appendix H. As shown in Figure 3 (a&b), attention heatmap comparisons reveal that CATE-MIL focuses more intensely on cancerous regions, with a clearer delineation between high and low attention areas. By comparing the similarities of original and calibrated features to class concept features in Figure 3 (c&d), it is evident that the enhanced similarity in cancerous regions is significantly higher than in original features. Moreover, the disparity between cancerous and non-cancerous regions’ similarities is also expanded, which further verifies the ability of CATE to enhance task-relevant information and suppress irrelevant information. We further performed a UMAP visualization of class concept features, original features, and calibrated features. As depicted in Figure 3 (f), calibrated features are notably closer to the corresponding class (IDC) concept features compared to the original features, which demonstrates CATE’s ability to effectively align features with task-relevant concepts and enhance task-relevant information.

Refer to caption
Figure 3: (a) Attention heatmap of CATE-MIL. (b) Attention heatmap of the original ABMIL. (c) similarity between the calibrated features and the corresponding class concept feature. (d) similarity between the original features and the corresponding class concept feature. (e) Original WSI. (f) UMAP visualization of class concept features, original features, and enhanced features.

4.3 Results on Additional Datasets

For clarity and to highlight the superiority of ABMIL when enhanced with CATE, we developed CATE-MIL by incorporating CATE into ABMIL and compared it against other leading MIL models on NSCLC and RCC datasets. The comparative results in Table 2 confirm that CATE-MIL consistently outperforms other models in both OOD and IND performance. However, it is noted that CATE-MIL performs poorly on the in-domain testing data for NSCLC and RCC. This underperformance may be attributed to the elimination of task-irrelevant information, including site-specific patterns, by CATE, potentially degrading performance on in-domain data for these datasets. Consequently, OOD performance more accurately reflects the discriminative and generalization capabilities of the models.

4.4 Ablation Analysis

We conduct ablation studies to assess the effectiveness of each component within CATE, and the results are shown in Table 3. Initially, incorporating Predictive Information Maximization (PIM) enables ABMIL to achieve improved performance in most experiments, which demonstrates PIM’s efficacy in extracting task-relevant information. However, using Superfluous Information Minimization (SIM) alone results in performance degradation across most experiments, which suggests that SIM may discard some task-relevant information without guidance from a task-relevant concept anchor. Incorporating both PIM and SIM consistently enhances ABMIL’s performance in all experiments, which further verifies that their combination effectively boosts the generalization capabilities of MIL models. We also conduct experiments by only using the interference features in CFI as the input of ABMIL, and the results show that the interference features are also informative for WSI classification tasks. More ablation analysis about the weights of PIM and SIM in CIB module the number of representative patches are in Appendix C.

Table 3: Ablation study of CATE.
Method PIM SIM CFI BRCA (NINDN_{\text{IND}}=1) BRCA (NINDN_{\text{IND}}=2) NSCLC (NINDN_{\text{IND}}=2) NSCLC (NINDN_{\text{IND}}=4) RCC (NINDN_{\text{IND}}=3) RCC (NINDN_{\text{IND}}=6)
ABMIL 0.914±\pm0.015 0.899±\pm0.035 0.874±\pm0.021 0.951±\pm0.023 0.973±\pm0.005 0.971±\pm0.007
0.932±\pm0.011 0.939±\pm0.012 0.906±\pm0.026 0.901±\pm0.047 0.976±\pm0.005 0.973±\pm0.005
0.895±\pm0.022 0.885±\pm0.140 0.656±\pm0.041 0.898±\pm0.028 0.952±\pm0.014 0.954±\pm0.015
0.936±\pm0.010 0.942±\pm0.010 0.910±\pm0.030 0.960±\pm0.011 0.979±\pm0.004 0.977±\pm0.005
0.913±\pm0.024 0.884±\pm0.032 0.850±\pm0.036 0.918±\pm0.029 0.975±\pm0.008 0.941±\pm0.022
CATE-MIL 0.951±\pm0.003 0.943±\pm0.006 0.945±\pm0.016 0.970±\pm0.003 0.983±\pm0.002 0.979±\pm0.007

5 Conclusion and Discussion

In this paper, we introduce CATE, a new approach that offers a "free lunch" for task-specific adaptation of pathology VLM by leveraging the inherent consistency between image and text modalities. CATE shows the potential to enhance the generic features extracted by pathology VLM for specific downstream tasks, using task-specific concept anchors as guidance. The proposed CIB module calibrates the image features by enhancing task-relevant information while suppressing task-irrelevant information, while the CFI module obtains the interference vector for each patch to generate discriminative task-specific features. Extensive experiments on WSI datasets demonstrate the effectiveness of CATE in improving the performance and generalizability of state-of-the-art MIL methods.

Limitations and Social Impact. The proposed CATE offers a promising solution to customize the pathology VLM for specific tasks, significantly improving the performance and applicability of MIL methods in WSI analysis. However, the performance of CATE heavily depends on the quality of the concept anchors, which, in turn, relies on domain knowledge and the quality of the pre-trained pathology VLM. Additionally, while CATE is optimized for classification tasks such as cancer subtyping, it may not be readily applicable to other analytical tasks, such as survival prediction. However, there might be a potential solution to address this challenge. For instance, we could leverage LLMs or retrieval-based LLMs to generate descriptive prompts about the general morphological appearance of WSIs for specific cancer types. By asking targeted questions, we can summarize reliable and general morphological descriptions associated with different survival outcomes or biomarker expressions and further verify these prompts with pathologists. Moreover, since medical data may contain sensitive information, ensuring the privacy and security of such data is crucial.

6 Acknowledgements

This work was supported in part by the Research Grants Council of Hong Kong (27206123 and T45-401/22-N), in part by the Hong Kong Innovation and Technology Fund (ITS/274/22), in part by the National Natural Science Foundation of China (No. 62201483), and in part by Guangdong Natural Science Fund (No. 2024A1515011875).

References

  • [1] Alexander A Alemi, Ian Fischer, Joshua V Dillon, and Kevin Murphy. Deep variational information bottleneck. arXiv preprint arXiv:1612.00410, 2016.
  • [2] Richard J Chen, Chengkuan Chen, Yicong Li, Tiffany Y Chen, Andrew D Trister, Rahul G Krishnan, and Faisal Mahmood. Scaling vision transformers to gigapixel images via hierarchical self-supervised learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 16144–16155, 2022.
  • [3] Richard J Chen, Tong Ding, Ming Y Lu, Drew FK Williamson, Guillaume Jaume, Andrew H Song, Bowen Chen, Andrew Zhang, Daniel Shao, Muhammad Shaban, et al. Towards a general-purpose foundation model for computational pathology. Nature Medicine, 30(3):850–862, 2024.
  • [4] Shenghua Cheng, Sibo Liu, Jingya Yu, Gong Rao, Yuwei Xiao, Wei Han, Wenjie Zhu, Xiaohua Lv, Ning Li, Jing Cai, et al. Robust whole slide image analysis for cervical cancer screening using deep learning. Nature communications, 12(1):5639, 2021.
  • [5] Kevin de Haan, Yijie Zhang, Jonathan E Zuckerman, Tairan Liu, Anthony E Sisk, Miguel FP Diaz, Kuang-Yu Jen, Alexander Nobori, Sofia Liou, Sarah Zhang, et al. Deep learning-based transformation of h&e stained tissues into special stains. Nature communications, 12(1):1–13, 2021.
  • [6] Marco Federici, Anjan Dutta, Patrick Forré, Nate Kushman, and Zeynep Akata. Learning robust representations via multi-view information bottleneck. arXiv preprint arXiv:2002.07017, 2020.
  • [7] Alexandre Filiot, Ridouane Ghermi, Antoine Olivier, Paul Jacob, Lucas Fidon, Alice Mac Kain, Charlie Saillard, and Jean-Baptiste Schiratti. Scaling self-supervised learning for histopathology with masked image modeling. medRxiv, pages 2023–07, 2023.
  • [8] Bernhard C Geiger. On information plane analyses of neural network classifiers—a review. IEEE Transactions on Neural Networks and Learning Systems, 33(12):7039–7051, 2021.
  • [9] R. Devon Hjelm, Alex Fedorov, Samuel Lavoie-Marchildon, Karan Grewal, Philip Bachman, Adam Trischler, and Yoshua Bengio. Learning deep representations by mutual information estimation and maximization. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net, 2019.
  • [10] Wentai Hou, Lequan Yu, Chengxuan Lin, Helong Huang, Rongshan Yu, Jing Qin, and Liansheng Wang. H^ 2-mil: exploring hierarchical representation with heterogeneous multiple instance learning for whole slide image analysis. In Proceedings of the AAAI conference on artificial intelligence, volume 36, pages 933–941, 2022.
  • [11] Frederick M Howard, James Dolezal, Sara Kochanny, Jefree Schulte, Heather Chen, Lara Heij, Dezheng Huo, Rita Nanda, Olufunmilayo I Olopade, Jakob N Kather, et al. The impact of site-specific digital histology signatures on deep learning model accuracy and bias. Nature communications, 12(1):4423, 2021.
  • [12] Gao Huang, Zhuang Liu, Laurens Van Der Maaten, and Kilian Q Weinberger. Densely connected convolutional networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4700–4708, 2017.
  • [13] Yanyan Huang, Weiqin Zhao, Shujun Wang, Yu Fu, Yuming Jiang, and Lequan Yu. Conslide: Asynchronous hierarchical interaction transformer with breakup-reorganize rehearsal for continual whole slide image analysis. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 21349–21360, 2023.
  • [14] Zhi Huang, Federico Bianchi, Mert Yuksekgonul, Thomas J Montine, and James Zou. A visual–language foundation model for pathology image analysis using medical twitter. Nature medicine, 29(9):2307–2316, 2023.
  • [15] Wisdom Ikezogwo, Saygin Seyfioglu, Fatemeh Ghezloo, Dylan Geva, Fatwir Sheikh Mohammed, Pavan Kumar Anand, Ranjay Krishna, and Linda Shapiro. Quilt-1m: One million image-text pairs for histopathology. Advances in Neural Information Processing Systems, 36, 2024.
  • [16] Maximilian Ilse, Jakub Tomczak, and Max Welling. Attention-based deep multiple instance learning. In International conference on machine learning, pages 2127–2136. PMLR, 2018.
  • [17] Syed Ashar Javed, Dinkar Juyal, Harshith Padigela, Amaro Taylor-Weiner, Limin Yu, and Aaditya Prakash. Additive mil: Intrinsically interpretable multiple instance learning for pathology. Advances in Neural Information Processing Systems, 35:20689–20702, 2022.
  • [18] Diederik P Kingma and Max Welling. Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114, 2013.
  • [19] Günter Klambauer, Thomas Unterthiner, Andreas Mayr, and Sepp Hochreiter. Self-normalizing neural networks. Advances in neural information processing systems, 30, 2017.
  • [20] Zhengfeng Lai, Zhuoheng Li, Luca Cerny Oliveira, Joohi Chauhan, Brittany N Dugger, and Chen-Nee Chuah. Clipath: Fine-tune clip with visual feature fusion for pathology image analysis towards minimizing data collection efforts. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 2374–2380, 2023.
  • [21] Bin Li, Yin Li, and Kevin W Eliceiri. Dual-stream multiple instance learning network for whole slide image classification with self-supervised contrastive learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 14318–14328, 2021.
  • [22] Hao Li, Ying Chen, Yifei Chen, Rongshan Yu, Wenxian Yang, Liansheng Wang, Bowen Ding, and Yuchen Han. Generalizable whole slide image classification with fine-grained visual-semantic interaction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 11398–11407, 2024.
  • [23] Jana Lipkova, Richard J Chen, Bowen Chen, Ming Y Lu, Matteo Barbieri, Daniel Shao, Anurag J Vaidya, Chengkuan Chen, Luoting Zhuang, Drew FK Williamson, et al. Artificial intelligence for multimodal data integration in oncology. Cancer cell, 40(10):1095–1110, 2022.
  • [24] Jiaxuan Lu, Fang Yan, Xiaofan Zhang, Yue Gao, and Shaoting Zhang. Pathotune: Adapting visual foundation model to pathological specialists. arXiv preprint arXiv:2403.16497, 2024.
  • [25] Ming Y Lu, Bowen Chen, Drew FK Williamson, Richard J Chen, Ivy Liang, Tong Ding, Guillaume Jaume, Igor Odintsov, Long Phi Le, Georg Gerber, et al. A visual-language foundation model for computational pathology. Nature Medicine, pages 1–12, 2024.
  • [26] Ming Y Lu, Drew FK Williamson, Tiffany Y Chen, Richard J Chen, Matteo Barbieri, and Faisal Mahmood. Data-efficient and weakly supervised computational pathology on whole-slide images. Nature biomedical engineering, 5(6):555–570, 2021.
  • [27] Aaron van den Oord, Yazhe Li, and Oriol Vinyals. Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748, 2018.
  • [28] Maxime Oquab, Timothée Darcet, Théo Moutakanni, Huy Vo, Marc Szafraniec, Vasil Khalidov, Pierre Fernandez, Daniel Haziza, Francisco Massa, Alaaeldin El-Nouby, et al. Dinov2: Learning robust visual features without supervision. arXiv preprint arXiv:2304.07193, 2023.
  • [29] Linhao Qu, Kexue Fu, Manning Wang, Zhijian Song, et al. The rise of ai language pathologists: Exploring two-level prompt learning for few-shot weakly-supervised whole slide image classification. Advances in Neural Information Processing Systems, 36, 2024.
  • [30] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International conference on machine learning, pages 8748–8763. PMLR, 2021.
  • [31] Abtin Riasatian, Morteza Babaie, Danial Maleki, Shivam Kalra, Mojtaba Valipour, Sobhan Hemati, Manit Zaveri, Amir Safarpoor, Sobhan Shafiei, Mehdi Afshari, et al. Fine-tuning and training of densenet for histopathology image representation using tcga diagnostic slides. Medical image analysis, 70:102032, 2021.
  • [32] Alexander Robey, George J Pappas, and Hamed Hassani. Model-based domain generalization. Advances in Neural Information Processing Systems, 34:20210–20229, 2021.
  • [33] Zhuchen Shao, Hao Bian, Yang Chen, Yifeng Wang, Jian Zhang, Xiangyang Ji, et al. Transmil: Transformer based correlated multiple instance learning for whole slide image classification. Advances in Neural Information Processing Systems, 34:2136–2147, 2021.
  • [34] Artem Shmatko, Narmin Ghaffari Laleh, Moritz Gerstung, and Jakob Nikolas Kather. Artificial intelligence in histopathology: enhancing cancer research and clinical oncology. Nature Cancer, 3(9):1026–1038, 2022.
  • [35] Wenhao Tang, Fengtao Zhou, Sheng Huang, Xiang Zhu, Yi Zhang, and Bo Liu. Feature re-embedding: Towards foundation model-level performance in computational pathology. arXiv preprint arXiv:2402.17228, 2024.
  • [36] Eugene Vorontsov, Alican Bozkurt, Adam Casson, George Shaikovski, Michal Zelechowski, Siqi Liu, Philippe Mathieu, Alexander van Eck, Donghun Lee, Julian Viret, et al. Virchow: A million-slide digital pathology foundation model. arXiv preprint arXiv:2309.07778, 2023.
  • [37] Xiyue Wang, Sen Yang, Jun Zhang, Minghui Wang, Jing Zhang, Wei Yang, Junzhou Huang, and Xiao Han. Transformer-based unsupervised contrastive learning for histopathological image classification. Medical image analysis, 81:102559, 2022.
  • [38] Guile Wu and Shaogang Gong. Collaborative optimization and aggregation for decentralized domain generalization and adaptation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6484–6493, 2021.
  • [39] Cui Yufei, Ziquan Liu, Xiangyu Liu, Xue Liu, Cong Wang, Tei-Wei Kuo, Chun Jason Xue, and Antoni B Chan. Bayes-mil: A new probabilistic perspective on attention-based multiple instance learning for whole slide images. In The Eleventh International Conference on Learning Representations, 2022.
  • [40] Hongrun Zhang, Yanda Meng, Yitian Zhao, Yihong Qiao, Xiaoyun Yang, Sarah E Coupland, and Yalin Zheng. Dtfd-mil: Double-tier feature distillation multiple instance learning for histopathology whole slide image classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 18802–18812, 2022.
  • [41] Yunkun Zhang, Jin Gao, Mu Zhou, Xiaosong Wang, Yu Qiao, Shaoting Zhang, and Dequan Wang. Text-guided foundation model adaptation for pathological image classification. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 272–282. Springer, 2023.
  • [42] Jinghao Zhou, Chen Wei, Huiyu Wang, Wei Shen, Cihang Xie, Alan Yuille, and Tao Kong. ibot: Image bert pre-training with online tokenizer. arXiv preprint arXiv:2111.07832, 2021.

Appendix A Overview

The structure of this supplementary material as shown below,

  • Appendix B presents additional experimental results, including changes in in-domain and out-of-domain site splitting, further comparisons on the BRCA dataset under traditional settings, additional results on the PRAD dataset, and results under site-preserved cross-validation.

  • Appendix C discusses additional ablation study results concerning hyperparameters, including the impact of loss weights for PIM and SIM, the effect of the number of representative patches kk, and further ablation studies on the CIB module.

  • Appendix D presents detailed descriptions of datasets and experimental settings.

  • Appendix E provides the detailed definition and formula derivation.

  • Appendix F elaborates on the Concept Information Bottleneck (CIB) module using the information plane.

  • Appendix G details the prompts used in our experiments.

  • Appendix H presents additional visualization results.

Appendix B Additional Results

B.1 Additional Cancer Subtyping Results with Different IND and OOD Sites

To further evaluate the generalization ability of the proposed CATE method, we conduct additional experiments on TCGA-BRCA dataset with different settings of in-domain (IND) and out-of-domain (OOD) sites. The detailed experimental settings are shown below:

  • BRCA (NINDN_{\text{IND}}=1): One site as the in-domain data, and the other sites as the out-of-domain data. The details of the in-domain site are shown below:

    • AR: 44 slides of IDC and 15 slides of ILC.

  • BRCA (NINDN_{\text{IND}}=2): Two sites as the in-domain data, and the other sites as the out-of-domain data. The details of the in-domain sites are shown below:

    • AR: 44 slides of IDC and 15 slides of ILC.

    • B6: 39 slides of IDC and 6 slides of ILC.

The additional results are presented in Table 4. It is evident that the proposed CATE-MIL consistently delivers superior performance in both in-domain and out-of-domain settings, demonstrating its effectiveness in enhancing task-specific information and minimizing the impact of irrelevant data.

B.2 Additional Cancer Subtyping Results with Traditional Experimental Settings

Table 4: Supplementary Cancer Subtyping Results on BRCA.
Method BRCA (NINDN_{\text{IND}}=1)
OOD-AUC OOD-ACC IND-AUC IND-ACC
ABMIL 0.916 ±\pm 0.017 0.882 ±\pm 0.022 0.977 ±\pm 0.029 0.953 ±\pm 0.036
CLAM 0.895 ±\pm 0.031 0.842 ±\pm 0.049 0.973 ±\pm 0.035 0.938 ±\pm 0.043
DSMIL 0.892 ±\pm 0.017 0.863 ±\pm 0.016 0.967 ±\pm 0.022 0.924 ±\pm 0.048
TransMIL 0.917 ±\pm 0.008 0.878 ±\pm 0.016 0.980 ±\pm 0.031 0.955 ±\pm 0.038
DTFD-MIL 0.920 ±\pm 0.023 0.881 ±\pm 0.024 0.982 ±\pm 0.030 0.957 ±\pm 0.038
R2T-MIL 0.902 ±\pm 0.022 0.835 ±\pm 0.063 0.963 ±\pm 0.037 0.944 ±\pm 0.039
CATE-MIL 0.938 ±\pm 0.014 0.895 ±\pm 0.021 0.984 ±\pm 0.021 0.959 ±\pm 0.038
Method BRCA (NINDN_{\text{IND}}=2)
OOD-AUC OOD-ACC IND-AUC IND-ACC
ABMIL 0.914 ±\pm 0.012 0.879 ±\pm 0.017 0.954 ±\pm 0.035 0.910 ±\pm 0.043
CLAM 0.914 ±\pm 0.021 0.890 ±\pm 0.013 0.958 ±\pm 0.034 0.926 ±\pm 0.050
DSMIL 0.931 ±\pm 0.007 0.900 ±\pm 0.012 0.971 ±\pm 0.020 0.919 ±\pm 0.034
TransMIL 0.929 ±\pm 0.009 0.897 ±\pm 0.006 0.962 ±\pm 0.030 0.903 ±\pm 0.039
DTFD-MIL 0.898 ±\pm 0.009 0.868 ±\pm 0.017 0.950 ±\pm 0.034 0.890 ±\pm 0.037
R2T-MIL 0.919 ±\pm 0.017 0.893 ±\pm 0.017 0.962 ±\pm 0.025 0.906 ±\pm 0.026
CATE-MIL 0.947 ±\pm 0.005 0.920 ±\pm 0.004 0.980 ±\pm 0.015 0.939 ±\pm 0.022
  • The best results are highlighted in bold, and the second-best results are underlined.

  • denotes the methods for feature re-embedding that utilize ABMIL as base MIL model.

Table 5: Results on BRCA under Traditional Settings.
Method AUC ACC Training Time Params Size
ABMIL 0.922±\pm0.046 0.902±\pm0.039 67.16 S 1.26 MB
CLAM 0.928±\pm0.030 0.912±\pm0.034 67.99 S 2.01 MB
DSMIL 0.934±\pm0.039 0.910±\pm0.033 72.91 S 1.26 MB
TransMIL 0.927±\pm0.046 0.906±\pm0.032 72.39 S 5.39 MB
DTFD-MIL 0.940±\pm0.030 0.901±\pm0.041 70.99 S 8.03 MB
R2T-MIL 0.936±\pm0.027 0.903±\pm0.028 68.87 S 9.28 MB (1.26+8.02)
CATE-MIL 0.945±\pm0.033 0.917±\pm0.029 68.53 S 4.26 MB (1.26+3.00)
  • The best results are highlighted in bold, and the second-best results are underlined.

  • denotes the methods for feature re-embedding that utilize ABMIL as base MIL model.

Table 6: Supplementary Gleason Grading Results on PRAD.
Method PRAD
OOD-AUC OOD-ACC IND-AUC IND-ACC
ABMIL 0.704 ±\pm 0.034 0.510 ±\pm 0.075 0.742 ±\pm 0.060 0.575 ±\pm 0.051
CATE-MIL 0.755 ±\pm 0.050 0.567 ±\pm 0.067 0.797 ±\pm 0.044 0.643 ±\pm 0.075
Table 7: Supplementary Results under Site-Preserved Cross-Validation.
Dataset Method OOD-AUC IND-AUC
BRCA ABMIL 0.912 ±\pm 0.012 0.905 ±\pm 0.043
CATE-MIL 0.935 ±\pm 0.014 0.942 ±\pm 0.038
NSCLC ABMIL 0.942 ±\pm 0.016 0.941 ±\pm 0.013
CATE-MIL 0.951 ±\pm 0.015 0.943 ±\pm 0.008
RCC ABMIL 0.980 ±\pm 0.001 0.986 ±\pm 0.010
CATE-MIL 0.983 ±\pm 0.001 0.989 ±\pm 0.007

In the main paper, we divided the dataset into in-domain and out-of-domain sites to evaluate the generalization ability of the proposed CATE method. To further evaluate its effectiveness, we conducted additional experiments under traditional settings, randomly splitting the dataset into training, validation, and testing sets. We report the performance metrics, including means and standard deviations over 10 Monte-Carlo Cross-Validation runs, in Table 5. Additionally, we provide comparisons of training times and parameter sizes for various methods.

It is evident that the proposed CATE-MIL consistently outperforms others in both AUC and ACC metrics, underscoring its superiority. Furthermore, CATE-MIL benefits from shorter training times and smaller parameter sizes compared to the R2T-MIL method.

B.3 Additional Gleason Grading Results on PRAD

CATE is a general framework that can be applied to more complex tasks beyond cancer subtyping, such as Gleason grading in prostate cancer. We have conducted conducted experiments on the TCGA-PRAD dataset to evaluate the performance of CATE-MIL in Gleason grading. Specifically, the samples in PRAD dataset are divided into Gleason pattern 3, 4, and 5, and the task is to classify the samples into these three categories. The results are shown in Table 6. It is evident that CATE-MIL consistently outperforms the base model ABMIL in both in-domain and out-of-domain settings, demonstrating its effectiveness in enhancing task-specific information and minimizing the impact of irrelevant data. In the future, as more studies reveal the connection between morphological features and molecular biomarkers and more powerful pathology VLMs are developed, our framework has the potential to benefit more complex tasks.

B.4 Additional Experiments using Site-Preserved Cross-Validation

To provide a more comprehensive evaluation of the proposed CATE-MIL, we conduct additional experiments using site-preserved cross-validation [11], where the samples from the same site are preserved in the same fold. For each fold, we split the data into training and testing sets, and these testing sets are regarded as in-domain testing data. And the other sites are used as out-of-domain testing data. The results are shown in Table 7. It is evident that CATE-MIL consistently outperforms ABMIL in both in-domain and out-of-domain settings.

Appendix C Additional Ablation Study

C.1 Ablation Study of the Weight of PIM and SIM Losses

The overall objective of the proposed CATE-MIL is a weighted sum of the PIM and SIM losses, along with the classification loss:

=CE+λPPIM+λSSIM.\mathcal{L}=\mathcal{L}_{CE}+\lambda_{P}\mathcal{L}_{PIM}+\lambda_{S}\mathcal{L}_{SIM}. (12)

We note that the hyperparameters for the weights of the PIM and SIM losses were not specifically tuned in the main paper. To investigate the effect of the weight of PIM and SIM losses on the model performance, we conduct an ablation study on BRCA (NINDN_{\text{IND}}=2) with CATE-MIL (ABMIL integrated with CATE), varying the weights of PIM and SIM losses. The results are shown in Figure 4.

Refer to caption
Figure 4: Ablation study of the weight of PIM and SIM losses on the model performance.

From Figure 4, it is evident that CATE enhances the performance of the base model ABMIL in most scenarios, particularly in out-of-domain settings. When the weight of PIM loss λP\lambda_{P} is too low, model performance suffers, underscoring the significant role of PIM loss in enhancing task-specific information. Conversely, an excessively high λP\lambda_{P} also diminishes performance, as the model overly prioritizes maximizing mutual information between the original and calibrated features at the expense of optimizing classification loss. Regarding the weight of SIM loss λS\lambda_{S}, optimal performance is achieved when λS\lambda_{S} is approximately 30. If the λS\lambda_{S} is too low, the model fails to effectively eliminate irrelevant information, thereby impairing performance. Conversely, if λS\lambda_{S} is too high, the model risks discarding task-relevant information, leading to performance degradation.

Refer to caption
Figure 5: Ablation study of kk.

C.2 Ablation Study of the Number of Representative Patches kk

In this section, we conduct an ablation study to investigate the impact of the number of representative patches, kk, on model performance, as discussed in Section 3.2. In practice, the representative patches are selected based on the similarities between the original image feature and the corresponding class-specific concept anchor.

As shown in Figure 5, it is evident that CATE generally enhances the performance of the base model, ABMIL. When the number of representative patches kk is too small, model performance degrades due to insufficient capture of task-specific information. Conversely, an excessively large kk also leads to performance degradation in out-of-domain scenarios, as it introduces noise from irrelevant information. Consequently, we have set the number of representative patches kk to 10 in the main paper.

C.3 Additional Ablation Study of CIB Module

We further conducted experiments on CATE-MIL without concept alignment (discarding PIM loss and SIM loss of the CIB module) and replaced the CIB module with an MLP to investigate the effect of concept alignment and the increased number of parameters. The results are shown in Table 8. The performance of CATE-MIL significantly decreases in both cases, demonstrating the importance of concept alignment in the CIB module and that the improvements of CATE are not due to the increased number of parameters.

Table 8: Supplementary Ablation Study of CIB Module.
Method BRCA (NINDN_{\text{IND}}=1) BRCA (NINDN_{\text{IND}}=2) NSCLC (NINDN_{\text{IND}}=2) NSCLC (NINDN_{\text{IND}}=4) RCC (NINDN_{\text{IND}}=3) RCC (NINDN_{\text{IND}}=6)
CATE-MIL w/o CFI (Baseline) 0.936±\pm0.010 0.942±\pm0.010 0.910±\pm0.030 0.960±\pm0.011 0.979±\pm0.004 0.977±\pm0.005
- w/o concept alignment 0.900±\pm0.017 0.884±\pm0.033 0.742±\pm0.059 0.897±\pm0.022 0.961±\pm0.011 0.932±\pm0.016
- Replace CIB with MLP 0.888±\pm0.027 0.902±\pm0.037 0.816±\pm0.040 0.931±\pm0.023 0.966±\pm0.006 0.951±\pm0.021

Appendix D Datasets Description and Detailed Experimental Settings

In this section, we provide detailed descriptions of the datasets used in the experiments and the detailed experimental settings that we used in the experiments in Section 4.2.

D.1 Datasets Description

  • TCGA-BRCA: The dataset contains nine disease subtypes, and this study focuses on the classification of Invasive ductal carcinoma (IDC, 726 slides from 694 cases) and invasive lobular carcinoma (ILC, 149 slides from 143 cases). The dataset is collected from 36 sites with 20 of them having both IDC and ILC slides, and the other 16 sites only have IDC slides or ILC slides.

  • TCGA-NSCLC: The dataset contains two disease subtypes, including lung adenocarcinoma (LUAD, 492 slides from 430 cases) and lung squamous cell carcinoma (LUSC, 466 slides from 432 cases). The dataset is collected from 66 sites. Different from TCGA-BRCA, each site only contains one disease subtype.

  • TCGA-RCC: The dataset contains three disease subtypes, including clear cell renal cell carcinoma (CCRCC, 498 slides from 492 cases), papillary renal cell carcinoma (PRCC, 289 slides from 267 cases), and chromophobe renal cell carcinoma (CHRCC, 118 slides from 107 cases). The dataset is collected from 55 sites. Similar to TCGA-NSCLC, each site only contains one disease subtype.

D.2 Detailed Experimental Settings

To validate that the proposed CATE effectively improves MIL model performance by enhancing the task-specific information and eliminating the disturbance of task-irrelevant information, we conduct extensive experiments across three datasets under various in-domain and out-of-domain settings.

Specifically, for TCGA-BRCA dataset, we conduct the following experiments:

  • BRCA (NINDN_{\text{IND}}=1): One site as the in-domain data, and the other sites as the out-of-domain data. The details of the in-domain site are shown below:

    • D8: 59 slides of IDC and 8 slides of ILC.

  • BRCA (NINDN_{\text{IND}}=2): Two sites as the in-domain data, and the other sites as the out-of-domain data. The details of the in-domain sites are shown below:

    • A8: 48 slides of IDC and 5 slides of ILC.

    • D8: 59 slides of IDC and 8 slides of ILC.

For TCGA-NSCLC dataset, we conduct the following experiments:

  • NSCLC (NINDN_{\text{IND}}=2): Two sites as the in-domain data, and the other sites as the out-of-domain data. The details of the in-domain sites are shown below:

    • 44: 43 slides of LUAD and 0 slides of LUSC.

    • 22: 0 slides of LUAD and 36 slides of LUSC.

  • NSCLC (NINDN_{\text{IND}}=4): Four sites as the in-domain data, and the other sites as the out-of-domain data. The details of the in-domain sites are shown below:

    • 44: 43 slides of LUAD and 0 slides of LUSC.

    • 50: 20 slides of LUAD and 0 slides of LUSC.

    • 22: 0 slides of LUAD and 36 slides of LUSC.

    • 56: 0 slides of LUAD and 35 slides of LUSC.

For TCGA-RCC dataset, we conduct the following experiments:

  • RCC (NINDN_{\text{IND}}=3): Three sites as the in-domain data, and the other sites as the out-of-domain data. The details of the in-domain sites are shown below:

    • A3: 48 slides of CCRCC, 0 slides of CHRCC, and 0 slides of PRCC.

    • KL: 0 slides of CCRCC, 24 slides of CHRCC, and 0 slides of PRCC.

    • 2Z: 0 slides of CCRCC, 0 slides of CHRCC, and 23 slides of PRCC.

  • RCC (NINDN_{\text{IND}}=6): Six sites as the in-domain data, and the other sites as the out-of-domain data. The details of the in-domain sites are shown below:

    • A3: 48 slides of CCRCC, 0 slides of CHRCC, and 0 slides of PRCC.

    • AK: 15 slides of CCRCC, 0 slides of CHRCC, and 0 slides of PRCC.

    • KL: 0 slides of CCRCC, 24 slides of CHRCC, and 0 slides of PRCC.

    • KM: 0 slides of CCRCC, 20 slides of CHRCC, and 0 slides of PRCC.

    • 2Z: 0 slides of CCRCC, 0 slides of CHRCC, and 23 slides of PRCC.

    • 5P: 0 slides of CCRCC, 0 slides of CHRCC, and 15 slides of PRCC.

Appendix E Detailed Definition and Formula Derivation

E.1 Definition of Sufficiency and Consistency

In this part, we will define the sufficiency and consistency in the context of the concept information bottleneck (CIB) module.

First, we define the mutual information between two random variables XX and YY as:

I(X;Y)=p(x,y)logp(x,y)p(x)p(y).I(X;Y)=\sum p(x,y)\log\frac{p(x,y)}{p(x)p(y)}. (13)

Sufficiency: To quantify the requirement that the calibrated features should be maximally informative about the label information 𝐲\boldsymbol{\rm{y}}, we define the sufficiency as the mutual information between the label information and the calibrated features:

Sufficiency. 𝜶^\hat{\boldsymbol{\alpha}} is sufficient for 𝐲I(𝒙^;𝐲|𝜶^)=0I(𝒙^;𝐲)=I(𝜶^;𝐲)\boldsymbol{\rm{y}}\iff I(\hat{\boldsymbol{x}};\boldsymbol{\rm{y}}|\hat{\boldsymbol{\alpha}})=0\iff I(\hat{\boldsymbol{x}};\boldsymbol{\rm{y}})=I(\hat{\boldsymbol{\alpha}};\boldsymbol{\rm{y}}).

The sufficiency definition requires that the calibrated features 𝜶^\hat{\boldsymbol{\alpha}} encapsulate all information about the label information 𝐲\boldsymbol{\rm{y}} that is accessible from the original features 𝒙^\hat{\boldsymbol{x}}. In essence, the calibrated feature 𝜶^\hat{\boldsymbol{\alpha}} of original feature 𝒙^\hat{\boldsymbol{x}} is sufficient for determining label 𝐲\boldsymbol{\rm{y}} if and only if the amount of information regarding the specific task is unchanged after the transformation.

Consistency: Since the image feature 𝒙^\hat{\boldsymbol{x}} and the concept anchor 𝐜\boldsymbol{\rm{c}} are consistent in the pathology VLM, we posit that any representation containing all information accessible from both the original feature and the concept anchor also encompasses the necessary discriminative label information. Thus, we define the consistency between the concept anchor and the original feature as:

Consistency. 𝐜\boldsymbol{\rm{c}} is consistent with 𝒙^\hat{\boldsymbol{x}} for 𝐲I(𝐲;𝒙^|𝐜)=0\boldsymbol{\rm{y}}\iff I(\boldsymbol{\rm{y}};\hat{\boldsymbol{x}}|\boldsymbol{\rm{c}})=0.

The consistency definition requires that the concept anchor 𝐜\boldsymbol{\rm{c}} and the original feature 𝒙^\hat{\boldsymbol{x}} should be consistent with the label information 𝐲\boldsymbol{\rm{y}}.

E.2 Maximize Predictive Information with InfoNCE

In this part, we will prove that the predictive information for the concept anchor can be maximized by maximizing the InfoNCE mutual information lower bound, which is defined in Equation 4. As in [27], the optimal value for f(𝒄,𝜶^i)f\left(\boldsymbol{c},\hat{\boldsymbol{\alpha}}_{i}\right) should be proportional to the density ratio:

f(𝒄,𝜶^i)p(𝒄,𝜶^i)p(𝒄)p(𝜶^i).f\left(\boldsymbol{c},\hat{\boldsymbol{\alpha}}_{i}\right)\propto\frac{p\left(\boldsymbol{c},\hat{\boldsymbol{\alpha}}_{i}\right)}{p\left(\boldsymbol{c}\right)p\left(\hat{\boldsymbol{\alpha}}_{i}\right)}. (14)

And we can get:

INCEopt(𝜶^;𝐜)=𝔼𝜶^,𝒄poscs[i=1klogp(𝒄poscs,𝜶^i)p(𝒄poscs)p(𝜶^i)j=1mp(𝒄jcs,𝜶^i)p(𝒄jcs)p(𝜶^i)+j=1np(𝒄jca,𝜶^i)p(𝒄jca)p(𝜶^i)]=𝔼𝜶^,𝒄poscs[i=1klog(1+p(𝒄poscs)p(𝜶^i)p(𝒄poscs,𝜶^i)(jposp(𝒄jcs,𝜶^i)p(𝒄jcs)p(𝜶^i)+j=1np(𝒄jca,𝜶^i)p(𝒄jca)p(𝜶^i)))]𝔼𝜶^,𝒄poscs[i=1klog(1+p(𝒄poscs)p(𝜶^i)p(𝒄poscs,𝜶^i)((m1)𝔼𝜶^ip(𝒄jcs,𝜶^i)p(𝒄jcs)p(𝜶^i)+n𝔼𝜶^ip(𝒄jca,𝜶^i)p(𝒄jca)p(𝜶^i)))]=𝔼𝜶^,𝒄poscs[i=1klog(1+p(𝒄poscs)p(𝜶^i)p(𝒄poscs,𝜶^i)(m+n1))]𝔼𝜶^,𝒄poscs[i=1klog(p(𝒄poscs)p(𝜶^i)p(𝒄poscs,𝜶^i)(m+n))]=i=1kI(𝒄poscs;𝜶^i)klog(m+n)=I(𝒄poscs;𝜶^)klog(m+n),\begin{split}&I_{NCE}^{opt}\left(\hat{\boldsymbol{\rm{\alpha}}};\boldsymbol{\rm{c}}\right)=\operatornamewithlimits{\mathbb{E}}_{\hat{\boldsymbol{\rm{\alpha}}},\boldsymbol{c}^{\text{cs}}_{\text{pos}}}\left[\sum_{i=1}^{k}\log\frac{\frac{p\left(\boldsymbol{c}^{\text{cs}}_{\text{pos}},\hat{\boldsymbol{\alpha}}_{i}\right)}{p\left(\boldsymbol{c}^{\text{cs}}_{\text{pos}}\right)p\left(\hat{\boldsymbol{\alpha}}_{i}\right)}}{\sum_{j=1}^{m}\frac{p\left(\boldsymbol{c}^{\text{cs}}_{j},\hat{\boldsymbol{\alpha}}_{i}\right)}{p\left(\boldsymbol{c}^{\text{cs}}_{j}\right)p\left(\hat{\boldsymbol{\alpha}}_{i}\right)}+\sum_{j=1}^{n}\frac{p\left(\boldsymbol{c}^{\text{ca}}_{j},\hat{\boldsymbol{\alpha}}_{i}\right)}{p\left(\boldsymbol{c}^{\text{ca}}_{j}\right)p\left(\hat{\boldsymbol{\alpha}}_{i}\right)}}\right]\\ &=-\operatornamewithlimits{\mathbb{E}}_{\hat{\boldsymbol{\rm{\alpha}}},\boldsymbol{c}^{\text{cs}}_{\text{pos}}}\left[\sum_{i=1}^{k}\log\left(1+\frac{p\left(\boldsymbol{c}^{\text{cs}}_{\text{pos}}\right)p\left(\hat{\boldsymbol{\alpha}}_{i}\right)}{p\left(\boldsymbol{c}^{\text{cs}}_{\text{pos}},\hat{\boldsymbol{\alpha}}_{i}\right)}\left(\sum_{j\neq\text{pos}}\frac{p\left(\boldsymbol{c}_{j}^{\text{cs}},\hat{\boldsymbol{\alpha}}_{i}\right)}{p\left(\boldsymbol{c}_{j}^{\text{cs}}\right)p\left(\hat{\boldsymbol{\alpha}}_{i}\right)}+\sum_{j=1}^{n}\frac{p\left(\boldsymbol{c}_{j}^{\text{ca}},\hat{\boldsymbol{\alpha}}_{i}\right)}{p\left(\boldsymbol{c}_{j}^{\text{ca}}\right)p\left(\hat{\boldsymbol{\alpha}}_{i}\right)}\right)\right)\right]\\ &\approx-\operatornamewithlimits{\mathbb{E}}_{\hat{\boldsymbol{\rm{\alpha}}},\boldsymbol{c}^{\text{cs}}_{\text{pos}}}\left[\sum_{i=1}^{k}\log\left(1+\frac{p\left(\boldsymbol{c}^{\text{cs}}_{\text{pos}}\right)p\left(\hat{\boldsymbol{\alpha}}_{i}\right)}{p\left(\boldsymbol{c}^{\text{cs}}_{\text{pos}},\hat{\boldsymbol{\alpha}}_{i}\right)}\left(\left(m-1\right)\operatornamewithlimits{\mathbb{E}}_{\hat{\boldsymbol{\alpha}}_{i}}\frac{p\left(\boldsymbol{c}_{j}^{\text{cs}},\hat{\boldsymbol{\alpha}}_{i}\right)}{p\left(\boldsymbol{c}_{j}^{\text{cs}}\right)p\left(\hat{\boldsymbol{\alpha}}_{i}\right)}+n\operatornamewithlimits{\mathbb{E}}_{\hat{\boldsymbol{\alpha}}_{i}}\frac{p\left(\boldsymbol{c}_{j}^{\text{ca}},\hat{\boldsymbol{\alpha}}_{i}\right)}{p\left(\boldsymbol{c}_{j}^{\text{ca}}\right)p\left(\hat{\boldsymbol{\alpha}}_{i}\right)}\right)\right)\right]\\ &=-\operatornamewithlimits{\mathbb{E}}_{\hat{\boldsymbol{\rm{\alpha}}},\boldsymbol{c}^{\text{cs}}_{\text{pos}}}\left[\sum_{i=1}^{k}\log\left(1+\frac{p\left(\boldsymbol{c}^{\text{cs}}_{\text{pos}}\right)p\left(\hat{\boldsymbol{\alpha}}_{i}\right)}{p\left(\boldsymbol{c}^{\text{cs}}_{\text{pos}},\hat{\boldsymbol{\alpha}}_{i}\right)}\left(m+n-1\right)\right)\right]\\ &\leq-\operatornamewithlimits{\mathbb{E}}_{\hat{\boldsymbol{\rm{\alpha}}},\boldsymbol{c}^{\text{cs}}_{\text{pos}}}\left[\sum_{i=1}^{k}\log\left(\frac{p\left(\boldsymbol{c}^{\text{cs}}_{\text{pos}}\right)p\left(\hat{\boldsymbol{\alpha}}_{i}\right)}{p\left(\boldsymbol{c}^{\text{cs}}_{\text{pos}},\hat{\boldsymbol{\alpha}}_{i}\right)}\left(m+n\right)\right)\right]\\ &=\sum_{i=1}^{k}I\left(\boldsymbol{c}^{\text{cs}}_{\text{pos}};\hat{\boldsymbol{\alpha}}_{i}\right)-k\log\left(m+n\right)\\ &=I\left(\boldsymbol{c}^{\text{cs}}_{\text{pos}};\hat{\boldsymbol{\rm{\alpha}}}\right)-k\log\left(m+n\right),\end{split} (15)

where 𝑪neg\boldsymbol{C}_{\text{neg}} denotes the negative concepts, including both negative class concepts and other type concepts. We can see that the InfoNCE estimation is a lower bound of the mutual information between the concept anchor and the calibrated features:

I(𝒄poscs;𝜶^)INCEopt(𝜶^;𝐜)+klog(m+n)INCE(𝜶^;𝐜).I\left(\boldsymbol{c}^{\text{cs}}_{\text{pos}};\hat{\boldsymbol{\rm{\alpha}}}\right)\geq I_{NCE}^{opt}\left(\hat{\boldsymbol{\rm{\alpha}}};\boldsymbol{\rm{c}}\right)+k\log\left(m+n\right)\geq I_{NCE}\left(\hat{\boldsymbol{\rm{\alpha}}};\boldsymbol{\rm{c}}\right). (16)

Thus, the predictive information of calibrated features for the concept anchor can be maximized by maximizing the InfoNCE mutual information lower bound.

E.3 Superfluous Information Minimization

As shown in Equation 3, the mutual information between the original features and the calibrated features I(𝒙^;𝜶^)I(\hat{\boldsymbol{x}};\hat{\boldsymbol{\alpha}}) can be decomposed into the mutual information I(𝜶^;𝐜)I(\hat{\boldsymbol{\alpha}};\boldsymbol{\rm{c}}) between the calibrated features and the concept anchor and the mutual information I(𝒙^;𝜶^|𝐜)I(\hat{\boldsymbol{x}};\hat{\boldsymbol{\alpha}}|\boldsymbol{\rm{c}}) between the concept anchor and the calibrated features:

I(𝒙^;𝜶^)=I(𝜶^;𝐜)Predictive information for 𝐜+I(𝒙^;𝜶^|𝐜)Superfluous information.I(\hat{\boldsymbol{x}};\hat{\boldsymbol{\alpha}})=\underbrace{I(\hat{\boldsymbol{\alpha}};\boldsymbol{\rm{c}})}_{\text{Predictive information for }\boldsymbol{\rm{c}}}+\underbrace{I(\hat{\boldsymbol{x}};\hat{\boldsymbol{\alpha}}|\boldsymbol{\rm{c}})}_{\text{Superfluous information}}. (17)

As introduced in section 3.2, the first item is the predictive information for the concept anchor, which can be maximized by maximizing the InfoNCE mutual information lower bound. Thus, the second superfluous information item can be minimized by minimizing I(𝒙^;𝜶^)I(\hat{\boldsymbol{x}};\hat{\boldsymbol{\alpha}}). In practice, the superfluous information minimization is conducted on all patches in the subset 𝐱\boldsymbol{\rm{x}}, since the task-irrelevant information is distributed across all patches. Thus, the superfluous information can be minimized by minimizing I(𝒙;𝜶)I({\boldsymbol{x}};{\boldsymbol{\alpha}}). As in [1], this mutual information can be represented as:

I(𝒙;𝜶)=p(𝒙,𝜶)logp(𝜶|𝒙)p(𝜶)d𝒙d𝜶=p(𝒙,𝜶)logp(𝜶|𝒙)𝑑𝒙𝑑𝜶p(𝜶)logp(𝜶)𝑑𝜶.\begin{split}I({\boldsymbol{x}};{\boldsymbol{\alpha}})&=\int p\left({\boldsymbol{x}},{\boldsymbol{\alpha}}\right)\log\frac{p\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right)}{p\left({\boldsymbol{\alpha}}\right)}d{\boldsymbol{x}}d{\boldsymbol{\alpha}}\\ &=\int p\left({\boldsymbol{x}},{\boldsymbol{\alpha}}\right)\log p\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right)d{\boldsymbol{x}}d{\boldsymbol{\alpha}}-\int p\left({\boldsymbol{\alpha}}\right)\log p\left({\boldsymbol{\alpha}}\right)d{\boldsymbol{\alpha}}.\end{split} (18)

However, the computing of the marginal distribution p(𝜶)=p(𝜶|𝒙)p(𝒙)𝑑𝒙p\left({\boldsymbol{\alpha}}\right)=\int p\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right)p\left({\boldsymbol{x}}\right)d{\boldsymbol{x}} is intractable. Thus, we can let the Gaussian distribution r(𝜶)r\left({\boldsymbol{\alpha}}\right) approximate the marginal distribution p(𝜶)p\left({\boldsymbol{\alpha}}\right). Since the Kullback-Leibler divergence between two distributions is non-negative, we can get:

p(𝜶)logp(𝜶)𝑑𝜶p(𝜶)logr(𝜶)𝑑𝜶.\int p\left({\boldsymbol{\alpha}}\right)\log p\left({\boldsymbol{\alpha}}\right)d{\boldsymbol{\alpha}}\geq\int p\left({\boldsymbol{\alpha}}\right)\log r\left({\boldsymbol{\alpha}}\right)d{\boldsymbol{\alpha}}. (19)

Thus, the mutual information I(𝒙;𝜶)I({\boldsymbol{x}};{\boldsymbol{\alpha}}) have the upper bound:

I(𝒙;𝜶)p(𝒙,𝜶)logp(𝜶|𝒙)𝑑𝒙𝑑𝜶p(𝜶)logr(𝜶)𝑑𝜶=p(𝒙)p(𝜶|𝒙)logp(𝜶|𝒙)r(𝜶)d𝒙d𝜶.\begin{split}I({\boldsymbol{x}};{\boldsymbol{\alpha}})&\leq\int p\left({\boldsymbol{x}},{\boldsymbol{\alpha}}\right)\log p\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right)d{\boldsymbol{x}}d{\boldsymbol{\alpha}}-\int p\left({\boldsymbol{\alpha}}\right)\log r\left({\boldsymbol{\alpha}}\right)d{\boldsymbol{\alpha}}\\ &=\int p\left({\boldsymbol{x}}\right)p\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right)\log\frac{p\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right)}{r\left({\boldsymbol{\alpha}}\right)}d{\boldsymbol{x}}d{\boldsymbol{\alpha}}.\end{split} (20)

In practice, to compute this upper bound, we approximate distribution p(𝒙)p\left({\boldsymbol{x}}\right) with the empirical distribution:

p(𝒙)=1Ni=1Nδ𝒙i(𝒙).p\left({\boldsymbol{x}}\right)=\frac{1}{N}\sum^{N}_{i=1}\delta_{{\boldsymbol{x}}_{i}}\left({\boldsymbol{x}}\right). (21)

Thus, the upper bound of the mutual information I(𝒙;𝜶)I({\boldsymbol{x}};{\boldsymbol{\alpha}}) can be computed as:

p(𝒙)p(𝜶|𝒙)logp(𝜶|𝒙)r(𝜶)d𝒙d𝜶1Ni=1Nqθ(𝜶i|𝒙i)logqθ(𝜶i|𝒙i)r(𝜶i),\int p\left({\boldsymbol{x}}\right)p\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right)\log\frac{p\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right)}{r\left({\boldsymbol{\alpha}}\right)}d{\boldsymbol{x}}d{\boldsymbol{\alpha}}\approx\frac{1}{N}\sum^{N}_{i=1}q_{\theta}\left({\boldsymbol{\alpha}}_{i}|{\boldsymbol{x}}_{i}\right)\log\frac{q_{\theta}\left({\boldsymbol{\alpha}}_{i}|{\boldsymbol{x}}_{i}\right)}{r\left({\boldsymbol{\alpha}}_{i}\right)}, (22)

where qθ(𝜶|𝒙)q_{\theta}\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right) is a variational distribution with parameter θ\theta to approximate p(𝜶|𝒙)p\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right), and we implement this variational distribution with MLP:

qθ(𝜶|𝒙)=𝒩(𝜶|MLPμ(𝒙),MLPΣ(𝒙)),q_{\theta}\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right)=\mathcal{N}\left({\boldsymbol{\alpha}}|MLP^{\mu}\left({\boldsymbol{x}}\right),MLP^{\Sigma}\left({\boldsymbol{x}}\right)\right), (23)

where MLPμMLP^{\mu} and MLPΣMLP^{\Sigma} are implemented with MLPs which output the mean μ\mu and covariance matrix Σ\Sigma of the Gaussian distribution. In practice, we utilize the reparameterization trick [18] to sample from the Gaussian distribution to get an unbiased estimate of the gradient to optimize the opjective. Thus, the upper bound can be optimized by minimizing the KL divergence between the variational distribution qθ(𝜶|𝒙)q_{\theta}\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right) and the Gaussian distribution r(𝜶)r\left({\boldsymbol{\alpha}}\right):

minθ𝔼𝒙[DKL(qθ(𝜶|𝒙)||r(𝜶))].\min_{\theta}\mathbb{E}_{{\boldsymbol{x}}}\left[D_{KL}\left(q_{\theta}\left({\boldsymbol{\alpha}}|{\boldsymbol{x}}\right)||r\left({\boldsymbol{\alpha}}\right)\right)\right]. (24)

Appendix F Explanation of CIB with Information Plane

To better understand the concept information bottleneck (CIB) method, we provide an explanation of CIB with the information plane [8, 6].

Refer to caption
Figure 6: Explanation of CIB module with Information Plane.

As shown in Figure 6, the information plane is a two-dimensional space where the x-axis represents the mutual information between the original features and the calibrated features of CIB module I(𝒙^;𝜶^)I(\hat{\boldsymbol{x}};\hat{\boldsymbol{\alpha}}), and the y-axis represents the mutual information between the label information and the calibrated features I(𝐲;𝒙^)I(\boldsymbol{\rm{y}};\hat{\boldsymbol{x}}). The line colored orange in the information plane defines sufficiency, indicating where the calibrated features are maximally informative about the label information 𝐲\boldsymbol{\rm{y}}. The consistency definition falls to the left of the sufficiency line, denoting a state where the calibrated features, containing all information from both the original feature and the concept anchor, also include the label information. The InfoMAX principle [9] states that the optimal calibrated feature should be maximally informative about the original feature.

Ideally, a good calibrated feature 𝜶^\hat{\boldsymbol{\alpha}} should be maximally informative about the label information 𝐲\boldsymbol{\rm{y}} (the sufficiency line in the information plane) while minimally informative about the original features 𝒙^\hat{\boldsymbol{x}} (i.e., the ideal situation in the information plane). This situation can be achieved by directly supervised learning. However, the weakly supervised nature of MIL methods complicates this goal. As demonstrated in the ablation analysis in Section 4.4, if we directly perform superfluous information minimization without the guidance of concept anchor (i.e., predictive information maximization), the calibrated features will be less informative about the label information 𝐲\boldsymbol{\rm{y}} and the model performance will be degraded significantly. To address this issue, the proposed CIB module introduces the concept anchor 𝐜\boldsymbol{\rm{c}} to guide the learning process of information bottleneck for each instance. As shown in Figure 6, the CIB module could effectively reduce the mutual information between the original features and the calibrated features I(𝒙^;𝜶^)I(\hat{\boldsymbol{x}};\hat{\boldsymbol{\alpha}}) while preserving the mutual information between the label information and the calibrated features I(𝐲;𝜶^)I(\boldsymbol{\rm{y}};\hat{\boldsymbol{\alpha}}).

Appendix G Prompts

In this section, we provide the detailed prompt templates and concepts for the datasets, which are used to generate the concept anchor for the CIB module. The prompt templates are shown in Table 9, and the concepts for TCGA-BRCA, TCGA-NSCLC, and TCGA-RCC are shown in Table 10, Table 11, and Table 12, respectively. The templates and the class-agnostic prompts are referred from the original paper of CONCH [25], and the class-specific prompts are generated by querying the LLM with the question such as ‘In addition to tumor tissues, what types of tissue or cells are present in whole slide images of breast cancer?’ The quality of LLM generated prompts has been demonstrated in several recent studies [29, 22]. In principle, we can use LLM (e.g., GPT-4) to generate reliable expert-designed prompts and further verified by pathologists. This strategy can ensure the scalability and reliability of the prompts.

Table 9: Prompt Templates.
Templates
<CLASSNAME>.
a photomicrograph showing <CLASSNAME>.
a photomicrograph of <CLASSNAME>.
an image of <CLASSNAME>.
an image showing <CLASSNAME>.
an example of <CLASSNAME>.
<CLASSNAME> is shown.
this is <CLASSNAME>.
there is <CLASSNAME>.
a histopathological image showing <CLASSNAME>.
a histopathological image of <CLASSNAME>.
a histopathological photograph of <CLASSNAME>.
a histopathological photograph showing <CLASSNAME>.
shows <CLASSNAME>.
presence of <CLASSNAME>.
<CLASSNAME> is present.
an H&E stained image of <CLASSNAME>.
an H&E stained image showing <CLASSNAME>.
an H&E image showing <CLASSNAME>.
an H&E image of <CLASSNAME>.
<CLASSNAME>, H&E stain.
<CLASSNAME>, H&E
Table 10: Concepts for TCGA-BRCA.
Concept Type Classes Concept Prompts
Class-specific Concept IDC invasive ductal carcinoma breast invasive ductal carcinoma invasive ductal carcinoma of the breast invasive carcinoma of the breast, ductal pattern idc
ILC invasive lobular carcinoma breast invasive lobular carcinoma invasive lobular carcinoma of the breast invasive carcinoma of the breast, lobular pattern ilc
Class-agnostic Concept Adipocytes adipocytes adipose tissue fat cells fat tissue fat
Connective tissue connective tissue stroma fibrous tissue collagen
Necrotic Tissue necrotic tissue necrosis
Normal Breast Tissue Cells normal breast tissue normal breast cells normal breast
Table 11: Concepts for TCGA-NSCLC.
Concept Type Classes Concept Prompts
Class-specific Concept LUAD adenocarcinoma lung adenocarcinoma adenocarcinoma of the lung luad
LUSC squamous cell carcinoma lung squamous cell carcinoma squamous cell carcinoma of the lung lusc
Class-agnostic Concept Connective tissue connective tissue stroma fibrous tissue collagen
Necrotic Tissue necrotic tissue necrosis
Normal Lung Tissue Cells normal lung tissue normal lung cells normal lung
Table 12: Concepts for TCGA-RCC.
Concept Type Classes Concept Prompts
Class-specific Concept CCRCC clear cell renal cell carcinoma renal cell carcinoma, clear cell type renal cell carcinoma of the clear cell type clear cell rcc
PRCC papillary renal cell carcinoma renal cell carcinoma, papillary type renal cell carcinoma of the papillary type papillary rcc
CHRCC chromophobe renal cell carcinoma renal cell carcinoma, chromophobe type renal cell carcinoma of the chromophobe type chromophobe rcc
Class-agnostic Concept Adipocytes adipocytes adipose tissue fat cells fat tissue fat
Connective tissue connective tissue stroma fibrous tissue collagen
Necrotic Tissue necrotic tissue necrosis
Normal Kidney Tissue Cells normal kidney tissue normal kidney cells normal kidney

Appendix H More Visualization

To further illustrate the effectiveness of the proposed CATE-MIL, we provide more visualization results in this section. The visualization results for IDC and ILC in TCGA-BRCA are shown in Figure 7 and Figure 8, respectively.

Refer to caption
Figure 7: Visualization results for samples of IDC in TCGA-BRCA. (a) Attention heatmap of CATE-MIL. (b) Attention heatmap of original ABMIL. (c) similarity between the calibrated features and the corresponding class concept feature. (d) similarity between the original features and the corresponding class concept feature. (e) Original WSI. (f) UMAP visualization.
Refer to caption
Figure 8: Visualization results for samples of ILC in TCGA-BRCA. (a) Attention heatmap of CATE-MIL. (b) Attention heatmap of original ABMIL. (c) similarity between the calibrated features and the corresponding class concept feature. (d) similarity between the original features and the corresponding class concept feature. (e) Original WSI. (f) UMAP visualization.