Explainable and Position-Aware Learning in Digital Pathology
Abstract
Encoding whole slide images (WSI) as graphs is well motivated since it makes it possible for the gigapixel resolution WSI to be represented in its entirety for the purpose of graph learning. To this end, WSIs can be broken into smaller patches that represent the nodes of the graph. Then, graph-based learning methods can be utilized for the grading and classification of cancer. Message passing among neighboring nodes is the foundation of graph-based learning methods. However, they do not take into consideration any positional information for any of the patches, and if two patches are found in topologically isomorphic neighborhoods, their embeddings are nearly similar to one another. In this work, classification of cancer from WSIs is performed with positional embedding and graph attention. In order to represent the positional embedding of the nodes in graph classification, the proposed method makes use of spline convolutional neural networks (CNN). The algorithm is then tested with the WSI dataset for grading prostate cancer and kidney cancer. A comparison of the proposed method with leading approaches in cancer diagnosis and grading verify improved performance. The identification of cancerous regions in WSIs is another critical task in cancer diagnosis. In this work, the explainability of the proposed model is also addressed. A gradient-based explainbility approach is used to generate the saliency mapping for the WSIs. This can be used to look into regions of WSI that are responsible for cancer diagnosis thus rendering the proposed model explainable.
Index Terms:
Computational pathology, graph learning, position embedding, whole slide images.I Introduction
To diagnose and classify cancer, a tissue sample or excision need to be evaluated by a pathologist. To this end, tissue samples are processed and cut into small layers before being put on a glass slide and are then stained using histological hematoxylin and eosin (H&E). The pathologist can diagnose any malignancy by scanning and examining the slides under the microscope. Digitized versions of H&E slides are known as WSIs [1] where they can be digitized at different microscopic resolutions.
Despite the high-resolution and computational complexity of WSIs, their use in computational pathology and machine-based cancer diagnosis is increasing. In fact, recent advances in deep learning has facilitated the adoption of WSIs for the computer-aided diagnosis and grading of cancers. However, even with recent progresses made in computational methods for digital pathology, processing of WSIs for training in deep learning is considered a challenging task. The fact that WSIs have billions of pixels in a single file typically larger than a gigabyte, makes it challenging to train with common methods such as CNN.
Recently, multi-instance learning (MIL) approaches have been shown to be very successful in learning of the WSIs [2], [3], [4], [5], [6]. In MIL, instead of learning individual labels, the instances are bagged together and the training is performed on them. More specifically, the WSI can be broken into many patches and then bagged. In the case of binary classification tasks such as cancer diagnosis, if the bag consists of a single patch of cancerous tissue, the WSI is labeled as positive; otherwise it is categorized as negative. Once the patches are bagged together they are passed through CNN and transfer learning for training. Although, the accuracy obtained through this method is very promising, there are limitations that need to be considered. For example, MIL does not take into account the neighboring patches and some patches in WSI may not get selected for training.
Currently, the research in WSI learning is shifting from MIL learning to graph-based learning [7], [8], [9], [10]. Representation learning in graphs is performed through an encoder that maps nodes to a low-dimensional embedding space. Encoders encode nodes as low-dimensional vectors that summarize their local neighborhood. These embeddings enable tasks like node classification, link prediction, graph classification, and graph generation [11], [12], [13]. Recently, graph neural networks (GNNs) has become a very popular way to encode nodes for embedding. To learn node embeddings, GNNs take into account the local neighborhood information of a node. The advantage of graph-based learning over MIL is that full-resolution WSI can be represented as a graph, thus encoding tumor neighborhood information. Then, the whole graph is learnt through a variant of message passing algorithms followed by the global aggregation method and multi-layer perceptron (MLP) for the cancer classification.

Despite all the recent advances in the adoption of GNN in WSI classification, they fall short in distinguishing nodes with similar local neighborhoods [7], [8], [9], [10], [14], [15], [16]. In fact, the position of the nodes with respect to the graph is not captured in traditional GNN and if two nodes are located in topologically isomorphic neighborhoods, their embedding is almost identical. To this end, modifying GNN such that it takes into account node features and captures the position of the nodes is well motivated for improving representation learning [17].
With WSI broken into patches, one needs to take into account the spatial information of each patch and capture the information between those patches. This can be done by capturing the position embedding of each patch. Through an spline CNN algorithm [18], we aim at capturing the position embedding of patches. Once the position embedding of the patches is captured, we then use an attention-based method to perform the message passing among patches. As not all patches are cancerous in WSI, we want to learn differently from each patch and give different attention to it. This way, we can identify cancerous regions, more efficiently.
The degree to which the deep learning model can be explained is of critical significance. The more readily explicable a model is, the more effectively it may be interpreted. When the model’s interpretability and explainability are improved, the prediction that the model generates can be comprehended more easily. This helps in constructing models that are more trustworthy, and thus more applicable in medical diagnosis. Simple machine learning models, such as linear regression, are highly interpretable, but deep learning models are generally less interpretable due to their complexity. However, deeper models often offer a higher accuracy compared to simpler models [19]. Recently, there has been more focus on developing algorithms to make complex models explainable. The use of CNN has huge success in vision related task such as image classification, segmentation and object detection. The explainablity of CNN-based algorithms has been widely explored in recent years. In [20], [19], [21], [22] CNN model is made explainbale by visualizing the area of interest in image classification. The explainability model for CNN model has been expanded to graph-based models as well in [23]. In graph-based learning the explainabilty amounts to identifying the importance of each node in prediction.
The explainability of a model is of paramount importance in the context of cancer detection, as it helps to increase trust in the model’s predictions. When working with WSI, it is essential to pinpoint the particular areas of the image that are accountable for making the cancer diagnosis. Because malignant cells may be concentrated in some areas, these regions might not necessarily be present throughout the entirety of the image. In this work, the goal is to use an explainability method that is based on graph neural networks in order to locate areas related top malignancy inside WSIs. By identifying the specific regions of the image that are responsible for the cancer diagnosis, the model can provide more accurate and trustworthy predictions.
To summarize, the contribution of this work is three-fold: 1) It introduces a graph-based self-supervised position-aware algorithm for cancer diagnosis using WSIs; 2) To encode position information of the patches in the WSIs obtained through spline CNN, an attention network is proposed to promote different weights on cancerous versus non-cancerous nodes; and 3) Based on gradient-weighted class activation mapping (Grad-CAM), an explainability algorithm for the proposed model is developed. Using Grad-CAM the heatmaps are generated to identify the regions of interest in WSIs which can be later used to visualize and detect the cancerous regions.
This paper is organized as follows. Section II addresses the related work. In section III, the proposed method is presented. The implementation details are shown in section IV. Section V presents the performance of the algorithm for different types of cancer WSIs and discussion of the results is followed in section VI. The conclusion is provided in section VII.

II Related Work
The use of GNN-based methods are increasing in WSI for cancer diagnosis. The WSIs are images of digitized histopathological slides and are not originally represented as graphs. So, the first challenge in using graph-based learning approaches is to find a way to represent the image data as a graph. Once WSIs are encoded as graphs, different graph representation learning methods can be deployed. The common approach to construct the graph is to form cell graph [24]. In this method, first the nuclei in the WSIs are identified and represented as the nodes in the graph. Then, the graph is formed with edges connecting other neighboring nuclei in the tissue. Another approach is to form patch-graph method. In this method, the WSI is broken into patches as nodes and the graph is constructed from those nodes.
In [8], the authors addressed the classification of WSIs for breast cancer using graph convoutional network (GCN) [12]. More specifically, the WSIs were constructed as graphs using the constituents cells with nuclear morphology as vertex features and the gland formation as edge attributes. In [15], a model for gleason score grading of prostate cancer has been proposed. The authors constructed a graph with nuclei as the nodes and the distance between each node as the edge. The features of each node are extracted using contrastive predictive coding and morphological features. Then, this graph was passed into GCN for the grading of high-risk and low-risk cancer patients. These methods use cell nuclei for the construction of the graphs from WSIs.
Other methods used in the graph construction from WSIs include patch-based methods. In [14] GCN-based multi-instance learning method is implemented for the classification of lymph node. First, the patches with regions of interest are selected and bagged together. Then, this bag of patches is passed into auto-encoder and GCN for feature selection and classification, respectively. In [10] the genetic mutation of colon cancer is modeled. Randomly selected patches from each WSIs act as nodes and CNN is used to extract features for each node. The constructed graph is passed through spatial-GCN to predict the mutations. Survival analysis is performed using graph-based methods in [9].
The existing methods for the graph-based learning of WSIs use popular message passing techniques such as GCN and graph attention network (GAT), which do not take node position in the graph into account. In this work, position embedding is also captured.
Explainability is another aspect in improving the interpretability of a model. This makes it easy to explain the model and its outcome. The literature in explainabilty of graph-based models in WSIs is very limited. In [25] the authors have presented an explainability method based on the importance of the nuclei for cancer classification.
III Method
The system diagram of the proposed model is presented in Fig. 1. The process begins by converting the WSI into a graph, with each patch of the WSI being represented as a node and edges connecting those nodes. In this work, the model aims to learn the position of each node in the graph. This allows to maintain the spatial information of each patch in WSI even though it is represented as graph data. We deploy Spline CNN to learn the position of each node in the WSI [18].
After capturing the position embedding, the graph is passed through the attention network. The model consists of 2 layers of attention network. The first layer of attention network consists of multi-headed attention followed by single attention head network. Following the attention layer, the model classifies the WSI using the graph classification. The results from attention layers are global pooled using mean pooling. This layer is forwarded to multi-layer perceptron network and then to a classification layer for cancer grading.
III-A Graph construction and feature extraction
The graph data structure is represented as where , and denote the set of nodes, the edge set, and node features, respectively. It is necessary to transform the WSI into graphs in order for each WSI to be represented as .
The aim of using graph-based learning for WSIs is to learn from the whole resolution of WSI. So, the graph is constructed from the highest level of resolution of WSI. The graph formation of WSI is implemented by breaking it into smaller patches. In this work, WSI is broken into non-overlapping patches of size . Since each patch acts as a node in the graph, its position in the WSI is used to construct connectivity between corresponding neighbors. Based on fast approximate k-nearest neighbor [26], the connectivity between nodes are established. Since each WSI is different in size, the number of nodes in each graph is different. For the constructed graph in this work, we consider as the set of patches (nodes) and the edge set is given by connection between those patches. With as the number of nodes in the graph , the set of features for each node, i.e., is obtained by extracting features from each node, where is the feature vector of node . For feature extraction of each patch, self-supervised learning method is deployed. Typically, medical data suffers from limited number of annotated data. The lack of annotated data makes self-supervised algorithms suitable for extracting features from patches. More specifically, contrastive learning framework is used to train patches [27], [7]. This way, feature extraction of patches is performed with an unsupervised learning approach. In Fig. 2, the feature extraction method for patches based on contrastive learning is visualized. The contrastive learning approach maximizes the similarity between the positive pair and dissimilarity among the negative pair. The augmentation is used to generate multiple view of the same images to create the positive pairs. When the network trains the augmented view of different patches, those are considered to be negative pairs.
III-B Position embedding in the graph
Graph learning entails embedding nodes into lower dimensions. The node embedding is performed by an encoder based on different information about the node and its neighbors. The embedding space is optimized so that the distance between nodes captures the relative position of the original graph. The encoder function maps nodes to embeddings , where refers to feature size in the embedding space. The most common approaches for embedding include matrix factorization [28], [29], random walk and neural learning methods such as graph neural learning. In matrix factorization, node embedding is constructed based on node similarity matrix. Some of the popular random walk methods include DeepWalk [30], node2vec [31]. With capability to learn from node features and inductive learning, GNNs are most commonly used in node embedding.

Due to unstructured nature of graph data, the positions of the nodes are not naturally encoded in them. In [17], it is shown that popular graph-based learning methods such as GNN are capable of capturing structural information, but the position of the node is not taken into account. As a result, the nodes with different labels having similar neighborhoods will end up with the same embedding and similar representations. To avoid this issue and make the learning aware of the position of the nodes in the graph, it is useful to have positional embedding.
In this work, WSIs are represented with 2D coordinates and the goal is to classify each WSI incorporating the position of each patch in the WSI graph. The coordinates of each patch is the midpoint of each patch in the WSI. The proposed algorithm to capture the position of the nodes is based on spline CNN [32], [33]. Spline CNN uses x,y coordinates to learn the geometry of the graph. We deploy B-splines [34] as presented in [18] to learn nodes positions. In Spline CNN, with set of neighboring nodes defined as , the feature vector of node , denoted by is given by [32], [18]:
(1) |
where is a kernel function based on the coordinates of nodes and . Suppose the coordinate of node is given by and the coordinate of node is , then . The kernel function is defined as follows:
(2) |
In the above equation, and are B-splines bases, is a trainable parameter for each element in the Cartesian product of and .
III-C Graph attention network
As defined in section III, a graph can be represented as where denotes the set of nodes with number of nodes , refers to the set of edges. Then, accounts for the nodes feature vectors. Graph learning entails embedding graph nodes to a lower dimensional space.
GNN is one of the most popular embedding methods used for graph representation learning where the learning is based on passing and updating messages among the nodes, iteratively.
Let us denote the feature vector after message passing as in the next layer. The message passing for each node in GNN [35] is based on the neighboring nodes and given by:
(3) |
where is the feature vector of the node whose message is being updated, and are the neighboring nodes feature vectors. Then, are the trainable weight parameters updated during the message passing from itself and neighboring nodes, respectively and is the activation function.
Graph convolutional neural network (GCN) [12], as a variant of GNN is very similar to CNN. In GCN message passing aggregates the message from its neighboring nodes features. The message update for the GCN follows:
(4) |
where is the activation function and accounts for the non-linearity, and are the trainable parameters updated during the training of GCN. is The degree of neighboring node is denoted by and accounts for the degree of the node for which message passing is being performed.
Using GCN, the message passing to a node is achieved via its neighboring nodes and all the neighboring nodes are equally accountable. The graph attention network (GAT) modifies this convolutional operator by assigning attention weight in message passing stage among nodes. The attention mechanism allows the network to learn the specific weights from the specific nodes. This allows to learn the importance of each neighboring nodes in message passing.
More specifically, the message passing in GAT [36] is given by:
(5) |
where is the attention weight between nodes and node . The attention weights are given by:
(6) |
where is the concatenation operator and Leaky Rectified Linear Unit (LeakyReLU) was used as activation function.
The stabilization of the self-attention mechanism is obtained by adding the multi-head attention. This operation is achieved by applying independent attention-based operators and then concatenating them. Adopting multi-head attention network with heads, (5) is replaced by:
(7) |
In this work, two layers of multi-head GAT is considered.
Use of attention-based method allows for different weight to nodes in the network. In cancer diagnosis, through a graph-based WSI, neighboring nodes might consist of both cancerous and non-cancerous patches. To this end, assigning different importance weights to different nodes in a neighboring region in message passing is well motivated.
III-D Explainability of the model
The application of GNN in cancer classification has the potential to be of significant assistance if the regions responsible for cancer classification can be pinpointed. The physicians and healthcare team need a model to be interpretable, so that the reasoning behind the suggested prediction can be well justified and understood. When it comes to the task of predicting cancer especially using deep learning models, explainability is of the paramount importance. When WSI is examined under the microscope by a pathologist, the pathologist is screening for possible areas inside the tissue that could be to blamed for cancer. The classification of WSI, together with the identification of the regions that causes a particular model for classification in WSI, could be of tremendous assistance to pathologists as well.
In this paper, we deploy explainability method implemented in graphs that are mainly tailored to CNN [19], [37]. In particular, graph Grad-CAM method for WSI is implemented. The choice of this approach is inspired by [23] as the authors address explainability for molecular graphs.
As shown in (4) in the layer of GNN, feature of the node is given by . Then, the average pooling of the feature at layer is given by:
(8) |
Based on this average pooling score , and weights for predicting class from feature , the class score can be calculated as shown below:
(9) |
The following equation for the Grad-CAM gives each feature importance score based on class and layer:
(10) |
Then, the heatmap for each node is generated by applying Rectified Linear Units (ReLU) [38] functions as:
(11) |
The advantage of using Grad-CAM is that the heatmaps for the nodes can be calculated from any layer of the network.
![]() |
(a) Sample WSIs from Prostate Cancer Dataset |
![]() |
(a) Sample WSIs from Kidney Cancer Dataset |
IV Implementation Details
IV-A Dataset
Using two different datasets, we examined how well the proposed position-aware and graph attention-based model works for grading WSIs. To be more specific, we tested the performance of models with samples from both prostate and kidney cancers. After lung cancer, prostate cancer is the type of cancer that is responsible for the second highest number of deaths in men [39]. Kidney cancer is a prevalent form of cancer that can impact both sex and ranks in the top 10 cancers that result in mortality [39]. The prostate dataset consists of biopsies from the WSIs taken from biopsies from the prostate glands and the WSIs for kidney dataset are obtained from the kidney resections (larger samples). The sample WSIs of each dataset are shown in Fig. 4. It can been seen that the structure of a biopsy is very different from a kidney resection as shown in Fig. 4,
The Cancer Genome Atlas (TCGA) has been used for the kidney cancer dataset [40]. The TCGA portal provides various kinds of data for cancer research, such as demographics of the patient, WSIs, and biomarkers. The data related to kidney cancer consists of the most prominent type of kidney cancer, renal cell carcinoma, which is addressed in this paper. In kidney cancer, the most prevalent malignancy Renal Cell Carcinoma [41] accounts for more than 90% of cases. In our work, three major types of RCC grouped as KIRC (Kidney Renal Clear Cell Carcinoma), KIRP (Kidney Renal Papillay Cell Carcinoma) and KICH (Kidney Chromophobe) are considered. The WSI dataset related to prostate cancer has been accessed from [42]. In prostate cancer, we look into the problem of grading the prostate cancer into different ISUP grades. ISUP grades consist of 5 grades of Gleason Grading (GG) [43] for the prostate cancer. In Table I the number of samples of each gleason grading for prostate cancer is listed. Similarly, Table II shows the number samples of each type of cancer in kidney dataset.
Description | Number of samples |
---|---|
GG1 | 2666 |
GG2 | 1343 |
GG3 | 1250 |
GG4 | 1242 |
GG5 | 1224 |
Description | Number of samples |
---|---|
KIRC | 519 |
KIRP | 300 |
KICH | 121 |
IV-B Training settings
Initially, the patches were trained for feature extraction for graph learning. To handle unlabeled patches, contrastive learning as a self-supervised method was used to train the patches. Then, MOCOv3 [45] with ResNet50 [46] as backbone was deployed to train the patches for feature extraction. The training was performed for 75 epochs with initial learning rate of and the weight decay of . Cosine scheduler was chosen to update the learning rate during the training.
For graph learning training, the network consists of two layers of Spline CNN followed by two layers of graph attention. This is then forwarded to a MLP for cancer grading. The training was done for 20 epochs with initial learning rate of and weight decay of . The AdamW [47] optimizer was used for training. The batch size for each epoch was chosen as 2. We use Kappa score [48] to measure the performance of the model.
![]() |
||
(i) | (ii) | (iii) |
(a) Activation map generated for the cancer with ISUP grade 4 | ||
![]() |
||
(i) | (ii) | (iii) |
(b) Activation map generated for the cancer with ISUP grade 5 |
V Results
The proposed approach was evaluated using the two different cancer datasets. The performance of the proposed model is compared with self-supervised GCN-based model [7] and MIL-based leading methods [44].
The position-aware method that was proposed for grading and classifying cancer was initially evaluated using prostate cancer dataset. To this end, a classification of cancer based on one of the the five ISUP grades can be obtained. On this dataset, the proposed approach yields a Kappa score of 0.912 as its outcome. Evaluating the cancer diagnosis and grading with a GCN-based and a MIL-based models for comparison, the Kappa scores are 0.899 and 0.87, respectively.
The proposed method was also applied to a dataset consisting of kidney cancer data. The obtained kappa scores are very comparable. The proposed model obtaines a Kappa score of 0.941 when it was applied to the kidney dataset. In comparison, the Kappa score for the GCN-based and MIL-based models are 0.939, and 0.868, respectively. As can be seen in Table III, the addition of patch position into WSI graph learning results in a significant improvement in cancer grading and classification.
In general, these results point to the fact that the proposed position-aware method is a promising approach for the diagnosis and treatment of cancer. This was verified with different types of WSI samples including both prostate biopsies and Kidney resections to classify tumors affecting the prostate and the kidney.
V-A Explainability visualized to the WSI
To better demonstrate the power of the proposed algorithm, it is crucial to investigate the explainabilty of the model. In other words, the expectation is that the algorithm can identify the regions and areas in the WSI that are responsible for the cancer diagnosis. This result verifies that the proposed model is interpretable. To this end, the heatmaps for each node is generated after the activation layer and is superimposed with the WSI. Then, the generated heatmap is compared with the original annotation provided in the dataset.
The heatmaps are generated for two cases of prostate cancer, one with an ISUP grade of 4 and the other with an ISUP grade of 5. In Fig. 5 the top image shows the heatmap for prostate cancer with an ISUP grade of 4, while the bottom image shows the heatmap for the prostate cancer with an ISUP grade of 4. The top left image (i) shows the original WSI for ISUB grade 4. As we can see in the image the areas where the cancer is spread the tissue is not smooth and the growth pattern is irregular. This is further confirmed in the top right image in (iii) which shows the annotation provided for the WSI and clearly depicts the cancerous regions in the WSI. The image on the top middle (ii) is generated using Grad-CAM for ISUB grade 4 WSI and the red regions demonstrate the areas responsible for the cancer diagnosis. The identified regions by Grad-CAM are very aligned with the annotations provided by the pathologist as shown in (iii). Similar trends hold for WSI of ISUB grade 5, as shown in Fig 5, (b). As it can be seen, the annotation matches with the cancerous areas identified by the Grad-CAM algorithm.
VI Discussion
In this work, graph-based leaning incorporating the position of each node embedded in the WSI is developed and used to classify the prostate and kidney cancer. WSIs are large dimensional, multi-resolution digital slides that store microscopic images of tissues to be used for cancer diagnois. However, due to their size and complexity, it can be difficult to extract useful information from them using traditional methods such as CNN.
To address this challenge, WSIs are broken into smaller patches. Then, the patches are used to form a graph structure. When broken into smaller patches, the position of each patch in the WSI is not encoded in the graph data. This paper takes into account these position embeddings in the training. This is accomplished with the use of a graph kernel-based technique known as Spline CNN, which incorporates the location of each patch into the graph itself.
An attention-based mechanism is employed for message passing between the nodes in the graph once the position of each patch has been incorporated in the network. Since the attention layer gives varying attention weights to the nodes, the model is able to focus in on specific regions of the image and concentrate on some of them to a greater extent than others. This attention-based strategy is used in combination with graph Grad-CAM to generate heatmaps to identify the regions responsible for the diagnosis. These heatmaps provide visual explanations of the model’s prediction and can be compared with the pathologists annotation to better understand and judge the performance of the model. In addition, one can use these heatmaps as initial screening of the pathologist to detect cancerous regions before final cancer grading.
The kappa score, which is a statistical measure of inter-rater agreement, was used to make a comparison between the performance of the position-aware model and that of other approaches. The results of this comparison indicated that the position-embedded method performed better than the other methods when applied to the datasets pertaining to prostate and kidney cancers.
To address the explianabilty of the approach, this paper tailors the graph Grad-CAM algorithm, which was originally designed for use in other graph learning-based applications, to the proposed self-supervised position-aware algorithm for WSIs. The explainabilty of the model is based on the attention score of the second activation layer of the model. The Grad-CAM approach takes into account each node’s attention score before generating a heatmap for that node. The score that is associated with each node is the one that is used for the patch that is associated with that node in the WSI. All the heatmaps for node is superimposed with the complete WSI to generate the heatmap in the WSI. By visualizing the heatmaps for each node, it is possible to understand which regions of the WSI are most important to the GNN’s prediction, and how these regions contribute to the overall prediction made by the model.
VII Conclusion
A self-supervised position-aware attention-based model for cancer grading and classification utilizing WSIs has been developed in this paper. The position embedding of the nodes in the WSI was accomplished by using spline CNN. In order to evaluate the effectiveness of the proposed approach, data sets of WSIs for prostate and kidney cancers were utilized. The proposed method surpasses established methodologies in cancer diagnosis, such as GCN-based and MIL-based models using kappa score as evaluation metric. In addition to cancer classification and grading, it was shown that the proposed approach is very explainable and the diagnosis supports the annotated regions in the WSI by the pathologist.
References
- [1] F. Aeffner, M. D. Zarella, N. Buchbinder, M. M. Bui, M. R. Goodman, D. J. Hartman, G. M. Lujan, M. A Molani, A. V Parwani, K. Lillard, O. C. Turner, V. N. P. Vemuri, A.G. Yuil-Valdes, and D. Bowman, “Introduction to digital image analysis in whole-slide imaging: A white paper from the digital pathology association,” J. Pathol. Inform., vol. 10, no. 1, pp. 9, Mar. 2019.
- [2] C. Mercan, S. Aksoy, E. Mercan, L.G. Shapiro, D.L. Weaver, and J.G. Elmore, “Multi-instance multi-label learning for multi-class classification of whole slide breast histopathology images,” IEEE Transactions on Medical Imaging, vol. 37, no. 1, pp. 316–325, 2018.
- [3] J. Melendez, B. Ginneken, P. Maduskar, R. Philipsen, H. Ayles, and C. Sanchez, “On combining multiple-instance learning and active learning for computer-aided detection of tuberculosis,” IEEE Transactions on Medical Imaging, vol. 35, pp. 1–1, 12 2015.
- [4] G. Quellec, M. Lamard, M. Cozic, G. Coatrieux, and G. Cazuguel, “Multiple-instance learning for anomaly detection in digital mammography,” IEEE Transactions on Medical Imaging, vol. 35, no. 7, pp. 1604–1614, 2016.
- [5] G. Ampanella, M.G. Hanna, Geneslaw, A. Miraflor, W. Krauss, V. Silva, K.J. Busam, E. Brogi, V.E. Reuter, D.S. Klimstra, and T.J. Fuchs, “Clinical-grade computational pathology using weakly supervised deep learning on whole slide images.,” Nature Medicine, 2019.
- [6] M.Y. Lu, R.J. Chen, J. Wang, D. Dillon, and F. Mahmood, “Semi-supervised histology classification using deep multiple instance learning and contrastive predictive coding,” Advances in Neural Information Processing Systems (NeurIPS) Workshop in Machine Learning for Health, 2019.
- [7] M. Aryal and N.Y. Soltani, “Context-aware graph-based self-supervised learning of whole slide images,” in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2022.
- [8] D. Anand, S. Gadiya, and A. Sethi, “Histographs: graphs in histopathology,” in Medical Imaging 2020: Digital Pathology, John E. Tomaszewski and Aaron D. Ward, Eds. International Society for Optics and Photonics, 2020.
- [9] R.J. Chen, M.Y. Lu, M. Shaban, C. Chen, T.Y. Chen, D.F.K Williamson, and F. Mahmood, “Whole slide images are 2d point clouds: Context-aware survival prediction using patch-based graph convolutional networks,” in Medical Image Computing and Computer Assisted Intervention – MICCAI 2021, pp. 339–349. Springer International Publishing, 2021.
- [10] K. Ding, Q. Liu, E.H. Lee, M. Zhou, A. Lu, and S. Zhang, “Feature-enhanced graph networks for genetic mutational prediction using histopathological images in colon cancer,” in MICCAI, 2020.
- [11] W. Hamilton, Z. Ying, and J. Leskovec, “Inductive representation learning on large graphs,” in Advances in Neural Information Processing Systems, 2017.
- [12] T.N. Kipf and M. Welling, “Semi-supervised classification with graph convolutional networks,” in International Conference on Learning Representations (ICLR), 2017.
- [13] Z. Ying, J. You, C. Morris, X. Ren, W. Hamilton, and J. Leskovec, “Hierarchical graph representation learning with differentiable pooling,” in Advances in Neural Information Processing Systems, 2018.
- [14] Y. Zhao, F. Yang, Y. Fang, H. Liu, N. Zhou, J. Zhang, J. Sun, S. Yang, B. Menze, X. Fan, and J. Yao, “Predicting lymph node metastasis using histopathological images based on multiple instance learning with deep graph convolution,” in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020.
- [15] J. Wang, R.J. Chen, M.Y. Lu, A. Baras, and F. Mahmood, “Weakly supervised prostate tma classification via graph convolutional networks,” in ISBI 2020 - 2020 IEEE International Symposium on Biomedical Imaging, Apr. 2020.
- [16] V. Anklin, P. Pati, G. Jaume, B. Bozorgtabar, A. Foncubierta-Rodríguez, J.P. Thiran, M. Sibony, M. Gabrani, and O. Goksel, “Learning whole-slide segmentation from inexact and incomplete labels using tissue graphs,” 2021.
- [17] J. You, R. Ying, and J. Leskovec, “Position-aware graph neural networks,” in Proceedings of the 36th International Conference on Machine Learning, Kamalika Chaudhuri and Ruslan Salakhutdinov, Eds., 09–15 Jun 2019.
- [18] M. Fey, J. Lenssen, F. Weichert, and H. Muller, “Splinecnn: Fast geometric deep learning with continuous b-spline kernels,” in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2018.
- [19] R.R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and D. Batra, “Grad-cam: Visual explanations from deep networks via gradient-based localization,” in 2017 IEEE International Conference on Computer Vision (ICCV), 2017.
- [20] C. Vondrick, A. Khosla, T. Malisiewicz, and A. Torralba, “Hoggles: Visualizing object detection features,” in 2013 IEEE International Conference on Computer Vision, 2013.
- [21] M. D. Zeiler and R. Fergus, “Visualizing and understanding convolutional networks,” in Computer Vision – ECCV 2014, 2014.
- [22] K. Simonyan, A. Vedaldi, and A. Zisserman, “Deep inside convolutional networks: Visualising image classification models and saliency maps,” 2013.
- [23] P. E. Pope, S. Kolouri, M. Rostami, C. E. Martin, and H. Hoffmann, “Explainability methods for graph convolutional neural networks,” in 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
- [24] D. Ahmedt-Aristizabal, M. A. Armin, S. Denman, C. Fookes, and L. Petersson, “A survey on graph-based deep learning for computational histopathology,” 2021.
- [25] G. Jaume, P. Pati, B. Bozorgtabar, A. Foncubierta-Rodríguez, F. Feroce, A. M. Anniciello, T. Rau, J.P. Thiran, M. Gabrani, and O. Goksel, “Quantifying explainers of graph neural networks in computational pathology,” 2020.
- [26] M. Muja and D. G. Lowe, “Fast approximate nearest neighbors with automatic algorithm configuration,” International Conference on Computer Vision Theory and Applications, pp. 331–340, 2009.
- [27] T. Chen, S. Kornblith, M. Norouzi, and G. Hinton, “A simple framework for contrastive learning of visual representations,” Proceedings of the 37th International Conference on Machine Learning, vol. 119, pp. 1597–1607, 13–18 Jul 2020.
- [28] A. Ahmed, N. Shervashidze, S. Narayanamurthy, V. Josifovski, and A. J. Smola, “Distributed large-scale natural graph factorization,” in Proceedings of the 22nd International Conference on World Wide Web, New York, NY, USA, 2013.
- [29] M. Belkin and P. Niyogi, “Laplacian eigenmaps and spectral techniques for embedding and clustering,” in Proceedings of the 14th International Conference on Neural Information Processing Systems: Natural and Synthetic, Cambridge, MA, USA, 2001.
- [30] B. Perozzi, R. Al-Rfou, and S. Skiena, “Deepwalk: Online learning of social representations,” in Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 2014.
- [31] A. Grover and J. Leskovec, “Node2vec: Scalable feature learning for networks,” in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 2016.
- [32] Z. Jiang, H. Rahmani, P. Angelov, S. Black, and B.M. Williams, “Graph-context attention networks for size-varied deep graph matching,” in 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022.
- [33] M. Rolínek, P. Swoboda, D. Zietlow, A. Paulus, V. Musil, and G. Martius, “Deep graph matching via blackbox differentiation of combinatorial solvers,” in Computer Vision – ECCV 2020, 2020.
- [34] L. Piegl and W. Tiller, The NURBS book, 1996.
- [35] W. L. Hamilton, “Graph representation learning,” Synthesis Lectures on Artificial Intelligence and Machine Learning, pp. 1–159, 2020.
- [36] P. Veličković, G. Cucurull, A. Casanova, A. Romero, P. Liò, and Y. Bengio, “Graph Attention Networks,” International Conference on Learning Representations, 2018.
- [37] K. Simonyan, A. Vedaldi, and A. Zisserman, “Deep inside convolutional networks: Visualising image classification models and saliency maps,” 2013.
- [38] V. Nair and G. E. Hinton, “Rectified linear units improve restricted boltzmann machines,” in Proceedings of the 27th International Conference on International Conference on Machine Learning, 2010.
- [39] R. L. Siegel, K. D. Miller, H. E. Fuchas, and A. Jemal, “Cancer statistics, 2021,” CA: A Cancer Journal for Clinicians, pp. 7–33, 2021.
- [40] O. Akin, P. Heller, M. Jarsosz, B.J. Kirk, and J. Filippni, “Radiology data from the cancer genome atlas kidney renal clear cell carcinoma [tcga-kirc] collection. the cancer imaging archive.,” 2016.
- [41] J.J. Hsieh, M. P. Purdue, S. Signoretti, C. Swanton, L. Albiges, M. Schmidinger, D.Y. Heng, J. Larkin, and V. Ficarra, “Renal cell carcinoma,” Nat. Rev. Dis. Primers, vol. 3, pp. 17009, Mar. 2017.
- [42] “Artificial intelligence for diagnosis and gleason grading of prostate cancer: the panda challenge,” NATURE MEDICINE, vol. 28, pp. 154–163, 2022.
- [43] N. Chen and Q. Zhou, “The evolving gleason grading system,” Chin. J. Cancer Res., vol. 28, no. 1, pp. 58–64, Feb. 2016.
- [44] X. Wang, Y. Yan, P. Tang, X. Bai, and W. Liu, “Revisiting multiple instance neural networks,” Pattern Recognition, vol. 74, pp. 15–24, Feb 2018.
- [45] X. Chen, S. Xie, and K. He, “An empirical study of training self-supervised vision transformers,” in 2021 IEEE/CVF International Conference on Computer Vision (ICCV), 2021.
- [46] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016.
- [47] I. Loshchilov and F. Hutter, “Decoupled weight decay regularization,” in International Conference on Learning Representations, 2019.
- [48] J. Cohen, “Weighted kappa: Nominal scale agreement with provision for scaled disagreement or partial credit.,” Psychological Bulletin 70 (4), pp. 213–220, 1968.