Department of Creative Informatics
Graduate School of Information Science and Technology
June \degreeyear1990 \thesisdateMay 18, 1990
William J. SupervisorAssociate Professor
Arthur C. ChairmanChairman, Department Committee on Graduate Theses
The University of Tokyo
Dedicated with love to my mother, for always believing and supporting;
and to my sister, for her unbounded trust and inspiration.
Acknowledgments
Most of all, I want to thank my mother and sister for raising me to value education and science—I would not be writing this thesis without their life-long dedicated support and disarming warmth. My second deepest appreciation goes to Prof. Hideki Nakayama for taking me under his wing and supporting me to explore creative and informative ideas, just like our department name. He kindly allowed me to challenge many different ideas despite my shortcomings. I am grateful to Research Center for Medical Big Data staff Prof. Fuyuko Kido, Dr. Youichirou Ninomiya, Fumika Tamura, and Prof. Shin’ichi Satoh, for offering me amazing research opportunities and helping me apply for Japanese permanent residency. In particular, I always enjoyed discussions and consequent publications together with you, Prof. Murao Kohei.
Many thanks to all people committed to our GAN projects. To physicians, including Dr. Yusuke Kawata and Dr. Fumiya Uchiyama (National Center for Global Health and Medicine), for their brilliant clinical insights and contributions. Especially, I owe entirely to Dr. Tomoyuki Noguchi (Kyushu Medical Center) for always sparing no effort to protect health and save lives with state-of-the-art technology—he introduced me many competent physicians for evaluating our GAN applications’ clinical relevance. Dr. Yujiro Furukawa (Jikei University School of Medicine) has always supported our GAN projects from the very first project, trusting its potential despite repeated urgent requests. Dr. Kazuki Umemoto (Juntendo University School of Medicine) is AI-enthusiastic and very passionate about advancing healthcare.
This thesis would not have been achievable without Ryosuke Araki (Chubu University) and Prof. Hideaki Hayashi (Kyushu University) because we launched and carried out these tough but fruitful projects together for so long. During my internship at FujiFilm with Dr. Yoshiro Kitamura, Akira Kudo, Akimichi Ichinose, and Dr. Yuanzhong Li, my best memory was drinking endless free beer.
I appreciate those participated in our questionnaire survey and GCL workshop: Dr. Yoshitaka Shida, Dr. Ryotaro Kamei (National Center for Global Health and Medicine), Dr. Toshifumi Masuda (Kyushu Medical Center), Dr. Ryoko Egashira, Dr. Yoshiaki Egashira, Dr. Tetsuya Kondo (Saga University), Dr. Takafumi Nemoto (Keio University School of Medicine), Dr. Yuki Kaji, Miwa Aoki Uwadaira, Hajime Takeno (The University of Tokyo), and Kohei Yamamoto (Corpy&Co., Inc.).
My sincere thanks go to Prof. Takeo Igarashi, Prof. Tatsuya Harada, Prof. Kazuhiko Ohe, Prof. Manabu Tsukada, and Prof. Murao Kohei for refereeing my Ph.D. thesis. Including Prof. Tsukada’s mentoring, I was privileged to have continuous financial and human support from GCL program of my University by MEXT.
Clinically valuable research requires international/interdisciplinary collaboration. Backed up by the GCL, I achieved this goal thanks to Dr. Leonardo Rundo, real friend who invited me to Università degli Studi di Milano-Bicocca twice, University of Cambridge once, and his hometown in Sicily twice. After working hard together everyday from everywhere, he understands this thesis better than anyone in the world. Special thanks also to Prof. Giancarlo Mauri on his 70th birthday in Milan and Prof. Evis Sala in Cambridge, along with the other (mostly Italian) guys in their labs, including Prof. Daniela Besozzi, Prof. Paolo Cazzaniga, Prof. Marco Nobile, Dr. Andrea Tangherloni, and Simone Spolaor. After visiting 19 Italian cities, sono quasi Italiano! Nicolas Y. Kröner was my geek tutor at Technische Universität München.
I wish to thank all my pleasant friends for making my Ph.D. days enjoyable and memorable in these intense years. Cheers to our long friendship, UTDS members, including Marishi Mochida and Kazuki Taniyoshi. Again, congratulations on your happy wedding, Yuki/Mari Inoue and Akito/Hitomi Misawa. Moreover, I shared far too many hilarious memories with KEK members Zoltán Ádám Milacski and Florian Gesser. Kurumi Nagasako drew this thesis’ very first figure. Prof. Yumiko Furuichi and Miwako Hayasaka are always cheerful and smiling like Mother Teresa.
“My true title of glory is that I will live forever” Last, but not least, I would like to thank Napoleon Bonaparte who taught me that nothing is impossible. With courage and hope, I could conduct this research for both human prosperity and individual happiness. Listening to Beethoven’s Symphony No.3 Eroica in the morning, I always try to be an everyday hero with great ambition like you. Your reign will never cease.
Abstract
Convolutional Neural Networks (CNNs) can play a key role in Medical Image Analysis under large-scale annotated datasets. However, preparing such massive dataset is demanding. In this context, Generative Adversarial Networks (GANs) can generate realistic but novel samples, and thus effectively cover the real image distribution. In terms of interpolation, the GAN-based medical image augmentation is reliable because medical modalities can display the human body’s strong anatomical consistency at fixed position while clearly reflecting inter-subject variability; thus, we propose to use noise-to-image GANs (e.g., random noise samples to diverse pathological images) for (i) medical Data Augmentation (DA) and (ii) physician training. Regarding the DA, the GAN-generated images can improve Computer-Aided Diagnosis based on supervised learning. For the physician training, the GANs can display novel desired pathological images and help train medical trainees despite infrastructural/legal constraints. This thesis contains four GAN projects aiming to present such novel applications’ clinical relevance in collaboration with physicians. Whereas the methods are more generally applicable, this thesis only explores a few oncological applications.
In the first project, after proposing the two applications, we demonstrate that GANs can generate realistic/diverse 128 128 whole brain Magnetic Resonance (MR) images from noise samples—despite difficult training, such noise-to-image GAN can increase image diversity for further performance boost. Even an expert fails to distinguish the synthetic images from the real ones in Visual Turing Test.
The second project tackles image augmentation for 2D classification. Most CNN architectures adopt around 256 256 input sizes; thus, we use the noise-to-noise GAN, Progressive Growing of GANs (PGGANs), to generate realistic/diverse 256 256 whole brain MR images with/without tumors separately. Multimodal UNsupervised Image-to-image Translation further refines the synthetic images’ texture and shape. Our two-step GAN-based DA boosts sensitivity 93.7% to 97.5% in 2D tumor/non-tumor classification. An expert classifies a few synthetic images as real.
The third project augments images for 2D detection. Further DA applications require pathology localization for detection and advanced physician training needs atypical image generation, respectively. To meet both clinical demands, we propose Conditional PGGANs (CPGGANs) that incorporates highly-rough bounding box conditions incrementally into the noise-to-image GAN (i.e., the PGGANs) to place realistic/diverse brain metastases at desired positions/sizes on MR images; the bounding box-based detection requires much less physicians’ annotation effort than segmentation. Our CPGGAN-based DA boosts sensitivity 83% to 91% in tumor detection with clinically acceptable additional False Positives (FPs). In terms of extrapolation, such pathology-aware GANs are promising because common and/or desired medical priors can play a key role in the conditioning—theoretically, infinite conditioning instances, external to the training data, exist and enforcing such constraints have an extrapolation effect via model reduction.
Finally, we solve image augmentation for 3D detection. Because lesions vary in 3D position/appearance, 3D multiple pathology-aware conditioning is important. Therefore, we propose 3D Multi-Conditional GAN (MCGAN) that translates noise boxes into realistic/diverse lung nodules placed naturally at desired position/size/attenuation on Computed Tomography scans. Our 3D MCGAN-based DA boosts sensitivity in 3D nodule detection under any nodule size/attenuation at fixed FP rates. Considering the realism confirmed by physicians, it could perform as a physician training tool to display realistic medical images with desired abnormalities.
We confirm our pathology-aware GANs’ clinical relevance for diagnosis via two discussions: (i) Conducting a questionnaire survey about our GAN projects for 9 physicians; (ii) Holding a workshop about how to develop medical Artificial Intelligence (AI) fitting into a clinical environment in five years for 7 professionals with various AI and/or Healthcare background.
List of Abbreviations
AI | Artificial Intelligence |
---|---|
BRATS | Brain Tumor Image Segmentation Benchmark |
CAD | Computer-Aided Diagnosis |
CNN | Convolutional Neural Network |
CPGGANs | Conditional Progressive Growing of Generative Adversarial Networks |
CPM | Competition Performance Metric |
CT | Computed Tomography |
DA | Data Augmentation |
DCGAN | Deep Convolutional Generative Adversarial Network |
FLAIR | FLuid Attenuation Inversion Recovery |
FP | False Positive |
FROC | Free Receiver Operation Characteristic |
GAN | Generative Adversarial Network |
HGG | High-Grade Glioma |
IoU | Intersection over Union |
JS | Jensen-Shannon |
LGG | Low-Grade Glioma |
LSGANs | Least Squares Generative Adversarial Networks |
MCGAN | Multi-Conditional Generative Adversarial Network |
MRI | Magnetic Resonance Imaging |
MUNIT | Multimodal UNsupervised Image-to-image Translation |
PGGANs | Progressive Growing of Generative Adversarial Networks |
ReLU | Rectified Linear Unit |
ROI | Region Of Interest |
---|---|
SGD | Stochastic Gradient Descent |
SimGAN | Simulated and unsupervised learning Generative Adversarial Network |
t-SNE | t-distributed Stochastic Neighbor Embedding |
T1 | T1-weighted |
T1c | contrast enhanced T1-weighted |
T2 | T2-weighted |
UNIT | UNsupervised Image-to-image Translation |
VAE | Variational AutoEncoder |
VOI | Voxel Of Interest |
WGAN | Wasserstein Generative Adversarial Network |
WGAN-GP | Wasserstein Generative Adversarial Network with Gradient Penalty |
Chapter 1 Introduction
“Life is short, and the Art long; the occasion fleeting; experience fallacious, and judgment difficult. The physician must not only be prepared to do what is right himself, but also to make the patient, the attendants, and externals cooperate.”
Hippocrates [460-375 BC]
1.1 Aims and Motivations
Convolutional Neural Networks (CNNs) have revolutionized Medical Image Analysis by extracting valuable insights for better clinical examination and medical intervention; the CNNs occasionally outperformed even expert physicians in diagnostic accuracy when large-scale annotated datasets were available [1, 2]. However, obtaining such massive datasets often involves the following intrinsic challenges [3, 4]: (i) it is costly and laborious to collect medical images, such as Magnetic Resonance (MR) and Computed Tomography (CT) images, especially for rare disease; (ii) it is time-consuming and observer-dependent, even for expert physicians, to annotate them due to the low pathological-to-healthy ratio. To tackle these issues, researchers have mainly focused on extracting as much information as possible from the available limited data [5, 6]. Instead, Generative Adversarial Networks (GANs) [7] can generate realistic but completely new samples via many-to-many mappings, and thus effectively cover the real image distribution; they showed great promise in Data Augmentation (DA) using natural images, such as 21% performance improvement in eye-gaze estimation [8].
Interpolation refers to new data point construction within a discretely-sampled data distribution. In terms of the interpolation, the GAN-based image augmentation is reliable on the medical images because medical modalites (e.g., X-ray, CT, MRI) can display the human body’s strong anatomical consistency at fixed position while clearly reflecting inter-subject variability [9, 10]—this is different from the natural images, where various objects can appear at any position; accordingly, to tackle large inter-subject, inter-pathology, and cross-modality variability [3, 4], we propose to use noise-to-image GANs (e.g., random noise samples to diverse pathological images) for (i) medical DA and (ii) physician training [11]. The noise-to-image GAN training is much more difficult than training image-to-image GANs (e.g., a benign image to a malignant one); but, it can perform more global regularization (i.e., adding constraints when fitting a loss function on a training set to prevent overfitting) and increase image diversity for further performance boost.
Regarding the DA, the GAN-generated images can improve Computer-Aided Diagnosis (CAD) based on supervised learning [12]. For the physician training, the GANs can display novel desired pathological images and help train medical trainees despite infrastructural and legal constraints [13]. However, we cannot directly use conventional GANs for realistic/diverse high-resolution medical image augmentation. Moreover, we have to find effective loss functions and training schemes for each of those applications [14]; the diversity matters more for the DA to sufficiently fill the real image distribution whereas the realism matters more for the physician training not to confuse the medical students and radiology trainees.
So, how can we perform clinically relevant GAN-based DA/physician training using only limited annotated training images? Always in collaboration with physicians, for improving 2D classification, we combine the noise-to-image [15, 16] (i.e., Progressive Growing of GANs, PGGANs [17]) and image-to-image GANs (i.e., Multimodal UNsupervised Image-to-image Translation, MUNIT [18]); the two-step GAN can generate and refine realistic/diverse original-sized brain MR images with/without tumors separately. Nevertheless, further DA applications require pathology localization for detection (i.e., identifying target pathology positions in medical images) and advanced physician training needs atypical image generation, respectively. To meet both clinical demands, we propose novel 2D/3D bounding box-based GANs conditioned on pathology position/size/appearance; the bounding box-based detection requires much less physicians’ annotation effort than rigorous segmentation.
Extrapolation refers to new data point estimation beyond a discretely-sampled data distribution. While it is not mutually-exclusive with the interpolation and both rely on a model’s restoring force, it is more subject to uncertainty and thus a risk of meaningless data generation. In terms of the extrapolation, the pathology-aware GANs (i.e., the conditional GANs controlling pathology, such as tumors and nodules, based on position/size/appearance) are promising because common and/or desired medical priors can play a key role in the conditioning—theoretically, infinite conditioning instances, external to the training data, exist and enforcing such constraints have an extrapolation effect via model reduction [19]; inevitable errors, not limited between two data points, caused by the model reduction forces a generator to synthesize images that the generator has never synthesized before.
For improving 2D detection, we propose Conditional PGGANs (CPGGANs) that incorporates highly-rough bounding box conditions incrementally into the noise-to-image GAN (i.e., the PGGANs) to place realistic/diverse brain metastases at desired positions/sizes on MR images [20]. As its pathology-aware conditioning, we use 2D tumor position/size on MR images. Since lesions vary in 3D position/appearance, for improving 3D detection, we propose 3D Multi-Conditional GAN (MCGAN) that translates noise boxes into realistic/diverse lung nodules placed naturally at desired position/size/attenuation on CT scans [21]; inputting the noise box with the surrounding tissues has the effect of combining the noise-to-image and image-to-image GANs. As its pathology-aware conditioning, we use 3D nodule position/size/attenuation on CT scans.
Lastly, we confirm our pathology-aware GANs’ clinical relevance for diagnosis as a clinical decision support system and non-expert physician training tool via two discussions: (i) Conducting a questionnaire survey about our GAN projects for 9 physicians; (ii) Holding a workshop about how to develop medical Artificial Intelligence (AI) fitting into a clinical environment in five years for 7 professionals with various AI and/or Healthcare background.
Contributions. Our main contributions are as follows:
-
•
Noise-to-Image GAN Applications: We propose clinically-valuable novel noise-to-image GAN applications, medical DA and physician training, focusing on their ability to generate realistic and diverse images.
-
•
Pathology-Aware GANs: For required extrapolation, always in collaboration with physicians, we propose novel 2D/3D GANs controlling pathology (i.e., tumors and nodules) on most major modalities (i.e., brain MRI and lung CT).
-
•
Clinical Validation: After detailed discussions with many physicians and professionals with various AI and/or Healthcare background, we confirm our pathology-aware GANs’ clinical relevance as a (i) clinical decision support system and (ii) non-expert physician training tool.
1.2 Thesis Overview
This Ph.D. thesis aims to present the clinical relevance of our novel pathology-aware GAN applications, medical DA and physician training, always in collaboration with physicians.
The thesis is organized as follows (Fig. 1.1). Chapter 2 covers the background of Medical Image Analysis and Deep Learning, as well as methods to address data paucity to bridge them. Chapter 3 describes related work on the GAN-based medical DA and physician training, which emerged after our proposal to use noise-to-image GANs for those applications in Chapter 4. Chapter 5 presents a two-step GAN for 2D classification that combines both noise-to-image and image-to-image GANs. Chapter 6 proposes CPGGANs for 2D detection that incorporates highly-rough bounding box conditions incrementally into the noise-to-image GAN. Finally, we propose 3D MCGAN for 3D detection that translates noise boxes into desired pathology in Chapter 7. Chapter 8 discusses both our pathology-aware GANs’ clinical relevance via a questionnaire survey and how to develop medical AI fitting into a clinical environment in five years via a workshop. Lastly, Chapter 9 provides the conclusive remarks and future directions for further GAN-based extrapolation.
Chapter 2 Background
This chapter introduces basic concepts in Medical Image Analysis and Deep Learning. Afterwards, we describe methods to address data paucity because they play the greatest role in bridging the Medical Image Analysis and Deep Learning.
2.1 Medical Image Analysis
Medical Image Analysis refers to the process of increasing clinical examination/medical intervention efficiency, based on several imaging modalities and digital image processing techniques [22, 23]; to effectively visualize the human body’s anatomical and physiological features, it covers various modalities including X-ray, CT, MRI, positron emission tomography, endoscopy, optical coherence tomography, pathology, ultrasound imaging, and fundus imaging. Its tasks are mainly classified into three groups: (i) Early detection/diagnosis/prognosis of disease often based on pathology classification/detection/segmentation and survival prediction [24, 25]; (ii) Clinical workflow enhancement often based on body part segmentation, inter-modality registration, 3D reconstruction, flow measurement, and surgery simulation [26, 27]; (iii) Clinically impossible image analysis, such as radiogenomics that identifies the correlation between cancer imaging features and gene expression [28].
Among the various modalities, this thesis focuses on the most common 3D modalities for non-invasive diagnosis, CT and MRI. To get a detailed picture inside the body, the CT merges multiple X-rays at different angles using computational tomographic reconstruction [9, 29]. Since X-ray intensity is associated with the mass attenuation coefficient, higher-density tissues show higher attenuation and vice versa. Accordingly, each voxel possesses its attenuation value following the Hounsfield scale from to (e.g., Hounsfield units for air, 0 for water, and for dense bone). The CT can provide a outstanding contrast within soft-tissue, bone, and lung while the soft-tissue contrast is poor—accordingly, it is especially performed for comprehensive lung assessment.
The MRI uses magnetization properties of atomic nuclei [10, 30]. Since different tissues show various relaxation processes when the nuclei return to their resting alignment, the tissues’ proton density maps serve as both anatomical and functional images. Since the tissues possess two different relaxation times, T1 (i.e., longitudinal relaxation time) and T2 (i.e., transverse relaxation time), as MRI sequences, we can obtain both T1-weighted (T1) and T2-weighted (T2) images. Moreover, using very long repetition time and time to echo, we can obtain FLuid Attenuation Inversion Recovery (FLAIR) images. The MRI can provide a superior soft-tissue contrast to the CT—accordingly, it is especially performed for comprehensive brain assessment.
2.2 Deep Learning
Deep Learning is a kind of Machine Learning algorithms, based on Artificial Neural Networks [31]. The Deep Neural Networks consist of many linearly connected non-linear units whose parameters are optimized by gradient descent [32]; accordingly, their multiple layers can gradually grasp more-detailed features as training progresses (i.e., learning which features to place is automatic). Thanks to the good generalization ability, under large-scale data, the Deep Learning significantly outperforms classical Machine Learning algorithms relying on feature engineering. A visual cortex includes arrangements of simple and complex cells activated by a receptive field (i.e., subregions of a visual field); inspired by this biological structure [33], CNNs adopt a mathematical operation called convolution to achieve translation invariance [34]. Since the CNNs are excellent at image/video recognition, their diverse medical applications include pathology classification/detection/segmentation and survival prediction [24, 25].
Variational AutoEncoders (VAEs) often suffer from blurred samples despite easier training, due to the imperfect reconstruction using a single objective function [35]; meanwhile, GANs have revolutionized image generation in terms of realism and diversity [36], including denoising [37] and MRI-to-CT translation [38], based on a two-player objective function using two CNNs [7]: a generator tries to generate realistic images to fool a discriminator while maintaining diversity; attempts to distinguish between the real and synthetic images. However, difficult GAN training from the two-player objective function accompanies artifacts and mode collapse [39], when generating high-resolution images (e.g., pixels) [40]; to tackle this, multi-stage noise-to-image GANs have been proposed: AttnGAN generated images from text using attention-based multi-stage refinement [41]; PGGANs generated realistic images using low-to-high resolution multi-stage training [17].
Contrarily, to obtain images with desired texture and shape, some researchers have proposed image-to-image GANs: MUNIT translated images using both GANs and VAEs [18]; Simulated and unsupervised learning GAN (SimGAN) translated images for DA using the self-regularization term and local adversarial loss [8]; Isola et al. proposed Pix2Pix GAN to produce robust images using paired training samples [42]. Others have proposed conditional GANs: Reed et al. proposed bounding box-based conditional GAN to control generated images’ local properties [43]; Park et al. proposed multi-conditional GAN to refine base images based on texts describing desired position [44].
In Healthcare, medical images have generated the largest volume of data and this trend will no doubt increase due to equipment improvement [45, 46]. Accordingly, as the Deep Learning dominates Computer Vision, Medical Image Analysis is not an exception; their combination can analyze the large-scale medical images and extract valuable insights for better clinical examination and medical intervention. However, the biggest challenge to bridge them lies in the difficulty of obtaining desired pathological images, especially for rare disease [3, 4]. Moreover, it is time-consuming and observer-dependent, even for expert physicians, to annotate them.
2.3 Methods to Address Data Paucity
So, how can we tackle the data paucity? We can either attempt to (a) overcome the lack of generalization or (b) overcome difficulties in optimization. The most straightforward and effective way to address the generalization is DA [47, 48]; because the best model when given data is uncertain, we commonly increase training set size. Human perception is invariant to size, shape, brightness, and color [49]. Accordingly, we recognize the same objects while their such features change, and thus intentionally changing the features is plausible to obtain more data. Such classical DA include (i) x/y/z-axis flipping and rotating, (ii) zooming and scaling, (iii) cropping, (iv) translating, (v) elastic deformation, (vi) adding Gaussian noise (i.e, the distortion of high frequency features), and (vii) brightness and contrast fluctuation.
Recent DA techniques focus on regularization: Mixup [50] and Between-class learning [51] mixed two images during training, such as a dog image and a cat one, for regularization; Cutout randomly masked out square regions during training for regularization [52]; CutMix combined the Mixup and Cutout [53]. As a recent impressive DA approach, AutoAugment automatically searched for improved DA policies [54]. Moreover, similarly to the Mixup among all images within the same class, GAN-based DA can fill the uncovered real image distribution by generating realistic and diverse images via many-to-many mapping [55].
Along with the DA, researchers proposed many other techniques to improve the generalization: semi-supervised learning can considerably increase accuracy under limited labeled data by using pseudo labels for unlabeled data [5]; unsupervised anomaly detection allows to detect out-of-distribution images from normal ones, such as disease, without any labeled data [56]; regularization techniques, such as dropout [57], Lasso [58], and elastic net [59], are commonly used for reducing overfitting; similarly, ensembling multiple models via bagging [60] and boosting [61] can effectively increase the robustness; Lastly, in Medical Image Analysis, we can fuse multiple image modalities and/or sequences, such as MRI CT [62] and T1 MRI T2 MRI [63].
Moreover, many techniques exist for overcoming the difficulties in optimization: transfer learning can achieve better parameter initialization [64]; problem reduction, such as inputting 2D/3D image patches instead of a whole image, can eliminate unnecessary parameters [65]; learning methods with less data, such as zero-shot learning [66], one-shot learning [6], and neural Turing machine [67], are also promising; meta-learning promotes a versatile model applicable to various tasks without requiring multiple training from scratch [68].
Chapter 3 Investigated Contexts and Applications
In terms of interpolation, GAN-based medical image augmentation is reliable because medical modalities (e.g., X-ray, CT, MRI) can display the human body’s strong anatomical consistency at fixed position while clearly reflecting inter-subject variability [9, 10]—this is different from natural images, where various objects can appear at any position. Accordingly, we proposed to use noise-to-image GANs for (i) medical DA and (ii) physician training [11] in Chapter 4. Since then, research towards such clinically valuable applications has shown great promise. This chapter covers such related research works except our own works [15, 16, 20, 21] included in Chapters 5-7. Involving 9 physicians, we discuss in detail the clinical relevance of the GAN-based medical DA and physician training [69] in Chapter 8.
3.1 GAN-based Medical DA
Because the lack of annotated pathological images is the greatest challenge in CAD [3, 4], to handle various types of small/fragmented datasets from multiple scanners, researchers have actively conducted GAN-based DA studies especially in Medical Image Analysis. For better classification, some researchers adopted image-to-image GANs similarly to their conventional medical applications, such as denoising [37] and MRI-to-CT translation [38]: Wu et al. translated normal mammograms into lesion ones [70], Gupta et al. translated normal leg X-ray images into bone lesion ones [71], and Malygina et al. translated / normal chest X-ray images into pneumonia/pleural-thickening ones [72]. Meanwhile, others adopted the noise-to-image GANs as we proposed, to increase image diversity for further performance boost—the diversity matters more for the DA to sufficiently fill the real image distribution: Frid-Adar et al. augmented liver lesion CT images [12], Madani et al. augmented chest X-ray images with cardiovascular abnormality [73], and Konidaris et al. augmented brain MR images with Alzheimer’s disease [74].
To facilitate pathology detection and segmentation, researchers conditioned the image-to-image GANs, not the noise-to-image GANs like our work in Chapter 6, with pathology features (e.g., position, size, and appearance) and generated realistic/diverse pathology at desired positions in medical images. In terms of extrapolation, the pathology-aware GANs are promising because common and/or desired medical priors can play a key role in the conditioning—theoretically, infinite conditioning instances, external to the training data, exist and enforcing such constraints have an extrapolation effect via model reduction [19]. To the best of our knowledge, only Kanayama et al. tackled bounding box-based pathology detection using the image-to-image GAN [75]; they translated normal endoscopic images with various image sizes ( on average) into gastric cancer ones by inputting both a benign image and a black image (i.e., pixel value: 0) with a specific lesion Region Of Interest (ROI) at desired position. Without conditioning the noise-to-image GAN with nodule position, Gao et al. generated 3D nodule subvolumes only applicable to their subvolume-based detector using binary classification [76].
Since 3D imaging is spreading in radiology (e.g., CT and MRI), most GAN-based DA works for segmentation exploited 3D conditional image-to-image GANs. However, 3D medical image generation is more challenging than 2D one due to expensive computational cost and strong anatomical consistency; so, instead of generating a whole image including pathology, researchers only focused on a malignant Voxel Of Interest (VOI): Shin et al. translated normal brain MR images into tumor ones by inputting both a benign image and a tumor-conditioning image [77], similarly to the Kanayama et al.’s work [75]; Jin et al. generated CT images of lung nodules including the surrounding tissues by only inputting a VOI centered at a lung nodule, but with a central sphere region erased [78]. Recently, instead of generating realistic images and training classifiers on them separately, Chaitanya et al. directly optimized segmentation results on cardiac MR images [26]; however, it segmented body parts, instead of pathology. Since effective GAN-based medical DA generally requires much engineering effort, we also published a tutorial journal paper [14] about tricks to boost classification/detection/segmentation performance using the GANs, based on our experience and related work.
3.2 GAN-based Physician Training
While medical students and radiology trainees must view thousands of images to become competent [79], accessing such abundant medical images is often challenging due to infrastructural and legal constraints [80]. Because pathology-aware GANs can generate novel medical images with desired abnormalities (e.g., position, size, and appearance)—while maintaining enough realism not to confuse the medical trainees—GAN-based physician training concept is drawing attention: Chuquicusma et al. appreciated the GAN potential to train radiologists for educational purpose after successfully generating CT images of lung nodules that even deceived experts [81]; thanks to their anonymization ability, Shin et al. proposed to share pathology-aware GAN-generated images outside institutions after achieving considerable tumor segmentation results with only synthetic MR images for training [77]; more importantly, Finlayson et al. from Harvard Medical School are currently validating a class-conditional GANs’ radiology educational efficacy after succeeding in learning features that distinguish fractures from non-fractures on pelvic X-ray images [13].
Chapter 4 GAN-based Medical Image Generation
4.1 Prologue to First Project
4.1.1 Project Publication
-
•
GAN-based Synthetic Brain MR Image Generation. C. Han, H. Hayashi, L. Rundo, R. Araki, Y. Furukawa, W. Shimoda, S. Muramatsu, G. Mauri, H. Nakayama, In IEEE International Symposium on Biomedical Imaging (ISBI), pp. 734–738, April 2018.
4.1.2 Context
Prior to this work, it remained challenging to generate realistic and diverse medical images using noise-to-image GANs, not image-to-image GANs [37], due to their unstable training. GAN architectures well-suited for medical images were unclear. Yi et al. published results on the noise-to-image GAN-based brain MR image generation, proposing its potential for medical DA and physician training while our paper was under submission [82]; however, they only generated single-sequence low-resolution brain MR images without tumors.
4.1.3 Contributions
This project’s main contribution is to propose to use recently developed Wasserstein Generative Adversarial Network (WGAN) [83] for medical DA and physician training—the medical GAN applications are reliable in terms of interpolation because medical modalities can display the human body’s strong anatomical consistency at fixed position while clearly reflecting inter-subject variability. We also demonstrate the noise-to-image GAN’s such potential by generating multi-sequence realistic/diverse whole brain tumor MR images [83]; then, we confirm the superb realism via Visual Turing Test by a physician.
4.1.4 Recent Developments
Since proposing the GAN applications, we have successfully applied the noise-to-image GANs to improve 2D tumor classification/detection on brain MR images [15, 16, 20] as described in Chapters 5 and 6. For better 3D tumor segmentation, Shin et al. have translated normal brain MR images into tumor ones using the image-to-image GAN [77]. Finlayson et al. have generated pelvic fracture/non-fracture X-ray images using a class-conditional noise-to-image GAN, also introducing ongoing work on validating such GANs’ radiology educational efficacy [13]. Kwon et al. have generated realistic/diverse 3D brain MR images using the noise-to-image GAN [84].
4.2 Motivation
Along with classic methods [85], CNNs have recently revolutionized medical image analysis [86], including brain MRI segmentation [87]. However, CNN training demands extensive medical data that are laborious to obtain [88]. To overcome this issue, DA techniques via reconstructing original images are common for better performance, such as geometry and intensity transformations [89, 90].
However, those reconstructed images intrinsically resemble the original ones, leading to limited performance improvement in terms of generalization abilities; thus, generating realistic (similar to the real image distribution) but completely new images is essential. In this context, GAN-based DA has excellently performed in general computer vision tasks. It attributes to GAN’s good generalization ability from matching the noise-generated distribution to the real one with a sharp value function. Especially, Shrivastava et al. (SimGAN) outperformed the state-of-the-art with a relative 21% improvement in eye-gaze estimation [8].
So, how can we generate realistic medical images completely different from the original samples? Our aim is to generate synthetic multi-sequence brain MR images using GANs, which is essential in medical imaging to increase diagnostic reliability, such as via DA in CAD as well as physician training (Fig. 4.1) [91]. However, this is extremely challenging—MR images are characterized by low contrast, strong visual consistency in brain anatomy, and intra-sequence variability. Our novel GAN-based approach for medical DA adopts Deep Convolutional Generative Adversarial Network (DCGAN) [40] and WGAN [83] to generate realistic images, and an expert physician validates them via Visual Turing Test [92].
Research Questions. We mainly address two questions:
-
•
GAN Selection: Which GAN architecture is well-suited for realistic medical image generation?
-
•
Medical Image Processing: How can we handle MR images with specific intra-sequence variability?
Contributions. Our main contributions are as follows:
-
•
MR Image Generation: This research shows that WGAN can generate realistic multi-sequence brain MR images, possibly leading to valuable clinical applications: DA and physician training.
-
•
Medical Image Generation: This research provides how to exploit medical images with intrinsic intra-sequence variability towards GAN-based DA for medical imaging.
4.3 Materials and Methods
Towards clinical applications utilizing realistic brain MR images, we generate synthetic brain MR images from the original samples using GANs. Here, we compare the most used two GANs, DCGAN and WGAN, to find a well-suited GAN between them for medical image generation—it must avoid mode collapse and generate realistic MR images with high resolution.
4.3.1 BRATS 2016 Dataset
This project exploits a dataset of multi-sequence brain MR images to train GANs with sufficient data and resolution, which was originally produced for the Multimodal Brain Tumor Image Segmentation Benchmark (BRATS) Challenge [93]. In particular, the BRATS 2016 training dataset contains 220 High-Grade Glioma (HGG) and 54 Low-Grade Glioma (LGG) cases, with T1-weighted (T1), contrast enhanced T1-weighted (T1c), T2-weighted, and FLAIR sequences—they were skull stripped and resampled to isotropic resolution with voxels; among the different sectional planes, we use sagittal multi-sequence scans of the HGG patients to show that our GANs can generate a complete view of the whole brain anatomy (allowing for visual consistency among the different brain lobes), including also severe tumors for clinical purpose.
4.3.2 DCGAN/WGAN-based Image Generation
Pre-processing We select the slices from to among the whole slices to omit initial/final slices, since they convey a negligible amount of useful information and could affect the training. The images are resized to both / pixels from for better GAN training (DCGAN architecture results in stable training on 64 64 pixels [40], and so is reasonably a high-resolution). Fig. 4.2 shows example real MR images used for training; each sequence contains 15,400 images with 220 patients 70 slices (61,600 in total).
MR Image Generation DCGAN and WGAN generate six types of images as follows:
-
•
T1 sequence () from the real T1;
-
•
T1c sequence () from the real T1c;
-
•
T2 sequence () from the real T2;
-
•
FLAIR sequence () from the real FLAIR;
-
•
Concat sequence () from concatenating the real T1, T1c, T2, and FLAIR (i.e., feeding the model with samples from all the MRI sequences);
-
•
Concat sequence () from concatenating the real T1, T1c, T2, and FLAIR.
Concat sequence refers to a new ensemble sequence for an alternative DA, containing features of all four sequences. We also generate Concat images to compare the generation performance in terms of image resolution.
DCGAN [40] is a standard GAN [7] with a convolutional architecture for unsupervised learning; this generative model uses up-convolutions interleaved with Rectified Lineaer Unit (ReLU) non-linearity and batch normalization.
Let be a generating distribution over data . The generator is a mapping to data space that takes a prior on input noise variables , where is a neural network with parameters . Similarly, the discriminator is a neural network with parameters that takes either real data or synthetic data and outputs a single scalar probability that came from the real data. The discriminator maximizes the probability of classifying both training examples and samples from correctly while the generator minimizes the likelihood; it is formulated as a minimax two-player game with value function :
(4.1) |
This can be reformulated as the minimization of the Jensen-Shannon (JS) divergence between the distribution and another distribution derived from and .
DCGAN Implementation Details We use the same DCGAN architecture [40] with no in the generator, ELU as the discriminator, all filters of size , and a half channel size for DCGAN training. A batch size of and Adam optimizer with learning rate were implemented.
WGAN [83] is an alternative to traditional GAN training, as the JS divergence is limited, such as when it is discontinuous; this novel GAN achieves stable learning with less mode collapse by replacing it to the Earth Mover (EM) distance (i.e., the Wasserstein-1 metrics):
(4.2) |
where is the set of all joint distributions whose marginals are and , respectively. In other words, implies how much mass must be transported from one distribution to another. This distance intuitively indicates the cost of the optimal transport plan.
WGAN Implementation Details We use the same DCGAN architecture [40] for WGAN training. A batch size of 64 and Root Mean Square Propagation (RMSprop) optimizer with learning rate were implemented.
4.3.3 Clinical Validation via Visual Turing Test
To quantitatively evaluate how realistic the synthetic images are, an expert physician was asked to constantly classify a random selection of real/ synthetic MR images as real or synthetic shown in random order for each GAN/sequence, without previous training stages revealing which is real/synthetic; Concat images were classified together with real T1, T1c, T2, and FLAIR images in equal proportion. The so-called Visual Turing Test [92] uses binary questions to probe a human ability to identify attributes and relationships in images. For these motivations, it is commonly used to evaluate GAN-generated images, such as for SimGAN [8]. This applies also to medical images in clinical environments [81], wherein physicians’ expertise is critical.
4.4 Results
This section shows how DCGAN and WGAN generate synthetic brain MR images. The results include instances of synthetic images and their quantitative evaluation of the realism by an expert physician.
4.4.1 MR Images Generated by DCGAN/WGAN
DCGAN Fig. 4.3 illustrates examples of synthetic images by DCGAN. The images look similar to the real samples. Concat images combine appearances and patterns from all the four sequences used in training. Since DCGAN’s value function could be unstable, it often generates hyper-intense T1-like images analogous to mode collapse for 64 64 Concat images, while sharing the same hyper-parameters with 128 128.
WGAN Fig. 4.4 shows the example output of WGAN in each sequence. Remarkably outperforming DCGAN, WGAN successfully captures the sequence-specific texture and tumor appearance while maintaining the realism of the original brain MR images. As expected, Concat images tend to have more messy and unrealistic artifacts than Concat ones, especially around boundaries of the brain, due to the introduction of unexpected intensity patterns.
4.4.2 Visual Turing Test Results
Table 4.1 shows the confusion matrix concerning the Visual Turing Test. Even the expert physician found classifying real and synthetic images challenging, especially in lower resolution due to their less detailed appearances unfamiliar in clinical routine, even for highly hyper-intense Concat images by DCGAN; distinguishing Concat images was easier compared to the case of T1, T1c, T2, and FLAIR images because the physician often felt odd from the artificial sequence. WGAN succeeded to deceive the physician significantly better than DCGAN for all the MRI sequences except FLAIR images ( to ).
Accuracy (%) | Real as Real (%) | Real as Synt (%) | Synt as Real (%) | Synt as Synt (%) | |
T1 (DCGAN, ) | 70 | 52 | 48 | 12 | 88 |
T1c (DCGAN, ) | 71 | 48 | 52 | 6 | 94 |
T2 (DCGAN, ) | 64 | 44 | 56 | 16 | 84 |
FLAIR (DCGAN, ) | 54 | 24 | 76 | 16 | 84 |
Concat (DCGAN, ) | 77 | 68 | 32 | 14 | 86 |
Concat (DCGAN, ) | 54 | 26 | 74 | 18 | 82 |
T1 (WGAN, ) | 64 | 40 | 60 | 12 | 88 |
T1c (WGAN, ) | 55 | 26 | 74 | 16 | 84 |
T2 (WGAN, ) | 58 | 38 | 62 | 22 | 78 |
FLAIR (WGAN, ) | 62 | 32 | 68 | 8 | 92 |
Concat (WGAN, ) | 66 | 62 | 38 | 30 | 70 |
Concat (WGAN, ) | 53 | 36 | 64 | 30 | 70 |
4.5 Conclusion
Our preliminary results show that GANs, especially WGAN, can generate realistic multi-sequence brain MR images that even an expert physician is unable to accurately distinguish from the real, leading to valuable clinical applications, such as DA and physician training. This attributes to WGAN’s good generalization ability with a sharp value function. In this context, DCGAN might be unsuitable due to both inferior realism and mode collapse in terms of intensity. We only use slices of interest in training to obtain desired MR images and generate both original/Concat sequence images for DA in medical imaging.
This study confirms the synthetic image quality by the human expert evaluation, but a more objective computational evaluation for GANs should also follow, such as Classifier Two-Sample Tests (C2ST) [94], which assesses whether two samples are drawn from the same distribution. Currently this work uses sagittal MR images alone, so we plan to generate coronal and transverse images. As this research uniformly selects middle slices in pre-processing, better data generation demands developing a classifier to only select brain MRI slices with/without tumors.
Towards DA, whereas realistic images give more insights on geometry/intensity transformations in classification, more realistic images do not always assure better DA, so we have to find suitable image resolutions and sequences; that is why we generate both high-resolution images and Concat images, yet they looked more unrealistic for the physician. For physician training, generating desired realistic tumors by adding conditioning requires exploring latent spaces of GANs extensively.
Overall, our novel GAN-based realistic brain MR image generation approach sheds light on diagnostic and prognostic medical applications; future studies on these applications are needed to confirm our encouraging results.
Chapter 5 GAN-based Medical Image Augmentation for 2D Classification
5.1 Prologue to Second Project
5.1.1 Project Publications
-
•
Infinite Brain MR Images: PGGAN-based Data Augmentation for Tumor Detection. C. Han, L. Rundo, R. Araki, Y. Furukawa, G. Mauri, H. Nakayama, H. Hayashi, In A. Esposito, M. Faundez-Zanuy, F. C. Morabito, E. Pasero (eds.) Neural Approaches to Dynamics of Signal Exchanges, Springer, pp. 291–303, September 2019.
-
•
Combining Noise-to-Image and Image-to-Image GANs: Brain MR Image Augmentation for Tumor Detection. C. Han, L. Rundo, R. Araki, Y. Nagano, Y. Furukawa, G. Mauri, H. Nakayama, H. Hayashi, IEEE Access, pp. 156966–156977, October 2019.
5.1.2 Context
At the time we wrote the former paper, high-resolution (e.g., 256 256) medical image generation using noise-to-image GANs had been challenging [73] while most CNN architectures adopt around 256 256 input sizes (e.g., InceptionResNetV2 [95]: , ResNet-50 [96]: ). Moreover, prior to the latter paper, analysis had been immature on GAN-generated additional training images for better CNN-based classification.
5.1.3 Contributions
This project’s core contribution is to firstly combine noise-to-image and image-to-image GANs for improved 2D classification. The former paper adopts a noise-to-image GAN called PGGANs to generate realistic/diverse original-sized 256 256 whole brain MR images with/without tumors separately; additionally, the latter paper exploits an image-to-image GAN called MUNIT to further refine the synthetic images’ texture and shape similarly to real ones. By so doing, our two-step GAN-based DA boosts sensitivity 93.7% to 97.5% in tumor/non-tumor classification. Moreover, we firstly analyze how medical GAN-based DA is associated with pre-training on ImageNet and discarding weird-looking synthetic images to humans to achieve high sensitivity. A physician classifies a few synthetic images as real in Visual Turing Test despite the high resolution.
5.1.4 Recent Developments
Since the former paper’s acceptance (the book chapter’s publication process took more than a year), to improve 2D classification, Konidaris et al. generated brain MR images with Alzheimer’s disease using the noise-to-image GAN [74]. No more recent developments to report exist for the latter paper because it is very recent.
5.2 Motivation
CNNs are playing a key role in Medical Image Analysis, updating the state-of-the-art in many tasks [87, 97, 98] when large-scale annotated training data are available. However, preparing such massive medical data is demanding; thus, for better diagnosis, researchers generally adopt classic DA techniques, such as geometric/intensity transformations of original images [89, 90]. Those augmented images, however, intrinsically have a similar distribution to the original ones, resulting in limited performance improvement. In this sense, GAN-based DA can considerably increase the performance [7]; since the generated images are realistic but completely novel samples, they can relieve the sampling biases and fill the real image distribution uncovered by the original dataset [99].
The main problem in CAD lies in small/fragmented medical imaging datasets from multiple scanners; thus, researchers have improved classification by augmenting images with noise-to-image GANs [11] or image-to-image GANs [70]. However, no research has achieved further performance boost by combining noise-to-image and image-to-image GANs.
So, how can we maximize the DA effect under limited training images using the GAN combinations? To generate and refine brain MR images with/without tumors separately (Fig. 5.1), we propose a two-step GAN-based DA approach: (i) PGGANs [17], low-to-high resolution noise-to-image GAN, first generates realistic/diverse images—the PGGANs helps DA since most CNN architectures adopt around input sizes (e.g., InceptionResNetV2 [95]: , ResNet-50 [96]: ); (ii) MUNIT [18] that combines GANs/VAEs [35] or SimGAN [8] that uses a DA-focused GAN loss, further refines the texture and shape of the PGGAN-generated images to fit them into the real image distribution. Since training a single sophisticated GAN system is already difficult, instead of end-to-end training, we adopt a two-step approach for performance boost via an ensemble generation process from those state-of-the-art GANs’ different algorithms.
We thoroughly investigate CNN-based tumor classification results, also considering the influence of pre-training on ImageNet [34] and discarding weird-looking GAN-generated images. Moreover, we evaluate the synthetic images’ appearance via Visual Turing Test [92] by an expert physician, and visualize the data distribution of real/synthetic images via t-Distributed Stochastic Neighbor Embedding (t-SNE) [100]. When combined with classic DA, our two-step GAN-based DA approach significantly outperforms the classic DA alone, boosting sensitivity to .
Research Questions. We mainly address two questions:
-
•
GAN Selection: Which GAN architectures are well-suited for realistic/diverse medical image generation?
-
•
Medical DA: How to use GAN-generated images as additional training data for better CNN-based diagnosis?
Contributions. Our main contributions are as follows:
-
•
Whole Image Generation: This research shows that PGGANs can generate realistic/diverse whole medical images—not only small pathological sub-areas—and MUNIT can further refine their texture and shape similarly to real ones.
-
•
Two-step GAN-based DA: This novel two-step approach, combining for the first time noise-to-image and image-to-image GANs, significantly boosts tumor classification sensitivity.
-
•
Misdiagnosis Prevention: This study firstly analyzes how medical GAN-based DA is associated with pre-training on ImageNet and discarding weird-looking synthetic images to achieve high sensitivity with small and fragmented datasets.
5.3 Materials and Methods
5.3.1 BRATS 2016 Dataset
We use a dataset of T1c brain axial MR images of HGG cases from BRATS 2016 [93]. T1c is the most common sequence in tumor classification thanks to its high-contrast [101].
5.3.2 PGGAN-based Image Generation
Pre-processing For better GAN/ResNet-50 training, we select the slices from to among the whole slices to omit initial/final slices, which convey negligible useful information; also, since tumor/non-tumor annotation in the BRATS 2016 dataset, based on 3D volumes, is highly incorrect/ambiguous on 2D slices, we exclude () tumor images tagged as non-tumor, () non-tumor images tagged as tumor, () borderline images with unclear tumor/non-tumor appearance, and () images with missing brain parts due to the skull-stripping procedure. For tumor classification, we divide the whole dataset ( patients) into:
-
•
Training set
( patients/ tumor/ non-tumor images); -
•
Validation set
( patients/ tumor/ non-tumor images); -
•
Test set
( patients/ tumor/ non-tumor images).
During the GAN training, we only use the training set to be fair; for better PGGAN training, the training set images are zero-padded to reach a power of : pixels from . Fig. 5.2 shows example real MR images.
PGGANs [17] is a GAN training method that progressively grows a generator and discriminator: starting from low resolution, new layers model details as training progresses. This study adopts the PGGANs to synthesize realistic/diverse brain MR images (Fig. 5.3); we train and generate tumor/non-tumor images separately.
PGGAN Implementation Details The PGGAN architecture adopts the Wasserstein loss with Gradient Penalty (WGAN-GP) [39]:
(5.1) |
where denotes the expected value, the discriminator (i.e., the set of -Lipschitz functions), is the data distribution defined by the true data sample , and is the model distribution defined by the generated sample ( is the input noise to the generator sampled from a Gaussian distribution). A gradient penalty is added for the random sample , where is the gradient operator towards the generated samples and is the gradient penalty coefficient.
We train the model (Table 5.1) for epochs with a batch size of and learning rate for the Adam optimizer (the exponential decay rates ) [102]. All experiments use with critic iteration per generator iteration. During training, we apply random cropping in - pixels as DA.
Generator | Activation | Output Shape |
---|---|---|
Latent vector | – | |
Conv | LReLU | |
Conv | LReLU | |
Upsample | – | |
Conv | LReLU | |
Conv | LReLU | |
Upsample | – | |
Conv | LReLU | |
Conv | LReLU | |
Upsample | – | |
Conv | LReLU | |
Conv | LReLU | |
Upsample | – | |
Conv | LReLU | |
Conv | LReLU | |
Upsample | – | |
Conv | LReLU | |
Conv | LReLU | |
Upsample | – | |
Conv | LReLU | |
Conv | LReLU | |
Conv | Linear | |
Discriminator | Activation | Output Shape |
---|---|---|
Input image | – | |
Conv | LReLU | |
Conv | LReLU | |
Conv | LReLU | |
Downsample | – | |
Conv | LReLU | |
Conv | LReLU | |
Downsample | – | |
Conv | LReLU | |
Conv | LReLU | |
Downsample | – | |
Conv | LReLU | |
Conv | LReLU | |
Downsample | – | |
Conv | LReLU | |
Conv | LReLU | |
Downsample | – | |
Conv | LReLU | |
Conv | LReLU | |
Downsample | – | |
Minibatch stddev | – | |
Conv | LReLU | |
Conv | LReLU | |
Fully-connected | Linear | |
5.3.3 MUNIT/SimGAN-based Image Refinement
Refinement Using resized images for ResNet-50, we further refine the texture and shape of PGGAN-generated tumor/non-tumor images separately to fit them into the real image distribution using MUNIT [18] or SimGAN [8]. SimGAN remarkably improved eye gaze estimation results after refining non-GAN-based synthetic images from the UnityEyes simulator image-to-image translation; thus, we also expect such performance improvement after refining synthetic images from a noise-to-image GAN (i.e., PGGANs) an image-to-image GAN (i.e., MUNIT/SimGAN) with considerably different GAN algorithms.
We randomly select real/ PGGAN-generated tumor images for tumor image training, and we perform the same for non-tumor image training. To find suitable refining steps for each architecture, we pick the MUNIT/SimGAN models with the highest accuracy on tumor classification validation, when pre-trained and combined with classic DA, among // steps, respectively.
MUNIT [18] is an image-to-image GAN based on both auto-encoding/translation; it extends UNIT [104] to increase the generated images’ realism/diversity via a stochastic model representing continuous output distributions.
MUNIT Implementation Details The MUNIT architecture adopts the following loss:
(5.2) | |||||
where denotes the loss function. Using the multiple encoders /, generators /, discriminators /, cycle-consistencies CC1/CC2, and domain-invariant perceptions VGG1/VGG2 [105], this framework jointly solves learning problems of the VAE1/VAE2 and GAN1/GAN2 for the image reconstruction streams, image translation streams, cycle-consistency reconstruction streams, and domain-invariant perception streams. Since we do not need the style loss for our experiments, instead of the MUNIT loss, we use the UNIT loss with the perceptual loss for the MUNIT architecture (as in the UNIT authors’ GitHub repository). The MUNIT architecture adopts the following loss:
(5.3) |
We train the model (Table 5.2) for steps with a batch size of 1 and learning rate for the Adam optimizer () [102]. The learning rate is reduced by half every steps. We use the following MUNIT weights: the adversarial loss weight ; the image reconstruction loss weight ; the Kullback-Leibler (KL) divergence loss weight for reconstruction ; the cycle consistency loss weight ; the KL divergence loss weight for cycle consistency ; the domain-invariant perceptual loss weight ; the Least Squares GAN objective function for the discriminators [106]. During training, we apply horizontal flipping as DA.
Generator | Activation | Output Shape |
Content Encoder | ||
Input image | – | |
Conv | ReLU | |
Conv | ReLU | |
Conv | ReLU | |
ResBlock4 | ReLU | |
– | ||
Decoder | ||
ResBlock4 | ReLU | |
– | ||
Upsample | – | |
Conv | ReLU | |
Upsample | – | |
Conv | ReLU | |
Conv | Tanh | |
Discriminator | Activation | Output Shape |
---|---|---|
Input image | – | |
Conv | LReLU | |
Conv | LReLU | |
Conv | LReLU | |
Conv | LReLU | |
Conv | – | |
AveragePool | – | |
Conv | LReLU | |
Conv | LReLU | |
Conv | LReLU | |
Conv | LReLU | |
Conv | – | |
AveragePool | – | |
Conv | LReLU | |
Conv | LReLU | |
Conv | LReLU | |
Conv | LReLU | |
Conv | – | |
AveragePool | – | |
SimGAN [8] is an image-to-image GAN designed for DA that adopts the self-regularization term/local adversarial loss; it updates a discriminator with a history of refined images.
SimGAN Implementation Details The SimGAN architecture (i.e., a refiner) uses the following loss:
(5.4) |
where denotes the loss function, is the function parameters, is the PGGAN-generated training image, and is the set of the real images . The first part adds realism to the synthetic images using a discriminator, while the second part preserves the tumor/non-tumor features.
We train the model (Table 5.3) for steps with a batch size of 10 and learning rate for the Stochastic Gradient Descent (SGD) optimizer [109] without momentum. The learning rate is reduced by half at 15,000 steps. We train the refiner first with just the self-regularization loss with for steps; then, for each update of the discriminator, we update the refiner times. During training, we apply horizontal flipping as DA.
Refiner | Activation | Output Shape |
---|---|---|
Input image | – | |
Conv | ReLU | |
ResBlock12 | ReLU | |
– | ||
Conv | Tanh | |
Discriminator | Activation | Output Shape |
---|---|---|
Input image | – | |
Conv | ReLU | |
Conv | ReLU | |
Maxpool | – | |
Conv | ReLU | |
Conv | ReLU | |
Maxpool | – | |
Conv | ReLU | |
Conv | ReLU | |
5.3.4 ResNet-50-based Tumor Classification
Pre-processing As ResNet-50’s input size is pixels, we resize the whole real images from and whole PGGAN-generated images from .
ResNet-50 [96] is a -layer residual learning-based CNN. We adopt it to conduct tumor/non-tumor binary classification on MR images due to its outstanding performance in image classification tasks [110], including binary classification [24]. Chang et al.[111] also used a similar -layer residual convolutional network for the binary classification of brain tumors (i.e., determining the Isocitrate Dehydrogenase status in LGG/HGG).
DA Setups To confirm the effect of PGGAN-based DA and its refinement using MUNIT/SimGAN, we compare the following DA setups under sufficient images both with/without ImageNet [34] pre-training (i.e., 20 DA setups):
-
1.
real images;
-
2.
+ k classic DA;
-
3.
+ k classic DA;
-
4.
+ k PGGAN-based DA;
-
5.
+ k PGGAN-based DA w/o clustering/discarding;
-
6.
+ k classic DA & k PGGAN-based DA;
-
7.
+ k MUNIT-refined DA;
-
8.
+ k classic DA & k MUNIT-refined DA;
-
9.
+ k SimGAN-refined DA;
-
10.
+ k classic DA & k SimGAN-refined DA.
Due to the risk of overlooking the tumor diagnosis, higher sensitivity matters much more than higher specificity [112]; thus, we aim to achieve higher sensitivity, using the additional synthetic training images. We perform McNemar’s test on paired tumor classification results [113] to confirm our two-step GAN-based DA’s statistically-significant sensitivity improvement; since this statistical analysis involves multiple comparison tests, we adjust their -values using the Holm-Bonferroni method [114].
Whereas medical imaging researchers widely use the ImageNet initialization despite different textures of natural/medical images, recent study found that such ImageNet-trained CNNs are biased towards recognizing texture rather than shape [115]; thus, we aim to investigate how the medical GAN-based DA affects classification performance with/without the pre-training. As the classic DA, we adopt a random combination of horizontal/vertical flipping, rotation up to degrees, width/height shift up to , shearing up to , zooming up to , and constant filling of points outside the input boundaries (Fig. 5.4). For the PGGAN-based DA and its refinement, we only use success cases after discarding weird-looking synthetic images (Fig. 5.5); DenseNet-169 [116] extracts image features and k-means++ [117] clusters the features into groups, and then we manually discard each cluster containing similar weird-looking images. To verify its effect, we also conduct a PGGAN-based DA experiment without the discarding step. Additionally, to confirm the effect of changing training data set sizes, we compare classification results with pre-training on /// real images vs real images + k classic DA vs real images + k classic DA & k PGGAN-based DA (i.e., setups).
ResNet-50 Implementation Details The ResNet-50 architecture adopts the binary cross-entropy loss for binary classification both with/without ImageNet pre-training. As shown in Table 5.4, for robust training, before the final sigmoid layer, we introduce a dropout [57], linear dense, and batch normalization [118] layers—training with GAN-based DA tends to be unstable especially without the batch normalization layer. We use a batch size of , learning rate for the SGD optimizer [109] with momentum, and early stopping of epochs. The learning rate was multiplied by every epochs for the training from scratch and by every epochs for the ImageNet pre-training.
Classifier | Activation | Output Shape |
---|---|---|
Input image | – | |
Conv | ReLU | |
Maxpool | – | |
ResBlock3 | ReLU | |
ReLU | ||
ReLU | ||
ResBlock4 | ReLU | |
ReLU | ||
ReLU | ||
ResBlock6 | ReLU | |
ReLU | ||
ReLU | ||
ResBlock3 | ReLU | |
ReLU | ||
ReLU | ||
AveragePool | – | |
Flatten | – | 2048 |
Dropout | – | 2048 |
Dense | – | 2 |
BatchNorm | Sigmoid | 2 |
5.3.5 Clinical Validation via Visual Turing Test
To quantify the (i) realism of synthetic images by PGGANs, MUNIT, and SimGAN against real images respectively (i.e., 3 setups) and (ii) clearness of their tumor/non-tumor features, we supply, in random order, to an expert physician a random selection of:
-
•
real tumor images;
-
•
real non-tumor images;
-
•
synthetic tumor images;
-
•
synthetic non-tumor images.
Then, the physician is asked to classify them as both (i) real/synthetic and (ii) tumor/non-tumor, without previously knowing which is real/synthetic and tumor/non-tumor.
5.3.6 Visualization via t-SNE
To visualize distributions of geometrically-transformed and each GAN-based images by PGGANs, MUNIT, and SimGAN against real images respectively (i.e., 4 setups), we adopt t-SNE [100] on a random selection of:
-
•
real tumor images;
-
•
real non-tumor images;
-
•
geometrically-transformed or each GAN-based tumor images;
-
•
geometrically-transformed or each GAN-based non-tumor images.
We select only images per each category for better visualization. The t-SNE method reduces the dimensionality to represent high-dimensional data into a lower-dimensional (2D/3D) space; it non-linearly balances between the input data’s local and global aspects using perplexity.
T-SNE Implementation Details The t-SNE uses a perplexity of for iterations to visually represent a 2D space. We input the images after normalizing pixel values to . For point locations of the real images, we compress all the images simultaneously and plot each setup (i.e., the geometrically-transformed or each GAN-based images against the real ones) separately; we maintain their locations by projecting all the data onto the same subspace.
5.4 Results
This section shows how PGGANs generates synthetic brain MR images and how MUNIT and SimGAN refine them. The results include instances of synthetic images, their quantitative evaluation by a physician, their t-SNE visualization, and their influence on tumor classification.
5.4.1 MR Images Generated by PGGANs
Fig. 5.5 illustrates examples of synthetic MR images by PGGANs. We visually confirm that, for about of cases, it successfully captures the T1c-specific texture and tumor appearance, while maintaining the realism of the original brain MR images; but, for the rest , the generated images lack clear tumor/non-tumor features or contain unrealistic features (i.e., hyper-intensity, gray contours, and odd artifacts).
5.4.2 MR Images Refined by MUNIT/SimGAN
MUNIT and SimGAN differently refine PGGAN-generated images—they render the texture and contours while maintaining the overall shape (Fig. 5.6). Non-tumor images change more remarkably than tumor images for both MUNIT and SimGAN; it probably derives from unsupervised image translation’s loss for consistency to avoid image collapse, resulting in conservative change for more complicated images.
5.4.3 Tumor Classification Results
Table 5.5 shows the brain tumor classification results with/without DA while Table 5.6 indicates their pairwise comparison (-values between our two-step GAN-based DA setups and the other DA setups) using McNemar’s test. ImageNet pre-training generally outperforms training from scratch despite different image domains (i.e., natural images to medical images). As expected, classic DA remarkably improves classification, while no clear difference exists between the / classic DA under sufficient geometrically-transformed training images. When pre-trained, each GAN-based DA (i.e., PGGANs/MUNIT/SimGAN) alone helps classification due to the robustness from GAN-generated images; but, without pre-training, it harms classification due to the biased initialization from the GAN-overwhelming data distribution. Similarly, without pre-training, PGGAN-based DA without clustering/discarding causes poor classification due to the synthetic images with severe artifacts, unlike the PGGAN-based DA’s comparable results with/without the discarding step when pre-trained.
DA Setups | Accuracy (%) | Sensitivity (%) | Specificity (%) | |
---|---|---|---|---|
(1) 8,429 real images | 93.1 (86.3) | 90.9 (88.9) | 95.9 (83.2) | |
(2) + 200k classic DA | 95.0 (92.2) | 93.7 (89.9) | 96.6 (95.0) | |
(3) + 400k classic DA | 94.8 (93.2) | 91.9 (90.9) | 98.4 (96.1) | |
(4) + 200k PGGAN-based DA | 93.9 (86.2) | 92.6 (87.3) | 95.6 (84.9) | |
(5) + 200k PGGAN-based DA w/o clustering/discarding | 94.8 (80.7) | 91.9 (80.2) | 98.4 (81.2) | |
(6) + 200k classic DA & 200k PGGAN-based DA | 96.2 (95.6) | 94.0 (94.2) | 98.8 (97.3) | |
(7) + 200k MUNIT-refined DA | 94.3 (83.7) | 93.0 (87.8) | 95.8 (78.5) | |
(8) + 200k classic DA & 200k MUNIT-refined DA | 96.7 (96.3) | 95.4 (97.5) | 98.2 (95.0) | |
(9) + 200k SimGAN-refined DA | 94.5 (77.6) | 92.3 (82.3) | 97.1 (72.0) | |
(10) + 200k classic DA & 200k SimGAN-refined DA | 96.4 (95.0) | 95.1 (95.1) | 97.9 (95.0) | |
DA Setup Comparison | Accu | Sens | Spec | DA Setup Comparison | Accu | Sens | Spec | DA Setup Comparison | Accu | Sens | Spec | |
(7) w/ PT vs (1) w/ PT | 0.693 | 0.206 | 1 | (7) w/ PT vs (1) w/o PT | 0.001 | 0.002 | 0.001 | (7) w/ PT vs (2) w/ PT | 1 | 1 | 1 | |
(7) w/ PT vs (2) w/o PT | 0.034 | 0.024 | 1 | (7) w/ PT vs (3) w/ PT | 1 | 1 | 0.035 | (7) w/ PT vs (3) w/o PT | 1 | 0.468 | 1 | |
(7) w/ PT vs (4) w/ PT | 1 | 1 | 1 | (7) w/ PT vs (4) w/o PT | 0.001 | 0.001 | 0.001 | (7) w/ PT vs (5) w/ PT | 1 | 1 | 0.003 | |
(7) w/ PT vs (5) w/o PT | 0.001 | 0.001 | 0.001 | (7) w/ PT vs (6) w/ PT | 0.009 | 1 | 0.001 | (7) w/ PT vs (6) w/o PT | 0.397 | 1 | 1 | |
(7) w/ PT vs (7) w/o PT | 0.001 | 0.001 | 0.001 | (7) w/ PT vs (8) w/ PT | 0.001 | 0.025 | 0.045 | (7) w/ PT vs (8) w/o PT | 0.008 | 0.001 | 1 | |
(7) w/ PT vs (9) w/ PT | 1 | 1 | 1 | (7) w/ PT vs (9) w/o PT | 0.001 | 0.001 | 0.001 | (7) w/ PT vs (10) w/ PT | 0.001 | 0.077 | 0.108 | |
(7) w/ PT vs (10) w/o PT | 1 | 0.206 | 1 | (7) w/o PT vs (1) w/ PT | 0.001 | 0.135 | 0.001 | (7) w/o PT vs (1) w/o PT | 0.026 | 1 | 0.014 | |
(7) w/o PT vs (2) w/ PT | 0.001 | 0.001 | 0.001 | (7) w/o PT vs (2) w/o PT | 0.001 | 1 | 0.001 | (7) w/o PT vs (3) w/ PT | 0.001 | 0.020 | 0.001 | |
(7) w/o PT vs (3) w/o PT | 0.001 | 0.147 | 0.001 | (7) w/o PT vs (4) w/ PT | 0.001 | 0.002 | 0.001 | (7) w/o PT vs (4) w/o PT | 0.044 | 1 | 0.001 | |
(7) w/o PT vs (5) w/ PT | 0.001 | 0.015 | 0.001 | (7) w/o PT vs (5) w/o PT | 0.011 | 0.001 | 1 | (7) w/o PT vs (6) w/ PT | 0.001 | 0.001 | 0.001 | |
(7) w/o PT vs (6) w/o PT | 0.001 | 0.001 | 0.001 | (7) w/o PT vs (8) w/ PT | 0.001 | 0.001 | 0.001 | (7) w/o PT vs (8) w/o PT | 0.001 | 0.001 | 0.001 | |
(7) w/o PT vs (9) w/ PT | 0.001 | 0.004 | 0.001 | (7) w/o PT vs (9) w/o PT | 0.001 | 0.001 | 0.001 | (7) w/o PT vs (10) w/ PT | 0.001 | 0.001 | 0.001 | |
(7) w/o PT vs (10) w/o PT | 0.001 | 0.001 | 0.001 | (8) w/ PT vs (1) w PT | 0.001 | 0.001 | 0.010 | (8) w/ PT vs (1) w/o PT | 0.001 | 0.001 | 0.001 | |
(8) w/ PT vs (2) w/ PT | 0.001 | 0.074 | 0.206 | (8) w/ PT vs (2) w/o PT | 0.001 | 0.001 | 0.001 | (8) w/ PT vs (3) w/ PT | 0.002 | 0.001 | 1 | |
(8) w/ PT vs (3) w/o PT | 0.001 | 0.001 | 0.112 | (8) w/ PT vs (4) w/ PT | 0.001 | 0.001 | 0.006 | (8) w/ PT vs (4) w/o PT | 0.001 | 0.001 | 0.001 | |
(8) w/ PT vs (5) w/ PT | 0.002 | 0.001 | 1 | (8) w/ PT vs (5) w/o PT | 0.001 | 0.001 | 0.001 | (8) w/ PT vs (6) w/ PT | 1 | 0.128 | 1 | |
(8) w/ PT vs (6) w/o PT | 0.222 | 0.760 | 1 | (8) w/ PT vs (8) w/o PT | 1 | 0.008 | 0.001 | (8) w/ PT vs (9) w/ PT | 0.001 | 0.001 | 1 | |
(8) w/ PT vs (9) w/o PT | 0.001 | 0.001 | 0.001 | (8) w/ PT vs (10) w/ PT | 1 | 1 | 1 | (8) w/ PT vs (10) w/o PT | 0.007 | 1 | 0 | |
(8) w/o PT vs (1) w/ PT | 0.001 | 0.001 | 1 | (8) w/o PT vs (1) w/o PT | 0.001 | 0.001 | 0.001 | (8) w/o PT vs (2) w/ PT | 0.179 | 0.001 | 0.588 | |
(8) w/o PT vs (2) w/o PT | 0.001 | 0.001 | 1 | (8) w/o PT vs (3) w/ PT | 0.101 | 0.001 | 0.001 | (8) w/o PT vs (3) w/o PT | 0.001 | 0.001 | 1 | |
(8) w/o PT vs (4) w/ PT | 0.001 | 0.001 | 1 | (8) w/o PT vs (4) w/o PT | 0.001 | 0.001 | 0.001 | (8) w/o PT vs (5) w/ PT | 0.197 | 0.001 | 0.001 | |
(8) w/o PT vs (5) w/o PT | 0.001 | 0.001 | 0.001 | (8) w/o PT vs (6) w/ PT | 1 | 0.001 | 0.001 | (8) w/o PT vs (6) w/o PT | 1 | 0.001 | 0.007 | |
(8) w/o PT vs (9) w/ PT | 0.023 | 0.001 | 0.256 | (8) w/o PT vs (9) w/o PT | 0.001 | 0.001 | 0.001 | (8) w/o PT vs (10) w/ PT | 1 | 0.002 | 0.001 | |
(8) w/o PT vs (10) w/o PT | 0.143 | 0.005 | 1 | (9) w/ PT vs (1) w/ PT | 0.387 | 1 | 1 | (9) w/ PT vs (1) w/o PT | 0.001 | 0.046 | 0.001 | |
(9) w/ PT vs (2) w/ PT | 1 | 1 | 1 | (9) w/ PT vs (2) w/o PT | 0.008 | 0.262 | 0.321 | (9) w/ PT vs (3) w/ PT | 1 | 1 | 0.931 | |
(9) w/ PT vs (3) w/o PT | 0.910 | 1 | 1 | (9) w/ PT vs (4) w/ PT | 1 | 1 | 0.764 | (9) w/ PT vs (4) w/o PT | 0.001 | 0.001 | 0.001 | |
(9) w/ PT vs (5) w/ PT | 1 | 1 | 0.639 | (9) w/ PT vs (5) w/o PT | 0.001 | 0.001 | 0.001 | (9) w/ PT vs (6) w/ PT | 0.014 | 0.660 | 0.066 | |
(9) w/ PT vs (6) w/o PT | 0.716 | 0.365 | 1 | (9) w/ PT vs (9) w/o PT | 0.001 | 0.001 | 0.001 | (9) w/ PT vs (10) w/ PT | 0.004 | 0.006 | 1 | |
(9) w/ PT vs (10) w/o PT | 1 | 0.017 | 0.256 | (9) w/o PT vs (1) w/ PT | 0.001 | 0.001 | 0.001 | (9) w/o PT vs (1) w/o PT | 0.001 | 0.001 | 0.001 | |
(9) w/o PT vs (2) w/ PT | 0.001 | 0.001 | 0.001 | (9) w/o PT vs (2) w/o PT | 0.001 | 0.001 | 0.001 | (9) w/o PT vs (3) w/ PT | 0.001 | 0.001 | 0.001 | |
(9) w/o PT vs (3) w/o PT | 0.001 | 0.001 | 0.001 | (9) w/o PT vs (4) w/ PT | 0.001 | 0.001 | 0.001 | (9) w/o PT vs (4) w/o PT | 0.001 | 0.001 | 0.001 | |
(9) w/o PT vs (5) w/ PT | 0.001 | 0.001 | 0.001 | (9) w/o PT vs (5) w/o PT | 0.022 | 1 | 0.001 | (9) w/o PT vs (6) w/ PT | 0.001 | 0.001 | 0.001 | |
(9) w/o PT vs (6) w/o PT | 0.001 | 0.001 | 0.001 | (9) w/o PT vs (10) w/PT | 0.001 | 0.001 | 0.001 | (9) w/o PT vs (10) w/o PT | 0.001 | 0.001 | 0.001 | |
(10) w/ PT vs (1) w/ PT | 0.001 | 0.001 | 0.049 | (10) w/ PT vs (1) w/o PT | 0.001 | 0.001 | 0.001 | (10) w/ PT vs (2) w/ PT | 0.039 | 0.515 | 1 | |
(10) w/ PT vs (2) w/o PT | 0.001 | 0.001 | 0.002 | (10) w/ PT vs (3) w/ PT | 0.017 | 0.001 | 1 | (10) w/ PT vs (3) w/o PT | 0.001 | 0.001 | 0.415 | |
(10) w/ PT vs (4) w/ PT | 0.001 | 0.019 | 0.028 | (10) w/ PT vs (4) w/o PT | 0.001 | 0.001 | 0.001 | (10) w/ PT vs (5) w/ PT | 0.015 | 0.001 | 1 | |
(10) w/ PT vs (5) w/o PT | 0.001 | 0.001 | 0.001 | (10) w/ PT vs (6) w/ PT | 1 | 1 | 1 | (10) w/ PT vs (6) w/o PT | 0.981 | 1 | 1 | |
(10) w/ PT vs (10) w/o PT | 0.054 | 1 | 0.002 | (10) w/o PT vs (1) w/ PT | 0.039 | 0.001 | 1 | (10) w/o PT vs (1) w/o PT | 0.001 | 0.001 | 0.001 | |
(10) w/o PT vs (2) w/ PT | 1 | 0.727 | 0.649 | (10) w/o PT vs (2) w/o PT | 0.001 | 0.001 | 1 | (10) w/o PT vs (3) w/ PT | 1 | 0.002 | 0.001 | |
(10) w/o PT vs (3) w/o PT | 0.039 | 0.001 | 1 | (10) w/o PT vs (4) w/ PT | 1 | 0.019 | 1 | (10) w/o PT vs (4) w/o PT | 0.001 | 0.001 | 0.001 | |
(10) w/o PT vs (5) w/ PT | 1 | 0.002 | 0.001 | (10) w/o PT vs (5) w/o PT | 0.001 | 0.001 | 0.001 | (10) w/o PT vs (6) w/ PT | 0.308 | 1 | 0.001 | |
(10) w/o PT vs (6) w/o PT | 1 | 1 | 0.035 | |||||||||
When combined with the classic DA, each GAN-based DA remarkably outperforms the GAN-based DA or classic DA alone in terms of sensitivity since they are mutually-complementary: the former learns the non-linear manifold of the real images to generate novel local tumor features (since we train tumor/non-tumor images separately) strongly associated with sensitivity; the latter learns the geometrically-transformed manifold of the real images to cover global features and provide the robustness on training for most cases. We confirm that test samples, originally-misclassified but correctly classified after DA, are obviously different for the GAN-based DA and classic DA; here, both image-to-image GAN-based DA, especially MUNIT, produce remarkably higher sensitivity than the PGGAN-based DA after refinement. Specificity is higher than sensitivity for every DA setup with pre-training, probably due to the training data imbalance; but interestingly, without pre-training, sensitivity is higher than specificity for both image-to-image GAN-based DA since our tumor classification-oriented two-step GAN-based DA can fill the real tumor image distribution uncovered by the original dataset under no ImageNet initialization. Accordingly, when combined with the classic DA, the MUNIT-based DA based on both GANs/VAEs achieves the highest sensitivity against the best performing classic DA’s , allowing to significantly alleviate the risk of overlooking the tumor diagnosis; in terms of sensitivity, it outperforms all the other DA setups, including two-step DA setups, with statistical significance.
Figure 5.7 shows that the PGGAN-based DA, even without further refinement, could moderately increase both accuracy/sensitivity on top of the classic DA in tumor classification; it achieves considerably high sensitivity with only / of the real training images. However, it should be noted that the MUNIT-based DA could outperform the PGGAN-based DA in return for more computational power.
5.4.4 Visual Turing Test Results
Table 5.7 indicates the confusion matrix for the Visual Turing Test. The expert physician classifies a few PGGAN-generated images as real, thanks to their realism, despite high resolution (i.e., pixels); meanwhile, the expert classifies less GAN-refined images as real due to slight artifacts induced during refinement. The synthetic images successfully capture tumor/non-tumor features; unlike the non-tumor images, the expert recognizes a considerable number of the mild/modest tumor images as non-tumor for both real/synthetic cases. It derives from clinical tumor diagnosis relying on a full 3D volume, instead of a single 2D slice.
Accuracy (%) | Accuracy (%) | Accuracy (%) | Accuracy (%) | Accuracy (%) | |
PGGAN | Real vs Synthetic | as | as | as | as |
Tumor vs Non-tumor | as | as | as | as | |
(, ) | () | ||||
MUNIT | Real vs Synthetic | as | as | as | as |
Tumor vs Non-tumor | as | as | as | as | |
(, ) | (, ) | ||||
SimGAN | Real vs Synthetic | as | as | as | as |
Tumor vs Non-tumor | as | as | as | as | |
(, ) | () | ||||
5.4.5 T-SNE Results
As Fig. 5.8 represents, the real tumor/non-tumor image distributions largely overlap while the non-tumor images distribute wider. The geometrically-transformed tumor/non-tumor image distributions also often overlap, and both images distribute wider than the real ones. All GAN-based synthetic images by PGGANs, MUNIT, and SimGAN distribute widely, while their tumor/non-tumor images overlap much less than the geometrically-transformed ones (i.e., a high discrimination ability associated with sensitivity improvement); the MUNIT-refined images show better tumor/non-tumor discrimination and a more similar distribution to the real ones than the PGGAN-based and SimGAN-based images. This trend derives from the MUNIT’s loss function adopting both GANs/VAEs that further fits the PGGAN-generated images into the real image distribution by refining their texture and shape; contrarily, this refinement could also induce slight human-recognizable but DA-irrelevant artifacts. Overall, the GAN-based images, especially the MUNIT-refined images, fill the distribution uncovered by the real or geometrically-transformed ones with less tumor/non-tumor overlap; this demonstrates the superiority of combining classic DA and GAN-based DA.
5.5 Conclusion
Visual Turing Test and t-SNE results show that PGGANs, multi-stage noise-to-image GAN, can generate realistic/diverse brain MR images with/without tumors separately. Unlike classic DA that geometrically covers global features and provides the robustness on training for most cases, the GAN-generated images can non-linearly cover local tumor features with much less tumor/non-tumor overlap; thus, combining them can significantly boost tumor classification sensitivity—especially after refining them with MUNIT or SimGAN, image-to-image GANs; thanks to an ensemble generation process from those GANs’ different algorithms, the texture/shape-refined images can replace missing data points of the training set with less tumor/non-tumor overlap, and thus handle the data imbalance by regularizing the model (i.e., improved generalization). Notably, MUNIT remarkably outperforms SimGAN in terms of sensitivity, probably due to the effect of combining both GANs/VAEs.
Regarding better medical GAN-based DA, ImageNet pre-training generally improves classification despite different textures of natural/medical images; but, without pre-training, the GAN-refined images may help achieve better sensitivity, allowing to alleviate the risk of overlooking the tumor diagnosis—this attributes to our tumor classification-oriented two-step GAN-based DA’s high discrimination ability to fill the real tumor image distribution under no ImageNet initialization. GAN-generated images typically include odd artifacts; however, only without pre-training, discarding them boosts DA performance.
Overall, by minimizing the number of annotated images required for medical imaging tasks, the two-step GAN-based DA can shed light not only on classification, but also on object detection [20] and segmentation [77]. Moreover, other potential medical applications exist: (i) A data anonymization tool to share patients’ data outside their institution for training without losing classification performance [77]; (ii) A physician training tool to show random pathological images for medical students/radiology trainees despite infrastructural/legal constraints [13]. As future work, we plan to define a new end-to-end GAN loss function that explicitly optimizes the classification results, instead of optimizing visual realism while maintaining diversity by combining the state-of-the-art noise-to-image and image-to-image GANs; towards this, we might extend a preliminary work on a three-player GAN for classification [119] to generate only hard-to-classify samples to improve classification; we could also (i) explicitly model deformation fields/intensity transformations and (ii) leverage unlabeled data during the generative process [26] to effectively fill the real image distribution.
Chapter 6 GAN-based Medical Image Augmentation for 2D Detection
6.1 Prologue to Third Project
6.1.1 Project Publication
-
•
Learning More with Less: Conditional PGGAN-based Data Augmentation for Brain Metastases Detection Using Highly-Rough Annotation on MR Images. C. Han, K. Murao, T. Noguchi, Y. Kawata, F. Uchiyama, L. Rundo, H. Nakayama, S. Satoh, In ACM International Conference on Information and Knowledge Management (CIKM), Beijing, China, pp. 119–127, November 2019.
6.1.2 Context
Further DA applications require pathology localization for detection and advanced physician training needs atypical image generation, respectively. To meet both clinical demands, developing pathology-aware GANs (i.e., GANs conditioned on pathology position and appearance) is the best solution—the pathology-aware GANs are promising in terms of extrapolation because common and/or desired medical priors can play a key role in the conditioning [19]. However, prior to this work, researchers had focused only on improving segmentation, instead of bounding box-based detection, while the detection requires much less physicians’ annotation effort [77, 78]. Moreover, they had relied on image-to-image GANs, instead of noise-to-image GANs, which sacrifices image diversity due to an input benign image.
6.1.3 Contributions
This project’s fundamental contribution is to propose a novel pathology-aware noise-to-image GAN called CPGGANs for improved 2D bounding box-based detection; it incorporates highly-rough bounding box conditions incrementally into the noise-to-image GAN (i.e., PGGANs) to place realistic/diverse brain metastases at desired positions/sizes on MR images. By so doing, our CPGGAN-based DA boosts sensitivity 83% to 91% with Intersection over Union (IoU) threshold in tumor detection with clinically acceptable additional False Positives (FPs). Moreover, we find that GAN training on additional normal images could increase synthetic images’ realism, including pathology, but decrease DA performance.
6.1.4 Recent Developments
Almost simultaneously, Kanayama et al. also tackle bounding box-based pathology detection using the image-to-image GAN, instead of the noise-to-image GAN [75]; they translated normal endoscopic images with various image sizes ( on average) into gastric cancer ones by inputting both a benign image and a black image (i.e., pixel value: 0) with a specific lesion ROI at desired position.
6.2 Motivation
Accurate CAD with high sensitivity can alleviate the risk of overlooking the diagnosis in a clinical environment. Specifically, CNNs have revolutionized medical imaging, such as diabetic eye disease diagnosis [120], mainly thanks to large-scale annotated training data. However, obtaining such annotated medical big data is demanding; thus, better diagnosis requires intensive DA techniques, such as geometric/intensity transformations of original images [89, 90]. Yet, those augmented images intrinsically have a similar distribution to the original ones, leading to limited performance improvement; in this context, GAN [7]-based DA can boost the performance by filling the real image distribution uncovered by the original dataset, since it generates realistic but completely new samples showing good generalization ability; GANs achieved outstanding performance in computer vision, including performance improvement in eye-gaze estimation [8].
Also in medical imaging, where the primary problem lies in small and fragmented imaging datasets from various scanners [97], GAN-based DA performs effectively: researchers improved classification by augmentation with noise-to-image GANs [12] and segmentation with image-to-image GANs [77, 78]. Such applications include brain MR image generation for tumor/non-tumor classification [15]. Nevertheless, unlike bounding box-based object detection, simple classification cannot locate disease areas and rigorous segmentation requires physicians’ expensive annotation.
So, how can we achieve high sensitivity in diagnosis using GANs with minimum annotation cost, based on highly-rough and inconsistent bounding boxes? We aim to generate GAN-based realistic and diverse brain MR images with brain metastases at desired positions/sizes for accurate CNN-based tumor detection (Fig. 6.1); this is clinically valuable for better diagnosis, prognosis, and treatment, since brain metastases are the most common intra-cranial tumors, getting prevalent as oncological therapies improve cancer patients’ survival [121]. Conventional GANs cannot generate realistic whole brain MR images conditioned on tumor positions/sizes under limited training data/highly-rough annotation [15]; since noise-to-image GANs cannot directly be conditioned on an image describing desired objects, we have to use image-to-image GANs (e.g., input both conditioning image/random noise samples or the conditioning image alone with dropout noises [57] on a generator [42])—it results in unrealistic high-resolution MR images with odd artifacts due to the limited training data/rough annotation, tumor variations, and strong consistency in brain anatomy, unless we also input a benign image sacrificing image diversity.
Such a high-resolution whole image generation approach, not involving ROIs alone, however, could facilitate detection because it provides more image details and most CNN architectures adopt around input pixels. Therefore, as a conditional noise-to-image GAN not relying on an input benign image, we propose CPGGANs, incorporating highly-rough bounding box conditions incrementally into PGGANs [17] to naturally place tumors of random shape at desired positions/sizes on MR images. Moreover, we evaluate the generated images’ realism via Visual Turing Test [92] by three expert physicians, and visualize the data distribution via t-SNE algorithm [100].
Using the synthetic images, our novel CPGGAN-based DA boosts sensitivity in diagnosis with clinically acceptable additional FPs.
Surprisingly, we confirm that further realistic tumor appearance, judged by the physicians, does not contribute to detection performance.
Research Questions. We mainly address two questions:
-
•
PGGAN Conditioning: How can we modify PGGANs to naturally place objects of random shape, unlike rigorous segmentation, at desired positions/sizes based on highly-rough bounding box masks?
-
•
Medical DA: How can we balance the number of real and additional synthetic training data to achieve the best detection performance?
Contributions. Our main contributions are as follows:
-
•
Conditional Image Generation: As the first bounding box-based whole pathological image generation approach, CPGGANs can generate realistic/diverse images with objects naturally at desired positions/sizes; the generated images can play a vital role in clinical oncology applications, such as DA, data anonymization, and physician training.
-
•
Misdiagnosis Prevention: This study allows us to achieve high sensitivity in automatic CAD using small/fragmented medical imaging datasets with minimum annotation efforts based on highly-rough/inconsistent bounding boxes.
-
•
Brain Metastases Detection: This first bounding box-based brain metastases detection method successfully detects tumors with CPGGAN-based DA.
6.3 Materials and Methods
6.3.1 Brain Metastases Dataset
As a new dataset for the first bounding box-based brain metastases detection, this project uses a dataset of T1c brain axial MR images, collected by the authors (National Center for Global Health and Medicine, Tokyo, Japan) and currently not publicly available for ethical restrictions; for robust clinical applications, it contains brain metastatic cancer cases from multiple MRI scanners—those images differ in contrast, magnetic field strength (i.e., T, T), and matrix size (i.e., , , , pixels). We also use additional brain MR images from normal subjects only for CPGGAN training, not in tumor detection, to confirm the effect of combining the normal and pathological images for training.
6.3.2 CPGGAN-based Image Generation
Data Preparation For tumor detection, our whole brain metastases dataset ( patients) is divided into: (i) a training set ( patients); (ii) a validation set ( patients); (iii) a test set ( patients); only the training set is used for GAN training to be fair. Our experimental dataset consists of:
-
•
Training set ( images/ bounding boxes);
-
•
Validation set ( images/ bounding boxes);
-
•
Test set ( images/ bounding boxes).
Our training set is relatively small/fragmented for CNN-based applications, considering that the same patient’s tumor slices could convey very similar information. To confirm the effect of realism and diversity—provided by combining PGGANs and bounding box conditioning—on tumor detection, we compare the following GANs: (i) CPGGANs trained only with the brain metastases images; (ii) CPGGANs trained also with additional brain images from normal subjects; (iii) Image-to-image GAN trained only with the brain metastases images. After skull-stripping on all images with various resolution, remaining brain parts are cropped and resized to pixels (i.e., a power of for better GAN training). As Fig. 6.2 shows, we lazily annotate tumors with highly-rough and inconsistent bounding boxes to minimize expert physicians’ labor.
CPGGANs is a novel conditional noise-to-image training method for GANs, incorporating highly-rough bounding box conditions incrementally into PGGANs [17], unlike conditional image-to-image GANs requiring rigorous segmentation masks [122]. The original PGGANs exploits a progressively growing generator and discriminator: starting from low-resolution, newly-added layers model fine-grained details as training progresses. As Fig. 6.3 shows, we further condition the generator and discriminator to generate realistic and diverse brain MR images with tumors of random shape at desired positions/sizes using only bounding boxes without an input benign image under limited training data/highly-rough annotation. Our modifications to the original PGGANs are as follows:
-
•
Conditioning image: prepare a black image (i.e., pixel value: ) with white bounding boxes (i.e., pixel value: ) describing tumor positions/sizes for attention;
-
•
Generator input: resize the conditioning image to the previous generator’s output resolution/channel size and concatenate them (noise samples generate the first images);
-
•
Discriminator input: concatenate the conditioning image with a real or synthetic image.
CPGGAN Implementation Details We use the CPGGAN architecture with the WGAN-GP loss [39]:
(6.1) |
where the discriminator belongs to the set of -Lipschitz functions, is the data distribution by the true data sample , and is the model distribution by the synthetic sample generated from the conditioning image uniform noise samples in . The last term is gradient penalty for the random sample .
Training lasts for steps with a batch size of and learning rate for the Adam optimizer [102]. We flip the discriminator’s real/synthetic labels once in three times for robustness. During testing, as tumor attention images, we use the annotation of training images with a random combination of horizontal/vertical flipping, width/height shift up to , and zooming up to ; these CPGGAN-generated images are used as additional training images for tumor detection.
Image-to-image GAN is a conventional conditional GAN that generates brain MR images with tumors, concatenating a conditioning image with noise samples for a generator input and concatenating the conditioning image with a real/synthetic image for a discriminator input, respectively. It uses a U-Net-like [89] generator with convolutional/deconvolutional layers in encoders/decoders respectively with skip connections, along with a discriminator with decoders. We apply batch normalization [118] to both convolution with LeakyReLU and deconvolution with ReLU. It follows the same implementation details as for the CPGGANs.
6.3.3 YOLOv3-based Brain Metastases Detection
You Only Look Once v3 (YOLOv3) [123] is a fast/accurate CNN-based object detector: unlike conventional classifier-based detectors, it divides the image into regions and predicts bounding boxes/probabilities for each region. We adopt YOLOv3 to detect brain metastases since its high efficiency can play a clinical role in real-time tumor alert; moreover, it shows very comparable results with network resolution against other state-of-the-art detectors, such as Faster RCNN [124].
To confirm the effect of GAN-based DA, the following detection results are compared: (i) real images without DA, (ii), (iii), (iv) with // CPGGAN-based DA, (v), (vi), (vii) with // CPGGAN-based DA, trained with additional normal brain images, (viii), (ix), (x) with // image-to-image GAN-based DA. Due to the risk of overlooking the diagnosis medical imaging, higher sensitivity matters more than less FPs; thus, we aim to achieve higher sensitivity with a clinically acceptable number of FPs, adding the additional synthetic training images. Since our annotation is highly-rough, we calculate sensitivity/FPs per slice with both IoU threshold 0.5 and 0.25.
YOLOv3 Implementation Details We use the YOLOv3 architecture with Darknet-53 as a backbone classifier and sum squared error between the predictions/ground truth as a loss:
(6.2) |
where are the centroid location of an anchor box, are the width/height of the anchor, is the Objectness (i.e., confidence score of whether an object exists), and is the classification loss. Let and be the size of a feature map and the number of anchor boxes, respectively. is when an object exists in cell and otherwise .
During training, we use a batch size of and learning rate for the Adam optimizer. The network resolution is set to pixels during training and pixels during validation/testing respectively to detect small tumors better. We recalculate the anchors at each DA setup. As classic DA, geometric/intensity transformations are also applied to both real/synthetic images during training to achieve the best performance. For testing, we pick the model with the best sensitivity on validation with detection threshold 0.1%/IoU threshold 0.5 between - steps to avoid severe FPs while achieving high sensitivity.
6.3.4 Clinical Validation via Visual Turing Test
To quantitatively evaluate how realistic the CPGGAN-based synthetic images are, we supply, in random order, to three expert physicians a random selection of real and synthetic brain metastases images. They take four tests in ascending order: (i), (ii) test 1, 2: real vs CPGGAN-generated resized tumor bounding boxes, trained without/with additional normal brain images; (iii), (iv) test 3, 4: real vs CPGGAN-generated MR images, trained without/with additional normal brain images.
Then, the physicians constantly classify them as real/synthetic, if needed, zooming/rotating them, without previous training stages revealing which is real/synthetic.
6.3.5 Visualization via t-SNE
To visually analyze the distribution of real/synthetic images, we use t-SNE [100] on a random selection of:
-
•
real tumor images;
-
•
CPGGAN-generated tumor images;
-
•
CPGGAN-generated tumor images, trained with additional normal brain images.
We normalize the input images to .
t-SNE Implementation Details We use t-SNE with a perplexity of for iterations to get a 2D representation.
6.4 Results
This section shows how CPGGANs and image-to-image GAN generate brain MR images. The results include instances of synthetic images and their influence on tumor detection, along with CPGGAN-generated images’ evaluation via Visual Turing Test and t-SNE.
6.4.1 MR Images Generated by CPGGANs
Fig. 6.4 illustrates example GAN-generated images. CPGGANs successfully captures the T1c-specific texture and tumor appearance at desired positions/sizes. Since we use highly-rough bounding boxes, the synthetic tumor shape largely varies within the boxes. When trained with additional normal brain images, it clearly maintains the realism of the original images with less odd artifacts, including tumor bounding boxes, which the additional images do not include. However, as expected, image-to-image GAN, without progressive growing, generates clearly unrealistic images without an input benign image due to the limited training data/rough annotation.
6.4.2 Brain Metastases Detection Results
Table 6.1 shows the tumor detection results with/without GAN-based DA. As expected, the sensitivity remarkably increases with the additional synthetic training data while FPs per slice also increase. Adding more synthetic images generally leads to a higher amount of FPs, also detecting blood vessels that are small/hyper-intense on T1c MR images, very similarly to the enhanced tumor regions (i.e., the contrast agent perfuses throughout the blood vessels). However, surprisingly, adding only CPGGAN-generated images achieves the best sensitivity improvement by with IoU threshold and by with IoU threshold , probably due to the real/synthetic training image balance—the improved training robustness achieves sensitivity with moderate IoU threshold despite our highly-rough bounding box annotation.
IoU 0.5 | IoU 0.25 | |||
Sensitivity (%) | FPs per slice | Sensitivity (%) | FPs per slice | |
2,813 real images | 67 | 4.11 | 83 | 3.59 |
+ 4,000 CPGGAN-based DA | 77 | 7.64 | 91 | 7.18 |
+ 8,000 CPGGAN-based DA | 71 | 6.36 | 87 | 5.85 |
+ 12,000 CPGGAN-based DA | 76 | 11.77 | 91 | 11.29 |
+ 4,000 CPGGAN-based DA (+ normal) | 69 | 7.16 | 86 | 6.60 |
+ 8,000 CPGGAN-based DA (+ normal) | 73 | 8.10 | 89 | 7.59 |
+ 12,000 CPGGAN-based DA (+ normal) | 74 | 9.42 | 89 | 8.95 |
+ 4,000 Image-to-Image GAN-based DA | 72 | 6.21 | 87 | 5.70 |
+ 8,000 Image-to-Image GAN-based DA | 68 | 3.50 | 84 | 2.99 |
+ 12,000 Image-to-Image GAN-based DA | 74 | 7.20 | 89 | 6.72 |
Fig. 6.5 also visually indicates that it can alleviate the risk of overlooking the tumor diagnosis with clinically acceptable FPs; in the clinical routine, the bounding boxes, highly-overlapping around tumors, only require a physician’s single check by switching on/off transparent alpha-blended annotation on MR images. It should be noted that we cannot increase FPs to achieve such high sensitivity without CPGGAN-based DA. Moreover, our results reveal that further realism—associated with the additional normal brain images during training—does not contribute to detection performance, possibly as the training focuses less on tumor generation. Image-to-image GAN-based DA just moderately facilitates detection with less additional FPs, probably because the synthetic images have a distribution far from the real ones and thus their influence on detection is limited during testing.
Accuracy (%) | Real as Real (%) | Real as Synt (%) | Synt as Real (%) | Synt as Synt (%) | ||
Test 1 | Physician 1 | 88 | 80 | 20 | 4 | 96 |
Physician 2 | 95 | 90 | 10 | 0 | 100 | |
Physician 3 | 97 | 98 | 2 | 4 | 96 | |
Test 2 | Physician 1 | 81 | 78 | 22 | 16 | 84 |
Physician 2 | 83 | 86 | 14 | 20 | 80 | |
Physician 3 | 91 | 90 | 10 | 8 | 92 | |
Test 3 | Physician 1 | 97 | 94 | 6 | 0 | 100 |
Physician 2 | 96 | 92 | 8 | 0 | 100 | |
Physician 3 | 100 | 100 | 0 | 0 | 100 | |
Test 4 | Physician 1 | 91 | 82 | 18 | 0 | 100 |
Physician 2 | 96 | 96 | 4 | 4 | 96 | |
Physician 3 | 100 | 100 | 0 | 0 | 100 | |
6.4.3 Visual Turing Test Results
Table 6.2 shows the confusion matrix for the Visual Turing Test. The expert physicians easily recognize synthetic images due to the lack of training data. However, when CPGGANs is trained with additional normal brain images, the experts classify a considerable number of synthetic tumor bounding boxes as real; it implies that the additional normal images remarkably facilitate the realism of both healthy and pathological brain parts while they do not include abnormality; thus, CPGGANs might perform as a tool to train medical students and radiology trainees when enough medical images are unavailable, such as abnormalities at rare positions/sizes. Such GAN applications are clinically prospective [13], considering the expert physicians’ positive comments about the tumor realism.
6.4.4 T-SNE Results
As presented in Fig. 6.6, synthetic tumor bounding boxes have a moderately similar distribution to real ones, but they also fill the real image distribution uncovered by the original dataset, implying their effective DA performance; especially, the CPGGAN-generated images trained without normal brain images distribute wider than the center-concentrating images trained with the normal brain images. Meanwhile, real/synthetic whole brain images clearly distribute differently, due to the real MR images’ strong anatomical consistency (Fig. 6.7). Considering the achieved high DA performance, the tumor (i.e., ROI) realism/diversity matter more than the whole image realism/diversity, since YOLOv3 look at an image patch instead of a whole image, similarly to most other CNN-based object detectors.
6.5 Conclusion
Without relying on an input benign image, our CPGGANs can generate realistic and diverse MR images with brain metastases of random shape, unlike rigorous segmentation, naturally at desired positions/sizes, and achieve high sensitivity in tumor detection—even with small/fragmented training data from multiple MRI scanners and lazy annotation using highly-rough bounding boxes; in the context of intelligent data wrangling, this attributes to the CPGGANs’ good generalization ability to incrementally synthesize conditional whole images with the real image distribution unfilled by the original dataset, improving the training robustness.
We confirm that the realism and diversity of the generated images, judged by three expert physicians Visual Turing Test, do not imply better detection performance; as the t-SNE results show, the CPGGAN-generated images, trained with additional non-tumor normal images, lack diversity probably because the training less focuses on tumors. Moreover, we notice that adding over-sufficient synthetic images leads to more FPs, but not always higher sensitivity, possibly due to the training data imbalance between real and synthetic images; as the t-SNE results reveal, the CPGGAN-generated tumor bonding boxes have a moderately similar—mutually complementary—distribution to the real ones; thus, GAN-overwhelming training images may decrease the necessary influence of the real samples and harm training, rather than providing robustness. Lastly, image-to-image GAN-based DA just moderately facilitates detection with less additional FPs, probably due to the lack of realism. However, further investigations are needed to maximize the effect of the CPGGAN-based medical image augmentation.
For example, we could verify the effect of further realism in return for less diversity by combining loss with the WGAN-GP loss for GAN training. We can also combine those CPGGAN-generated images, trained without/with additional brain images, similarly to ensemble learning [125]. Lastly, we plan to define a new GAN loss function that directly optimizes the detection results, instead of realism, similarly to the three-player GAN for optimizing classification results [119].
Overall, minimizing expert physicians’ annotation efforts, our novel CPGGAN-based DA approach sheds light on diagnostic and prognostic medical applications, not limited to brain metastases detection; future studies, especially on 3D bounding box detection with highly-rough annotation, are required to extend our promising results. Along with the DA, the CPGGANs has other potential clinical applications in oncology: (i) A data anonymization tool to share patients’ data outside their institution for training while preserving detection performance. Such a GAN-based application is reported in Shin et al. [77]; (ii) A physician training tool to display random synthetic medical images with abnormalities at both common and rare positions/sizes, by training CPGGANs on highly unbalanced medical datasets (i.e., limited pathological and abundant normal samples, respectively). It can help train medical students and radiology trainees despite infrastructural and legal constraints [13].
Chapter 7 GAN-based Medical Image Augmentation for 3D Detection
7.1 Prologue to Fourth Project
7.1.1 Project Publication
-
•
Synthesizing Diverse Lung Nodules Wherever Massively: 3D Multi-Conditional GAN-based CT Image Augmentation for Object Detection. C. Han, Y. Kitamura, A. Kudo, A. Ichinose, L. Rundo, Y. Furukawa, K. Umemoto, H. Nakayama, Y. Li, In International Conference on 3D Vision (3DV), Québec City, Canada, pp. 729–737, September 2019.
7.1.2 Context
Prior to this work, no researchers had tackled 3D GANs for general bounding box-based detection whereas 3D Medical Image Analysis can improve diagnosis by capturing anatomical and functional information. Jin et al. had used an image-to-image GAN to generate CT images of lung nodules including the surrounding tissues by inputting a VOI centered at a lung nodule, but with a central sphere region erased [78]; however, they had targeted annotation-expensive segmentation, instead of the detection, also translating both nodules/surroundings via expensive computation. Without conditioning a noise-to-image GAN with nodule position, Gao et al. had generated 3D nodule subvolumes only applicable to their subvolume-based detector using binary classification [76]. Unfortunately, no research had focused on multiple GAN conditions for more versatile 3D GANs while lesions vary in position, size, and attenuation.
7.1.3 Contributions
This project’s primary contribution is to propose a novel 3D pathology-aware multi-conditional GAN called 3D MCGAN for improved 3D bounding box-based detection in general; it translates noise boxes into realistic/diverse lung nodules placed naturally at desired position, size, and attenuation on CT scans—inputting the noise box with the surrounding tissues has the effect of combining the noise-to-image and image-to-image GANs. The nodule-only generation, not translating the surroundings, can decrease computational cost. By so doing, our 3D MCGAN-based DA boosts sensitivity in nodule detection under any nodule size/attenuation at fixed FP rates. Moreover, we find that GAN training with loss could increase synthetic images’ realism, but decrease DA performance. Using proper augmentation ratio (i.e., ) could improve the DA performance. Considering the outstanding realism confirmed by physicians, it could perform as a physician training tool to display realistic medical images with desired abnormalities (i.e., position, size, and attenuation).
7.1.4 Recent Developments
According to their arXiv paper, Xu et al. have generated realistic/diverse CT images of lung nodules combining the image-to-image GAN with gene expression profiles [126].
7.2 Motivation
Accurate CAD, thanks to recent CNNs, can alleviate the risk of overlooking the diagnosis in a clinical environment. Such great success of CNNs, including diabetic eye disease diagnosis [120], primarily derives from large-scale annotated training data to sufficiently cover the real data distribution. However, obtaining and annotating such diverse pathological images are laborious tasks; thus, the massive generation of proper synthetic training images matters for reliable diagnosis. Researchers usually use classical DA techniques, such as geometric/intensity transformations [89, 90]. However, those one-to-one translated images have intrinsically similar appearance and cannot sufficiently cover the real image distribution, causing limited performance improvement; in this regard, thanks to their good generalization ability, GANs [7] can generate realistic but completely new samples using many-to-many mappings for further performance improvement; GANs showed excellent DA performance in computer vision, including performance improvement in eye-gaze estimation [8].
This GAN-based DA trend especially applies to medical imaging, where the biggest problem lies in small and fragmented datasets from various scanners. For performance boost in various 2D medical imaging tasks, some researchers used noise-to-image GANs for classification [12, 16, 15]; others used image-to-image GANs for object detection [20] and segmentation [122]. However, although 3D imaging is spreading in radiology (e.g., CT and MRI), such 3D medical GAN-based DA approaches are limited, and mostly focus on segmentation [77, 78]—3D medical image generation is more challenging than 2D one due to expensive computational cost and strong anatomical consistency. Accordingly, no 3D conditional GAN-based DA approach exists for general bounding box-based 3D object detection, while it can locate disease areas with physicians’ minimum annotation cost, unlike rigorous 3D segmentation. Moreover, since lesions vary in position/size/attenuation, further GAN-based DA performance requires multiple conditions.
So, how can GAN generate realistic/diverse 3D nodules placed naturally on lung CT with multiple conditions to boost sensitivity in any 3D object detector? For accurate 3D CNN-based nodule detection (Fig. 7.1), we propose 3D MCGAN to generate nodules—such nodule detection is clinically valuable for the early diagnosis/treatment of lung cancer, the deadliest cancer [127]. Since nodules vary in position/size/attenuation, to improve CNN’s robustness, we adopt two discriminators with different loss functions for conditioning: the context discriminator learns to classify real vs synthetic nodule/surrounding pairs with noise box-centered surroundings; the nodule discriminator attempts to classify real vs synthetic nodules with size and attenuation conditions. We also evaluate the synthetic images’ realism via Visual Turing Test [92] by two expert physicians, and visualize the data distribution via t-SNE [100]. The 3D MCGAN-generated additional training images can achieve higher sensitivity under any nodule size/attenuation at fixed FP rates. Lastly, this study suggests training GANs without loss and using proper augmentation ratio (i.e., ) for better medical GAN-based DA performance.
Research Questions. We mainly address two questions:
-
•
3D Multiple GAN Conditioning: How can we condition 3D GANs to naturally place objects of random shape, unlike rigorous segmentation, at desired position/size/attenuation based on bounding box masks?
-
•
Synthetic Images for DA: How can we set the number of real/synthetic training data and GAN loss functions to achieve the best detection performance?
Contributions. Our main contributions are as follows:
-
•
3D Multi-conditional Image Generation: This first multi-conditional pathological image generation approach shows that 3D MCGAN can generate realistic and diverse nodules placed on lung CT at desired position/size/attenuation, which even expert physicians cannot distinguish from real ones.
-
•
Misdiagnosis Prevention: This first GAN-based DA method available for any 3D object detector allows to boost sensitivity at fixed FP rates in CAD with limited medical images/annotation.
-
•
Medical GAN-based DA: This study implies that training GANs without loss and using proper augmentation ratio (i.e., ) may boost CNN-based detection performance with higher sensitivity and less FPs in medical imaging.
7.3 Materials and Methods
7.3.1 3D MCGAN-based Image Generation
Data Preparation This study exploits the Lung Image Database Consortium image collection (LIDC) dataset [128] containing chest CT scans with lung nodules. Since the American College of Radiology recommends lung nodule evaluation using thin-slice CT scans [129], we only use scans with the slice thickness mm and mm in-plane pixel spacing mm. Then, we interpolate the slice thickness to mm and exclude scans with slice number .
To explicitly provide MCGAN with meaningful nodule appearance information and thus boost DA performance, the authors further annotate those nodules by size and attenuation for GAN training with multiple conditions: small (slice thickness mm); medium ( mm slice thickness mm); large (slice thickness mm); solid; part-solid; Ground-Glass Nodule (GGN). Afterwards, the remaining dataset ( scans) is divided into: (i) a training set ( scans/ nodules); (ii) a validation set ( scans/ nodules); (iii) a test set ( scans/ nodules); only the training set is used for MCGAN training to be methodologically sound. The training set contains more average nodules since we exclude patients with too many nodules for the validation/test sets; we arrange a clinical environment-like situation, where we could find more healthy patients than highly diseased ones to conduct anomaly detection.
3D MCGAN is a novel GAN training method for DA, generating realistic but new nodules at desired position/size/attenuation, naturally blending with surrounding tissues (Fig. 7.2). We crop/resize various nodules to voxels and replace them with noise boxes from a uniform distribution between , while maintaining their surroundings as VOIs—using those noise boxes, instead of boxes filled with the same voxel values, improves the training robustness; then, we concatenate the VOIs with size/attenuation conditions tiled to voxels (e.g., if the size is small, each voxel of the small condition is filled with , while the medium/large condition voxels are filled with to consider the effect of scaling factor). So, our generator uses the inputs to generate desired nodules in the noise box regions. The 3D U-Net [130]-like generator adopts convolutional layers in encoders and deconvolutional layers in decoders respectively with skip connections to effectively capture both nodule/context information.
We adopt two Pix2Pix GAN [42]-like discriminators with different loss functions: the context discriminator learns to classify real vs synthetic nodule/surrounding pairs with noise box-centered surroundings using Least Squares loss (LSGANs) [106]; the nodule discriminator attempts to classify real vs synthetic nodules with size/attenuation conditions using WGAN-GP [39]. The LSGANs in the context discriminator forces the model to learn surrounding tissue background by reacting more sensitively to every pixel in images than regular GANs. The WGAN-GP in the nodule discriminator allows the model to generate realistic/diverse nodules without focusing too much on details. Empirically, we confirm that such multiple discriminators with the mutually complementary loss functions, along with size/attenuation conditioning, help generate realistic/diverse nodules naturally placed at desired positions on CT scans; similar results are also reported by this work [131] for 2D pedestrian detection without label conditioning. We apply dropout to inject randomness and balance the generator/discriminators. Batch normalization is applied to both convolution (using LeakyReLU) and deconvolution (using ReLU).
Most GAN-based DA approaches use reconstruction loss [76] to generate realistic images, even modifying it for further realism [78]. However, no one has ever validated whether it really helps DA—it assures synthetic images resembling the original ones, sacrificing diversity; thus, to confirm its influence during classifier training, we compare our MCGAN objective without/with it, respectively:
(7.1) | |||||
(7.2) | |||||
We set 100 as a weight for the loss, since empirically it works well for reducing visual artifacts introduced by the GAN loss and most GAN works adopt the weight [42, 131].
3D MCGAN Implementation Details Training lasts for steps with a batch size of and learning rate for the Adam optimizer. We use horizontal/vertical flipping as DA and flip real/synthetic labels once in three times for robustness. During testing, we augment nodules with the same size/attenuation conditions by applying a random combination to real nodules of width/height/depth shift up to and zooming up to for better DA. As post-processing, we blend bounding boxes’ nearest surfaces from all the boundaries by averaging the values of nearest voxels/itself for iterations. We resample the resulting nodules to their original resolution and map back onto the original CT scans to prepare additional training data.
7.3.2 3D Faster RCNN-based Lung Nodule Detection
3D Faster RCNN is a 3D version of Faster RCNN [124] using multi-task loss with a -layer Region Proposal Network of 3D convolutional/batch normalization/ReLU layers. To confirm the effect of MCGAN-based DA, we compare the following detection results trained on (i) real images without GAN-based DA, (ii), (iii), (iv) with // MCGAN-based DA (i.e., // additional synthetic training images) , (v), (vi), (vii) with // MCGAN-based DA trained with loss. During training, we shuffle the real/synthetic image order. We evaluate the detection performance as follows: (i) Free Receiver Operation Characteristic (FROC) analysis, sensitivity as a function of FPs per scan; (ii) Competition Performance Metric (CPM) score [132], average sensitivity at seven pre-defined FP rates: 1/8, 1/4, 1/2, 1, 2, 4, and 8 FPs per scan—this quantifies if a CAD system can identify a significant percentage of nodules with both very few FPs and moderate FPs.
3D Faster RCNN Implementation Details During training, we use a batch size of and learning rate ( after steps) for the SGD optimizer with momentum. The input volume size to the network is set to voxels. As classical DA, a random combination of width/height/depth shift up to and zooming up to are also applied to both real/synthetic images to achieve the best performance. For testing, we pick the model with the highest sensitivity on validation between - steps under IoU threshold /detection threshold to avoid severe FPs.
7.3.3 Clinical Validation via Visual Turing Test
To quantitatively evaluate the realism of MCGAN-generated images, we supply, in random order, to two expert physicians a random selection of real and synthetic lung nodule images with all of 2D axial/coronal/sagittal views at the center. They take four classification tests in ascending order: Test1, 2: real vs MCGAN-generated nodules, trained without/with loss; Test3, 4: real vs MCGAN-generated nodules with surroundings without/with loss.
7.3.4 Visualization via t-SNE
To visually analyze the distribution of real/synthetic images, we use t-SNE [100] on a random selection of real, synthetic, and loss-added synthetic nodule images, with a perplexity of for iterations to get a 2D representation. We normalize the input images to .
7.4 Results
7.4.1 Lung Nodules Generated by 3D MCGAN
We generate realistic nodules in noise box regions at various position/size/attenuation, naturally blending with surrounding tissues including vessels, soft tissues, and thoracic walls (Fig. 7.3). Especially, when trained without loss, those synthetic nodules look clearly more different from the original real ones, including slight shading difference.
CPM by Size (%) | CPM by Attenuation (%) | ||||||
---|---|---|---|---|---|---|---|
CPM (%) | Small | Medium | Large | Solid | Part-solid | GGN | |
632 real images | 51.8 | 44.7 | 61.8 | 62.4 | 65.5 | 46.4 | 24.2 |
+ 3D MCGAN-based DA | 55.0 | 45.2 | 68.3 | 66.2 | 69.9 | 52.1 | 24.4 |
+ 3D MCGAN-based DA | 52.7 | 44.7 | 67.4 | 42.9 | 65.5 | 40.7 | 28.9 |
+ 3D MCGAN-based DA | 51.2 | 41.1 | 64.4 | 66.2 | 61.6 | 57.9 | 27.7 |
+ 3D MCGAN-based DA w/ | 50.8 | 43.0 | 63.3 | 55.6 | 62.6 | 47.1 | 27.1 |
+ 3D MCGAN-based DA w/ | 50.9 | 40.6 | 64.4 | 65.4 | 64.9 | 43.6 | 23.3 |
+ 3D MCGAN-based DA w/ | 47.9 | 38.9 | 59.4 | 61.7 | 59.6 | 50.7 | 22.6 |
7.4.2 Lung Nodule Detection Results
Table 7.1 and Fig. 7.4 show that it is easier to detect nodules with larger size/lower attenuation due to their clear appearance. 3D MCGAN-based DA with less augmentation ratio consistently increases sensitivity at fixed FP rates—especially, training with MCGAN-based DA without loss outperforms training only with real images under any size/attenuation in terms of CPM, achieving average CPM improvement by 0.032. It especially boosts nodule detection performance with larger size and lower attenuation. Fig. 7.5 visually reveals its ability to alleviate the risk of overlooking the nodule diagnosis with clinically acceptable FPs (i.e., the highly-overlapping bounding boxes around nodules only require a physician’s single check by switching on/off transparent alpha-blended annotation on CT scans). Surprisingly, adding more synthetic images tends to decrease sensitivity, due to the real/synthetic training image balance. Moreover, further nodule realism introduced by loss rather decreases sensitivity as loss sacrifices diversity in return for the realism.
7.4.3 Visual Turing Test Results
As Table 7.2 shows, expert physicians fail to classify real vs MCGAN-generated nodules without surrounding tissues—even regarding the synthetic nodules trained without loss more realistic than the real ones. Contrarily, they relatively recognize the synthetic nodules with surroundings due to slight shading difference between the nodules/surroundings, especially when trained without the reconstruction loss. Considering the synthetic images’ realism, CPGGANs might perform as a tool to train medical students and radiology trainees when enough medical images are unavailable, such as abnormalities at rare position/size/attenuation. Such GAN applications are clinically promising [13].
Accuracy (%) | Real as Real (%) | Real as Synt (%) | Synt as Real (%) | Synt as Synt (%) | ||
Test 1 | Physician 1 | 43 | 38 | 62 | 52 | 48 |
Physician 2 | 43 | 26 | 74 | 40 | 60 | |
Test 2 | Physician 1 | 57 | 44 | 56 | 30 | 70 |
Physician 2 | 53 | 22 | 78 | 16 | 84 | |
Test 3 | Physician 1 | 62 | 50 | 50 | 26 | 74 |
Physician 2 | 79 | 64 | 36 | 6 | 94 | |
Test 4 | Physician 1 | 58 | 42 | 58 | 26 | 74 |
Physician 2 | 66 | 72 | 28 | 40 | 60 | |
7.4.4 T-SNE Results
Implying their effective DA performance, synthetic nodules have a similar distribution to real ones, but concentrated in left inner areas with less real ones especially when trained without loss (Fig. 7.6)–using only GAN loss during training can avoid overwhelming influence from the real image samples, resulting in a moderately similar distribution; thus, those synthetic images can partially fill the real image distribution uncovered by the original dataset.
7.5 Conclusion
Our bounding box-based 3D MCGAN can generate diverse CT-realistic nodules at desired position/size/attenuation, naturally blending with surrounding tissues—those synthetic training data boost sensitivity under any size/attenuation at fixed FP rates in 3D CNN-based nodule detection. This attributes to the MCGAN’s good generalization ability coming from multiple discriminators with mutually complementary loss functions, along with informative size/attenuation conditioning; they allow to cover the real image distribution unfilled by the original dataset, improving the training robustness.
Surprisingly, we find that adding over-sufficient synthetic images produces worse results due to the real/synthetic image balance; as t-SNE results show, the synthetic images only partially cover the real image distribution, and thus GAN-overwhelming training images rather harm training. Moreover, we notice that GAN training without loss obtains better DA performance thanks to increased diversity providing robustness; also, expert physicians confirm their sufficient realism without loss.
Overall, our 3D MCGAN could help minimize expert physicians’ time-consuming annotation tasks and overcome the general medical data paucity, not limited to lung CT nodules. As future work, we will investigate the MCGAN-based DA results without size/attenuation conditioning to confirm their influence on DA performance. Moreover, we will compare our DA results with other non-GAN-based recent DA approaches, such as Mixup [50] and Cutout [52]. For further performance boost, we plan to directly optimize the detection results for MCGANs, instead of realism, similarly to the three-player GAN for classification [119]. Lastly, we will investigate how our MCGAN can perform as a physician training tool to display random realistic medical images with desired abnormalities (i.e., position/size/attenuation conditions) to help train medical students and radiology trainees despite infrastructural and legal constraints [13].
Chapter 8 Discussions on Developing Clinically
Relevant AI-Powered Diagnosis Systems
8.1 Prologue to First Project
8.1.1 Project Publication
-
•
Bridging the gap between AI and healthcare sides: towards developing clinically relevant AI-powered diagnosis systems. C. Han, L. Rundo, K. Murao, T. Nemoto, H. Nakayama, In IFIP International Conference on Artificial Intelligence Applications and Innovations (AIAI), pp. 320–333, June 2020.
8.2 Feedback from Physicians
8.2.1 Methods for Questionnaire Evaluation
To confirm the clinical relevance for diagnosis of our proposed pathology-aware GAN methods for DA and physician training respectively, we conduct a questionnaire survey for Japanese physicians who interpret MR and CT images in daily practice. The experimental settings are the following:
-
•
Subjects: physicians (i.e., a radiologist, a psychiatrist, and a physiatrist) committed to (at least one of) our pathology-aware GAN projects and project non-related radiologists without much AI background.
-
•
Experiments: Physicians are asked to answer the following questionnaire within 2 weeks from December 6th, 2019 after reading 10 summary slides written in Japanese111Available via Dropbox: https://www.dropbox.com/sh/bacowc3ilz1p1r3/AABNS9SyjArHq8BntgaODLb2a?dl=0 about general Medical Image Analysis and our pathology-aware GAN projects along with example synthesized images. We conduct both qualitative (i.e., free comments) and quantitative (i.e., five-point Likert scale [133]) evaluation: Likert scale 1 very negative, 2 negative, 3 neutral, 4 positive, 5 very positive.
-
•
Question 1: Are you keen to exploit medical AI in general when it achieves accurate and reliable performance in the near future? (five-point Likert scale) Please tell us your expectations, wishes, and worries (free comments).
-
•
Question 2: What do you think about using GAN-generated images for DA? (five-point Likert scale) Please tell us your expectations, wishes, and worries (free comments).
-
•
Question 3: What do you think about using GAN-generated images for physician training? (five-point Likert scale) Please tell us your expectations, wishes, and worries (free comments).
-
•
Question 4: Any comments or suggestions about our projects towards developing clinically relevant AI-powered systems based on your daily diagnosis experience?
8.2.2 Results
We show the questions and Japanese physicians’ corresponding answers.
Question 1: Are you keen to exploit medical AI in general when it achieves accurate and reliable performance in the near future?
-
•
Likert scale Project-related physicians: 5 5 5 (average: 5)
Project non-related radiologists: 5 5 3 4 5 5 (average: 4.5) -
•
Free comments (one comment for each physician)
-
•
As radiologists, we need AI-based diagnosis during image interpretation as soon as possible.
-
•
It is common to conduct further medical examinations when identifying disease is difficult from CT/MR images; thus, if AI-based diagnosis outperforms that of physicians, such clinical decision support systems could prevent unnecessary examinations. Moreover, recently lung cancer misdiagnosis occurred in Japan, but AI technologies may prevent such death caused by misdiagnosis.
-
•
The lack of diagnosticians is very evident in Healthcare, so AI has great potential to support us. It may be already applicable without severe problems for typical disease cases.
-
•
I am looking forward to its practical applications, especially at low or zero price.
-
•
I would like to use AI-based diagnosis as a kind of data, but it is yet uncertain how much I trust AI.
-
•
I am wondering whether such systems will become popular due to practical problems such as introduction cost.
-
•
The definition of accurate and reliable is unclear. Since a physician’s annotation is always subjective, we cannot claim that AI-based diagnosis is really correct even if AI diagnoses similarly to the specific physician. Because I do not believe other physicians’ diagnosis, but my own eyes, I would use AI just to identify abnormal candidates.
As expected, the project-related physicians are AI-enthusiastic while the project non-related radiologists are also generally very positive about the medical AI. Many of them appeal the necessity of AI-based diagnosis for more reliable diagnosis because of the lack of physicians. Meanwhile, other physicians worry about its cost and reliability. We may be able to persuade them by showing expected profitability (e.g., currently CT scanners have an earning rate 16% and CT scans require 2-20 minutes for interpretation in Japan); similarly, we can explain how experts annotate medical images and AI diagnoses disease based on them (e.g., multiple physicians, not a single one, can annotate the images via discussion).
Question 2: What do you think about using GAN-generated images for DA?
-
•
Likert scale Project-related physicians: 5 5 4 (average: 4.7)
Project non-related radiologists: 4 5 4 4 4 4 (average: 4.2) -
•
Free comments (one comment for each physician)
-
•
Achieved accuracy improvement shows its superiority in identifying diverse disease.
-
•
It would be effective, especially as rare disease training data.
-
•
I am looking forward to the future with advanced GAN technology.
-
•
It significantly improves detection sensitivity; but I am also curious about its influence on other metrics, such as specificity.
-
•
If Deep Learning could be more effective, we should introduce it; but anonymization would be important for privacy preservation.
-
•
Achieved accuracy improvement shows its superiority in identifying diverse disease.
-
•
It would be effective to train AI on data-limited disease, but which means that AI is inferior to humans.
-
•
It would be helpful if such DA improves accuracy and reliability. Since I am not familiar with AI and a generator/classifier’s failure judgment mechanisms, I am uncertain whether it will really increase reliability though.
As expected, the project-related physicians are very positive about the GAN-based DA while the project non-related radiologists are also positive. Many of them are satisfied with its achieved accuracy/sensitivity improvement when available annotated images are limited. However, similarly to their opinions on general Medical Image Analysis, some physicians question its reliability.
Question 3: What do you think about using GAN-generated images for physician training?
-
•
Likert scale Project-related physicians: 3 4 3 (average: 3.3)
Project non-related radiologists: 3 5 2 3 2 3 (average: 3) -
•
Free comments (one comment for each physician)
-
•
In future medical care, physicians should actively introduce and learn new technology; in this sense, GAN technology should be actively used for physician training in rare diseases.
-
•
It could be useful for medical student training, which aims for 85% accuracy by covering typical cases. But expert physician training aims for over 85% accuracy by comparing typical/atypical cases and acquiring new understanding—real atypical images are essential.
-
•
In physician training, we use radiological images after definite diagnosis, such as pathological examination—but, we actually lack rare disease cases. Since the GAN-generated images’ realism fluctuates based on image augmentation schemes and available training images, further realistic image generation of the rare cases would help the physician training.
-
•
It depends on how to construct the system.
-
•
Which specific usage is assumed for such physician training?
-
•
I cannot state an opinion before actually using the system, but I strongly recognize the importance of looking at real images.
-
•
I do not exactly understand in which situation such physician training is used, but eventually training with realistic images would be also helpful. However, if real images are available, using them would be better.
We generally receive neutral feedback because we do not provide a concrete physician training tool, but instead general pathology-aware generation ideas with example synthesized images—thus, some physicians are positive, and some are not. A physician provides a key idea about a pathology-coverage rate for medical student/expert physician training, respectively; for extensive physician training with GAN-generated atypical images, along with pathology-aware GAN-based extrapolation, further GAN-based extrapolation would be valuable.
Question 4: Any comments or suggestions about our pathology-aware GAN projects towards developing clinically relevant AI-powered systems based on your daily diagnosis experience?
-
•
This approach will change the way physicians work. I have high expectations for AI-based diagnosis, so I hope it to overcome the legal barrier.
-
•
For now, please show small abnormal findings, such as nodules and ground glass opacities—it would halve radiologists’ efforts. Then, we could develop accurate diagnosis step by step.
-
•
Showing abnormal findings with their shapes/sizes/disease names would increase diagnosis accuracy. But I also would like to know how diagnosticians’ roles change after all.
-
•
I hope that this approach will lead to physicians’ work reduction in the future.
-
•
Please develop reliable AI systems by increasing accuracy with the GAN-based image augmentation.
-
•
GANs can generate typical images, but not atypical images; this would be the next challenge.
-
•
AI can alert physicians to detect typical cases, and thus decrease interpretation time; however, it may lead to the diagnosticians’ easy diagnosis without much consideration. Especially in Japan, we currently often conduct unnecessary diagnostic tests, so the diagnosticians should be more responsible of their own duties after introducing AI.
Most physicians look excited about our pathology-aware GAN-based image augmentation projects and express their clinically relevant requests. The next steps lie in performing further GAN-based extrapolation, developing clinician-friendly systems with new practice guidelines, and overcoming legal/financial constraints.
8.3 AI and Healthcare Workshop
8.3.1 Methods for Workshop Evaluation
Convolutional Neural Networks (CNNs) have achieved accurate and reliable Computer-Aided Diagnosis (CAD), occasionally outperforming expert physicians [1, 134, 135]. However, such research results cannot be easily applied to a clinical environment: AI and Healthcare sides have a huge gap around technology, funding, and people, such as clinical significance/interpretation, data acquisition, commercial purpose, and anxiety about AI. Aiming to identify/bridge the gap between AI and Healthcare sides in Japan towards develop medical AI fitting into a clinical environment in five years, we hold a workshop for Japanese professionals with various AI and/or Healthcare background. The experimental settings are the following:
Time (mins) | Activity |
---|---|
Introduction | |
10 | 1. Explanation of the workshop’s purpose and flow* |
10 | 2. Self-introduction and explanation of motivation for participation |
5 | 3. Grouping into two groups based on background* |
Learning: Knowing Medical Image Analysis | |
15 | 1. TED speech video watching: Artificial Intelligence Can Change |
the future of Medical Diagnosis* | |
35 | 2. Lecture: Overview of Medical Image Analysis including |
state-of-the-art research, well-known challenges/solutions, and our | |
pathology-aware GAN projects summary* | |
(its video in Japanese: https://youtu.be/rTQLknPvnqs) | |
10 | 3. Sharing expectations, wishes, and worries about Medical Image |
Analysis (its video in Japanese: https://youtu.be/ILPEGga-hkY) | |
10 | Intermission |
Thinking: Finding How to Develop Robust Medical AI | |
25 | 1. Identifying the intrinsic gap between AI/Healthcare sides |
after sharing their common and different thinking/working styles | |
60 | 2. Finding how to develop gap-bridging medical AI fitting into |
a clinical environment in five years | |
10 | Intermission |
Summary | |
25 | 1. Presentation |
10 | 2. Sharing workshop impressions and ideas to apply obtained |
knowledge (its video in Japanese: https://youtu.be/F31tPR3m8hs) | |
5 | 3. Answering a questionnaire about satisfaction/further comments |
5 | 4. Closing remarks* |
-
•
Subjects: Medical Imaging experts (i.e., a Medical Imaging researcher and a medical AI startup entrepreneur), physicians (i.e., a radiologist and a psychiatrist), and generalists between Healthcare and Informatics (i.e., a nurse and researcher in medical information standardization, a general practitioner and researcher in medical communication, and a medical technology manufacturer’s owner and researcher in health disparities)
-
•
Experiments: As its program shows (Table 8.1), during the workshop, we conduct 2 activities: (Learning) Know the overview of Medical Image Analysis, including state-of-the-art research, well-known challenges/solutions, and the summary of our pathology-aware GAN projects; (Thinking) Find the intrinsic gap and its solutions between AI researchers and Healthcare workers after sharing their common and different thinking/working styles. Supported by GCL program, this workshop was held on March 17th, 2019 at Nakayama Future Factory, Open Studio, The University of Tokyo, Tokyo, Japan.
8.3.2 Results
We show the summary of clinically-relevant findings from this Japanese workshop.
Gap Between AI and Healthcare Sides
Gap 1: AI, including Deep Learning, does not provide clear decision criteria, does it make physicians reluctant to use it in a clinical environment, especially for diagnosis?
-
•
Healthcare side: We rather expect applications other than diagnosis. If we use AI for diagnosis, instead of replacing physicians, we suppose a reliable second opinion, such as alert to avoid misdiagnosis, based on various clinical data not limited to images—every single diagnostician is anxious about their diagnosis. AI only provides minimum explanation, such as a heatmap showing attention, which makes persuading not only the physicians but also patients difficult; so, the physicians’ intervention is essential for intuitive explanation. Methodological safety and feeling safe are different. In this sense, pursuing explainable AI generally decreases AI’s diagnostic accuracy [136], so physicians should still serve as mediators by engaging in high-level conversation or interaction with patients. Moreover, according to the medical law in most countries including Japan, only doctors can make the final decision. The first autonomous AI-based diagnosis without a physician was cleared by the Food and Drug Administration in the US in 2018 [137], but such a case is exceptional.
-
•
AI side: Compared with other systems or physicians, Deep Learning’s explanation is not particularly poor, so we require too severe standards for AI; the word AI is excessively promoting anxiety and perfection. If we could thoroughly verify the reliability of its diagnosis against physicians by exploring uncertainty measures [138], such intuitive explanation would be optional.
Gap 2: Are there any benefits to actually introducing medical AI?
-
•
Healthcare side: After all, even if AI can achieve high accuracy and convenient operation, hospitals would not introduce it without any commercial benefits. Moreover, small clinics, where physicians are desperately needed, often do not have CT or MRI scanners [139].
-
•
AI side: The commercial deployment of medical AI is strongly tied to diagnostic accuracy [140]; so, if it can achieve significantly outstanding accuracy at various tasks in the near future, patients would not visit hospitals/clinics without AI. Accordingly, introducing medical AI would become profitable in five years.
Gap 3: Is medical AI’s diagnostic accuracy reliable?
-
•
Healthcare side: To evaluate AI’s diagnostic performance, we should consider many metrics, such as sensitivity and specificity. Moreover, its generalization ability for medical data highly relies on inter-scanner/inter-individual variability [141]. How can we evaluate whether it is suitable as a clinically applicable system?
-
•
AI side: Generally, alleviating the risk of overlooking the diagnosis is the most important, so sensitivity matters more than specificity unless their balance is highly disturbed. Recently, such research on medical AI that is robust to different datasets is active [97].
How to Develop Medical AI Fitting into a Clinical Environment in Five Years
Why: Clinical significance/interpretation
-
•
Challenges: We need to clarify which clinical situations actually require AI introduction. Moreover, AI’s early diagnosis might not be always beneficial for patients.
-
•
Solutions: Due to nearly endless disease types and frequent misdiagnosis coming from physicians’ fatigue, we should use it as alert to avoid misdiagnosis [142] (e.g., reliable second opinion), instead of replacing physicians. It should help prevent oversight in diagnostic tests not only with CT and MRI, but also with blood data, chest X-ray, and mammography before taking CT and MRI [143]. It could be also applied to segmentation for radiation therapy [144], neurosurgery navigation [145], and pressure ulcers’ echo evaluation. Along with improving the diagnosis, it would also make the physicians’ workflow easier, such as by denoising [146]. Patients should decide whether they accept AI-based diagnosis under informed consent.
How: Data acquisition
-
•
Challenges: Ethical screening in Japan is exceptionally strict, so acquiring and sharing large-scale medical data/annotation are challenging—it also applies to Europe due to General Data Protection Regulation [147]. Considering the speed of technological advances in AI, adopting it for medical devices is difficult in Japan, unlike in medical AI-ready countries, such as the US, where the ethical screening is relatively loose in return for the responsibility of monitoring system stability. Moreover, whenever diagnostic criteria changes, we need further reviews and software modifications; for example, the Tumor-lymph Node-Metastasis (TNM) classification [148] criteria changed for oropharyngeal cancer in 2018 and for lung cancer in 2017, respectively. Diagnostic equipment/target changes also require large-scale data/annotation acquisition again.
-
•
Solutions: For Japan to keep pace, the ethical screening should be adequate to the other leading countries. Currently, overseas research and clinical trials are proceeding much faster, so it seems better to collaborate with overseas companies than to do it in Japan alone. Moreover, complete medical checkup, which is extremely costly, is unique in East Asia, so Japan could be superior in individuals’ multiple medical data—Japan is the only country, where most workers 40 or older are required to have medical checkups once a year independent of their health conditions by the Industrial Safety and Health Act [149]. To handle changes in diagnostic criteria/equipment and overcome dataset/task dependency, it is necessary to establish a common database creation workflow [150] by regularly entering electronic medical records into the database. For reducing data acquisition/annotation cost, AI techniques, such as GAN-based DA [21] and domain adaptation [151], would be effective.
How: Commercial deployment
-
•
Challenges: Hospitals currently do not have commercial benefits to actually introduce medical AI.
-
•
Solutions: For example, it would be possible to build AI-powered hospitals [152] operated with less staff. Medical manufacturers could also standardize data format [153], such as for X-ray, and provide some AI services. Many IT giants like Google are now working on medical AI to collect massive biomedical data [154], so they could help rural areas and developing countries, where physician shortage is severe [139], at relatively low cost.
How: Safety and feeling safe
- •
-
•
Solutions: We should integrate various clinical data, such as blood test biomarkers and multiomics, with images [143]. Moreover, developing bias-robust technology is important since confounding factors are inevitable [158]. To prevent oversight, prioritizing sensitivity over specificity is essential while maintaining a balance [159]. We should also devise education for medical AI users, such as result interpretation, to reassure patients [160].
Chapter 9 Conclusion
9.1 Final Remarks
Inspired by their excellent ability to generate realistic and diverse images, we propose to use noise-to-image GANs for (i) Medical DA and (ii) physician training [11]. Through information conversion, such applications can relieve the lack of pathological data and their annotation; this is uniquely and intrinsically important in Medical Image Analysis, as CNN generalization becomes unstable on unseen data due to large inter-subject, inter-pathology, and cross-modality variability [97, 161, 162]. Towards clinically relevant implementation for the DA and physician training, we find effective loss functions and training schemes for each of them [15, 16]—the diversity matters more for the DA to sufficiently fill the real image distribution whereas the realism matters more for the physician training not to confuse medical students and radiology trainees.
Specifically, our results imply that GAN training without loss, using proper augmentation ratio (i.e., ), and further refining synthetic images’ texture/shape could improve the DA performance, whereas discarding weird-looking synthetic images to humans is unnecessary; for example, adding over-sufficient GAN-generated training images leads to more FPs in detection, but not always higher sensitivity, due to the real/synthetic training data balance (both of their distributions are biased, but differently). Regarding the physician training, GAN training with loss, GAN training on additional normal images, and post-processing, such as by Poisson image editing [163], could improve the synthetic images’ realism; for instance, the GAN training on normal images along with pathological ones, remarkably facilitates the realism of both healthy and pathological parts while they do not include abnormality.
Because such excellent realism and diversity can be achieved by GAN-based interpolation and extrapolation, we propose novel 2D/3D pathology-aware GANs for bounding box-based pathology detection [20, 21]: (Interpolation) The GAN-based medical image augmentation is reliable because medical modalities (e.g., X-ray, CT, MRI) can display the human body’s strong anatomical consistency at fixed position while clearly reflecting inter-subject variability [9, 10]—this is different from natural images, where various objects can appear at any position; (Extrapolation) The pathology-aware GANs are promising because common and/or desired medical priors can play a key role in the conditioning—theoretically, infinite conditioning instances, external to the training data, exist and enforcing such constraints have an extrapolation effect via model reduction [19].
After conducting a questionnaire survey about our GAN projects for 9 physicians and holding a workshop about how to develop medical AI fitting into a clinical environment for 7 professionals with various AI and/or Healthcare background, we confirm our pathology-aware GANs’ clinical relevance for diagnosis: (DA) They could be integrated into a clinical decision support system; since CT has a much higher earning rate and longer interpretation time than MRI (16% to 3% and 2-20 minutes to 1 minute in Japan), alerting abnormal findings on CT, such as nodules/ground glass opacities, would halve radiologists’ efforts and increase hospitals’ financial outcomes; (Physician training) They could perform as a non-expert physician training tool; when the normal training images are sufficiently available, we can stably generate typical pathological images useful for medical student training, thanks to the excellent interpolation; but it is still challenging to generate atypical images needed for expert physician training. Whereas our pathology-aware bounding box conditioning largely improves extrapolation ability, better DA and physician training would require further GAN-based extrapolation.
9.2 Future Work
We believe that the next steps towards GAN-based extrapolation and thus atypical pathological image generation lie in (i) generation by parts with coordinate conditions [164], (ii) generation with both image and gene expression conditions [126], and (iii) transfer learning among different body parts and disease types [165]. Due to biological constraints, human interaction is restricted to part of the surrounding environment. Accordingly, we must reason spatial relationships across the surrounding parts to piece them together. Similarly, since machine performance also depends on computational constraints, it is plausible for a generator to generate partial images using the corresponding spatial coordinate conditions—meanwhile, a discriminator attempts to judge realism across the assembled patches by global coherence, local appearance, and edge-crossing continuity. This approach allowed COnditional COordinate GAN (COCO-GAN) to generate state-of-the-art realistic and seamless full images [164]. Since human anatomy has a much stronger local consistency than various object relationships in natural images, reasoning the body’s spatial relationships, like the COCO-GAN, would perform effective extrapolation both for medical DA and physician training.
We can also condition the GANs both on the image features and gene expression profiles to non-invasively identify molecular properties of disease. By so doing, Xu et al. succeeded to produce realistic synthetic CT images of lung nodules [126]. If the gene expression data are available, such condition fusing could be helpful for the medical DA and physician training.
Such information conversion, not limited to the GAN conditioning, should locate in the core of future Medical Image Analysis to overcome the data paucity. Whereas the transfer learning from large-scale natural image/video datasets for CNNs is already common in Medical Image Analysis, such pre-trained models cannot extract general human anatomical features. Accordingly, pre-training on large-scale 3D medical volumes for CNNs, such as CT and MRI, significantly outperformed the pre-training on natural videos or training from scratch for classification and segmentation [165] both by accuracy and training convergence speed. Similarly, transfer learning from mammography for the CNNs also significantly improved mass detection on digital breast tomosynthesis slices [166]. Such transfer learning across different body parts and disease types for the GANs would also largely improve their extrapolation ability.
References
- [1] E. J. Hwang, S. Park, K. Jin, J. I. Kim, S. Y. Choi, et al., “Development and validation of a deep learning–based automatic detection algorithm for active pulmonary tuberculosis on chest radiographs,” Clin. Infect. Dis., vol. 69, no. 5, pp. 739–747, 2018.
- [2] P. Rajpurkar, J. Irvin, K. Zhu, B. Yang, H. Mehta, et al., “Chexnet: Radiologist-level pneumonia detection on chest X-rays with deep learning,” arXiv preprint arXiv:1711.05225, 2017.
- [3] G. Litjens, T. Kooi, B. E. Bejnordi, A. A. A. Setio, F. Ciompi, et al., “A survey on deep learning in medical image analysis,” Med. Image Anal., vol. 42, pp. 60–88, 2017.
- [4] H. Greenspan, B. Van Ginneken, and R. M. Summers, “Guest editorial deep learning in medical imaging: Overview and future promise of an exciting new technique,” IEEE Trans. Med. imaging, vol. 35, no. 5, pp. 1153–1159, 2016.
- [5] O. Chapelle, B. Scholkopf, and A. Zien, “Semi-supervised learning [book reviews],” IEEE T. Neural Networ., vol. 20, no. 3, pp. 542–542, 2009.
- [6] O. Vinyals, C. Blundell, T. Lillicrap, k. kavukcuoglu, and D. Wierstra, “Matching networks for one shot learning,” in Proc. Advances in Neural Information Processing Systems (NIPS), pp. 3630–3638, 2016.
- [7] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, et al., “Generative adversarial nets,” in Proc. Advances in Neural Information Processing Systems (NIPS), pp. 2672–2680, 2014.
- [8] A. Shrivastava, T. Pfister, O. Tuzel, J. Susskind, W. Wang, and R. Webb, “Learning from simulated and unsupervised images through adversarial training,” in Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2107–2116, 2017.
- [9] J. Hsieh, Computed tomography: Principles, design, artifacts, and recent advances. SPIE Bellingham, WA, 2009.
- [10] R. W. Brown, Y. N. Cheng, E. M. Haacke, M. R. Thompson, and R. Venkatesan, Magnetic resonance imaging: Physical principles and sequence design. John Wiley & Sons, 2014.
- [11] C. Han, H. Hayashi, L. Rundo, R. Araki, W. Shimoda, et al., “GAN-based synthetic brain MR image generation,” in Proc. IEEE International Symposium on Biomedical Imaging (ISBI), pp. 734–738, 2018.
- [12] M. Frid-Adar, I. Diamant, E. Klang, M. Amitai, J. Goldberger, and H. Greenspan, “GAN-based synthetic medical image augmentation for increased CNN performance in liver lesion classification,” Neurocomputing, vol. 321, pp. 321–331, 2018.
- [13] S. G. Finlayson, H. Lee, I. S. Kohane, and L. Oakden-Rayner, “Towards generative adversarial networks as a new paradigm for radiology education,” in Proc. Machine Learning for Health (ML4H) Workshop arXiv preprint arXiv :1812.01547, 2018.
- [14] C. Han, K. Murao, S. Satoh, and H. Nakayama, “Learning more with less: GAN-based medical image augmentation,” Med. Imaging Tech., vol. 37, no. 3, pp. 137–142, 2019.
- [15] C. Han, L. Rundo, R. Araki, Y. Furukawa, G. Mauri, et al., “Infinite brain MR images: PGGAN-based data augmentation for tumor detection,” in Neural Approaches to Dynamics of Signal Exchanges, Smart Innovation, Systems and Technologies, pp. 291–303, Springer, 2019.
- [16] C. Han, L. Rundo, R. Araki, Y. Nagano, Y. Furukawa, et al., “Combining noise-to-image and image-to-image GANs: Brain MR image augmentation for tumor detection,” IEEE Access, vol. 7, pp. 156966–156977, 2019.
- [17] T. Karras, T. Aila, S. Laine, and J. Lehtinen, “Progressive growing of GANs for improved quality, stability, and variation,” in Proc. International Conference on Learning Representations (ICLR) arXiv preprint arXiv:1710.10196v3, 2018.
- [18] X. Huang, M. Y. Liu, S. Belongie, and J. Kautz, “Multimodal unsupervised image-to-image translation,” in Proc. European Conference on Computer Vision (ECCV), pp. 172–189, 2018.
- [19] P. Stinis, T. Hagge, A. M. Tartakovsky, and E. Yeung, “Enforcing constraints for interpolation and extrapolation in generative adversarial networks,” J. Comput. Phys., vol. 397, p. 108844, 2019.
- [20] C. Han, K. Murao, T. Noguchi, Y. Kawata, F. Uchiyama, et al., “Learning more with less: Conditional PGGAN-based data augmentation for brain metastases detection using highly-rough annotation on MR images,” in Proc. ACM International Conference on Information and Knowledge Management (CIKM), pp. 119–127, 2019.
- [21] C. Han, Y. Kitamura, A. Kudo, A. Ichinose, L. Rundo, et al., “Synthesizing diverse lung nodules wherever massively: 3D multi-conditional GAN-based CT image augmentation for object detection,” in Proc. International Conference on 3D Vision (3DV), pp. 729–737, 2019.
- [22] P. Lambin, R. T. H. Leijenaar, T. M. Deist, J. Peerlings, E. E. C. De Jong, et al., “Radiomics: The bridge between medical imaging and personalized medicine,” Nat. Rev. Clin. Oncol., vol. 14, no. 12, p. 749, 2017.
- [23] D. Rueckert, B. Glocker, and B. Kainz, “Learning clinically useful information from images: Past, present and future,” Med. Image Anal., vol. 33, p. 13, 18.
- [24] J. Yap, W. Yolland, and P. Tschandl, “Multimodal skin lesion classification using deep learning,” Exp. Dermatol., vol. 27, no. 11, pp. 1261–1267, 2018.
- [25] J. Lao, Y. Chen, Z. Li, Q. Li, J. Zhang, et al., “A deep learning-based radiomics model for prediction of survival in glioblastoma multiforme,” Sci. Rep., vol. 7, no. 1, p. 10353, 2017.
- [26] K. Chaitanya, N. Karani, C. F. Baumgartner, A. Becker, O. Donati, and E. Konukoglu, “Semi-supervised and task-driven data augmentation,” in Proc. International Conference on Information Processing in Medical Imaging (IPMI), pp. 29–41, 2019.
- [27] E. Yiannakopoulou, N. Nikiteas, D. Perrea, and C. Tsigris, “Virtual reality simulators and training in laparoscopic surgery,” Int. J. Surg., vol. 13, pp. 60–64, 2015.
- [28] Z. Zhu, E. Albadawy, A. Saha, J. Zhang, M. R. Harowicz, and M. Mazurowski, “Deep learning for identifying radiogenomic associations in breast cancer,” Comput. Biol. Med., vol. 109, pp. 85–90, 2019.
- [29] L. L. Geyer, U. J. Schoepf, F. G. Meinel, J. W. Nance Jr, G. Bastarrika, et al., “State of the art: Iterative CT reconstruction techniques,” Radiology, vol. 276, no. 2, pp. 339–357, 2015.
- [30] D. W. McRobbie, E. A. Moore, M. J. Graves, and M. R. Prince, MRI from picture to proton. Cambridge university press, 2017.
- [31] Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, vol. 521, no. 7553, pp. 436–444, 2015.
- [32] Y. Bengio and G. Marcus, “DEBATE: Yoshua Bengio and Gary Marcus on the best way forward for AI.” Debate, 2019.
- [33] D. H. Hubel and T. N. Wiesel, “Receptive fields of single neurones in the cat’s striate cortex,” J. Physiol., vol. 148, no. 3, pp. 574–591, 1959.
- [34] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, et al., “Imagenet large scale visual recognition challenge,” Int. J. Comput. Vis., vol. 115, no. 3, pp. 211–252, 2015.
- [35] D. P. Kingma and M. Welling, “Auto-encoding variational Bayes,” in Proc. International Conference on Learning Representations (ICLR), arXiv preprint arXiv:1312.6114, 2014.
- [36] J. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in Proc. IEEE International Conference on Computer Vision (ICCV), pp. 2223–2232, 2017.
- [37] J. M. Wolterink, T. Leiner, M. A. Viergever, and I. Išgum, “Generative adversarial networks for noise reduction in low-dose CT,” IEEE Trans. Med. Imaging, vol. 36, no. 12, pp. 2536–2545, 2017.
- [38] H. Emami, M. Dong, S. P. Nejad-Davarani, and C. K. Glide-Hurst, “Generating synthetic CTs from magnetic resonance images using generative adversarial networks,” Med. Phys., vol. 45, no. 8, pp. 3627–3636, 2018.
- [39] I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin, and A. C. Courville, “Improved training of Wasserstein GANs,” arXiv preprint arXiv:1704.00028, 2017.
- [40] A. Radford, L. Metz, and S. Chintala, “Unsupervised representation learning with deep convolutional generative adversarial networks,” in Proc. International Conference on Learning Representations (ICLR), arXiv preprint arXiv:1511.06434, 2016.
- [41] T. Xu, P. Zhang, Q. Huang, H. Zhang, Z. Gan, et al., “AttnGAN: Fine-grained text to image generation with attentional generative adversarial networks,” in Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1316–1324, 2018.
- [42] P. Isola, J. Y. Zhu, T. Zhou, and A. A. Efros, “Image-to-image translation with conditional adversarial networks,” in Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1125–1134, 2017.
- [43] S. E. Reed, Z. Akata, S. Mohan, S. Tenka, B. Schiele, and H. Lee, “Learning what and where to draw,” in Advances in Neural Information Processing Systems (NIPS), pp. 217–225, 2016.
- [44] H. Park, Y. Yoo, and N. Kwak, “MC-GAN: Multi-conditional generative adversarial network for image synthesis,” in Proc. British Machine Vision Conference (BMVC) arXiv preprint arXiv:1805.01123, 2018.
- [45] K. Hung, C. C. Lee, and S. Choy, “Ubiquitous health monitoring: Integration of wearable sensors, novel sensing techniques, and body sensor networks,” in Mobile Health, pp. 319–342, Springer, 2015.
- [46] M. J. Yaffe, “Emergence of big data and its potential and current limitations in medical imaging,” Semin. Nucl. Med., vol. 49, no. 2, pp. 94–104, 2019.
- [47] L. Perez and J. Wang, “The effectiveness of data augmentation in image classification using deep learning,” arXiv preprint arXiv:1712.04621, 2017.
- [48] C. Shorten and T. M. Khoshgoftaar, “A survey on image data augmentation for deep learning,” J. Big Data, vol. 6, no. 1, p. 60, 2019.
- [49] D. J. Heeger, “The representation of visual stimuli in primary visual cortex,” Curr. Dir. Psychol. Sci., vol. 3, no. 5, pp. 159–163, 1994.
- [50] H. Zhang, M. Cisse, Y. N. Dauphin, and D. Lopez-Paz, “Mixup: Beyond empirical risk minimization,” arXiv preprint arXiv:1710.09412, 2017.
- [51] Y. Tokozume, Y. Ushiku, and T. Harada, “Between-class learning for image classification,” in Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5486–5494, 2018.
- [52] T. DeVries and G. W. Taylor, “Improved regularization of convolutional neural networks with cutout,” arXiv preprint arXiv:1708.04552, 2017.
- [53] S. Yun, D. Han, S. J. Oh, S. Chun, J. Choe, and Y. Yoo, “Cutmix: Regularization strategy to train strong classifiers with localizable features,” in Proc. International Conference on Computer Vision (ICCV), pp. 6023–6032, 2019.
- [54] E. D. Cubuk, B. Zoph, D. Mane, V. Vasudevan, and Q. V. Le, “Autoaugment: Learning augmentation policies from data,” in Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 113–123, 2019.
- [55] A. Antoniou, A. Storkey, and H. Edwards, “Data augmentation generative adversarial networks,” arXiv preprint arXiv:1711.04340, 2017.
- [56] C. Han, L. Rundo, K. Murao, Z. Á. Milacski, et al., “GAN-based multiple adjacent brain MRI slice reconstruction for unsupervised Alzheimer’s disease diagnosis,” in Proc. International Conference on Computational Intelligence methods for Bioinformatics and Biostatistics (CIBB), 2019. In press.
- [57] N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, “Dropout: A simple way to prevent neural networks from overfitting,” J. Mach. Learn. Res., vol. 15, no. 1, pp. 1929–1958, 2014.
- [58] T. Park and G. Casella, “The bayesian Lasso,” J. Am. Stat. Assoc., vol. 103, no. 482, pp. 681–686, 2008.
- [59] H. Zou and T. Hastie, “Regularization and variable selection via the elastic net,” J. Royal Stat. Soc. Ser. B Methodol., vol. 67, no. 2, pp. 301–320, 2005.
- [60] L. Breiman, “Bagging predictors,” Mach. Learn., vol. 24, no. 2, pp. 123–140, 1996.
- [61] Y. Freund, R. Schapire, and N. Abe, “A short introduction to boosting,” J. Jap. Soc. Art., vol. 14, no. 771-780, p. 1612, 1999.
- [62] R. Hou, D. Zhou, R. Nie, D. Liu, and X. Ruan, “Brain CT and MRI medical image fusion using convolutional neural networks and a dual-channel spiking cortical model,” Med. Biol. Eng. Comput., vol. 57, no. 4, pp. 887–900, 2019.
- [63] H. Chen, Y. Qi, Y. Yin, T. Li, X. Liu, et al., “Mmfnet: A multi-modality MRI fusion network for segmentation of nasopharyngeal carcinoma,” arXiv preprint arXiv:1812.10033, 2018.
- [64] H. Shin, H. R. Roth, M. Gao, L. Lu, Z. Xu, et al., “Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning,” IEEE Trans. Med. imaging, vol. 35, no. 5, pp. 1285–1298, 2016.
- [65] J. Zbontar and Y. LeCun, “Stereo matching by training a convolutional neural network to compare image patches.,” J. Mach. Learn. Res., vol. 17, no. 1-32, p. 2, 2016.
- [66] B. Romera-Paredes and P. Torr, “An embarrassingly simple approach to zero-shot learning,” in Proc. International Conference on Machine Learning (ICML), pp. 2152–2161, 2015.
- [67] A. Graves, G. Wayne, and I. Danihelka, “Neural Turing machines,” arXiv preprint arXiv:1410.5401, 2014.
- [68] C. Finn, P. Abbeel, and S. Levine, “Model-agnostic meta-learning for fast adaptation of deep networks,” in Proc. International Conference on Machine Learning (ICML), pp. 1126–1135, 2017.
- [69] C. Han, L. Rundo, K. Murao, T. Nemoto, and H. Nakayama, “Bridging the gap between ai and healthcare sides: towards developing clinically relevant ai-powered diagnosis systems,” in Proc. IFIP International Conference on Artificial Intelligence Applications and Innovations (AIAI), pp. 320–333, 2020.
- [70] E. Wu, K. Wu, D. Cox, and W. Lotter, “Conditional infilling GANs for data augmentation in mammogram classification,” in Image Analysis for Moving Organ, Breast, and Thoracic Images, pp. 98–106, Springer, 2018.
- [71] A. Gupta, S. Venkatesh, S. Chopra, and C. Ledig, “Generative image translation for data augmentation of bone lesion pathology,” in Proc. International Conference on Medical Imaging with Deep Learning (MIDL) arXiv preprint arXiv:1902.02248, 2019.
- [72] T. Malygina, E. Ericheva, and I. Drokin, “Data augmentation with GAN: Improving chest X-ray pathologies prediction on class-imbalanced cases,” in Proc. International Conference on Analysis of Images, Social Networks and Texts (AIST), pp. 321–334, 2019.
- [73] A. Madani, M. Moradi, A. Karargyris, and T. Syeda-Mahmood, “Chest X-ray generation and data augmentation for cardiovascular abnormality classification,” in Proc. Medical Imaging: Image Processing, vol. 10574, p. 105741M, 2018.
- [74] F. Konidaris, T. Tagaris, M. Sdraka, and A. Stafylopatis, “Generative adversarial networks as an advanced data augmentation technique for MRI data,” in Proc. International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP), pp. 48–59, 2019.
- [75] T. Kanayama, Y. Kurose, K. Tanaka, K. Aida, S. Satoh, et al., “Gastric cancer detection from endoscopic images using synthesis by GAN,” in Proc. International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), pp. 530–538, 2019.
- [76] C. Gao, S. Clark, J. Furst, and D. Raicu, “Augmenting LIDC dataset using 3D generative adversarial networks to improve lung nodule detection,” in Proc. Medical Imaging: Computer-Aided Diagnosis, vol. 10950, p. 109501K, 2019.
- [77] H. C. Shin, N. A. Tenenholtz, J. K. Rogers, C. G. Schwarz, M. L. Senjem, et al., “Medical image synthesis for data augmentation and anonymization using generative adversarial networks,” in International Workshop on Simulation and Synthesis in Medical Imaging (SASHIMI), pp. 1–11, 2018.
- [78] D. Jin, Z. Xu, Y. Tang, A. P. Harrison, and D. J. Mollura, “CT-realistic lung nodule simulation from 3D conditional generative adversarial networks for robust lung segmentation,” in Proc. International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), pp. 732–740, 2018.
- [79] S. Wang, “Competencies and experiential requirements in radiology training,” in Radiology Education, pp. 55–66, Springer, 2012.
- [80] T. Ching, D. S. Himmelstein, B. K. Beaulieu-Jones, A. A. Kalinin, B. T. Do, et al., “Opportunities and obstacles for deep learning in biology and medicine,” J. R. Soc. Interface, vol. 15, no. 141, p. 20170387, 2018.
- [81] M. J. M. Chuquicusma, S. Hussein, J. Burt, and U. Bagci, “How to fool radiologists with generative adversarial networks? A visual Turing test for lung cancer diagnosis,” in IEEE International Symposium on Biomedical Imaging (ISBI), pp. 240–244, 2018.
- [82] F. Calimeri, A. Marzullo, C. Stamile, and G. Terracina, “Biomedical data augmentation using generative adversarial neural networks,” in Proc. International Conference on Artificial Neural Networks (ICANN), pp. 626–634, 2017.
- [83] M. Arjovsky, S. Chintala, and L. Bottou, “Wasserstein generative adversarial networks,” in Proc. International Conference on Machine Learning (ICML), pp. 214–223, 2017.
- [84] G. Kwon, C. Han, and D.-s. Kim, “Generation of 3D brain MRI using auto-encoding generative adversarial networks,” in Proc. International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), pp. 118–126, 2019.
- [85] L. Rundo, C. Militello, G. Russo, S. Vitabile, M. C. Gilardi, and G. Mauri, “GTVcut for neuro-radiosurgery treatment planning: An MRI brain cancer seeded image segmentation method based on a cellular automata model,” Nat. Comput., pp. 1–16, 2017.
- [86] D. Shen, G. Wu, and H. I. Suk, “Deep learning in medical image analysis,” Annu. Rev. Biomed. Eng., vol. 19, pp. 221–248, 2017.
- [87] M. Havaei, A. Davy, D. Warde-Farley, A. Biard, A. Courville, et al., “Brain tumor segmentation with deep neural networks,” Med. Image Anal., vol. 35, pp. 18–31, 2017.
- [88] D. Ravì, C. Wong, F. Deligianni, M. Berthelot, J. Andreu-Perez, et al., “Deep learning for health informatics,” IEEE J. Biomed. Health Inform., vol. 21, no. 1, pp. 4–21, 2017.
- [89] O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional networks for biomedical image segmentation,” in Proc. International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), pp. 234–241, 2015.
- [90] F. Milletari, N. Navab, and S. Ahmadi, “V-Net: Fully convolutional neural networks for volumetric medical image segmentation,” in Proc. International Conference on 3D Vision (3DV), pp. 565–571, 2016.
- [91] M. Prastawa, E. Bullitt, and G. Gerig, “Simulation of brain tumors in MR images for evaluation of segmentation efficacy,” Med. Image Anal., vol. 13, no. 2, pp. 297–311, 2009.
- [92] T. Salimans, I. Goodfellow, W. Zaremba, V. Cheung, A. Radford, and X. Chen, “Improved techniques for training GANs,” in Proc. Advances in Neural Information Processing Systems (NIPS), pp. 2234–2242, 2016.
- [93] B. H. Menze, A. Jakab, S. Bauer, J. Kalpathy-Cramer, K. Farahani, et al., “The multimodal brain tumor image segmentation benchmark (BRATS),” IEEE Trans. Med. Imaging, vol. 34, no. 10, pp. 1993–2024, 2015.
- [94] D. Lopez-Paz and M. Oquab, “Revisiting classifier two-sample tests,” in Proc. International Conference on Learning Representations (ICLR), arXiv preprint arXiv:1610.06545, 2017.
- [95] C. Szegedy, S. Ioffe, V. Vanhoucke, and A. A. Alemi, “Inception-v4, inception-resnet and the impact of residual connections on learning,” in Proc. AAAI Conference on Artificial Intelligence (AAAI), 2017.
- [96] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770–778, 2016.
- [97] L. Rundo, C. Han, Y. Nagano, J. Zhang, R. Hataya, et al., “USE-Net: Incorporating squeeze-and-excitation blocks into U-Net for prostate zonal segmentation of multi-institutional MRI datasets,” Neurocomputing, vol. 365, pp. 31–43, 2019.
- [98] J. Ker, L. Wang, J. Rao, and T. Lim, “Deep learning applications in medical image analysis,” IEEE Access, vol. 6, pp. 9375–9389, 2017.
- [99] X. Yi, E. Walia, and P. Babyn, “Generative adversarial network in medical imaging: A review,” Med. Image Anal., vol. 58, p. 101552, 2019.
- [100] L. van der Maaten and G. Hinton, “Visualizing data using t-SNE,” J. Mach. Learn. Res., vol. 9, pp. 2579–2605, 2008.
- [101] S. Koley, A. K. Sadhu, P. Mitra, B. Chakraborty, and C. Chakraborty, “Delineation and diagnosis of brain tumors from post contrast T1-weighted MR images using rough granular computing and random forest,” Appl. Soft Comput., vol. 41, pp. 453–465, 2016.
- [102] D. P. Kingma and J. L. Ba, “Adam: A method for stochastic optimization,” in Proc. International Conference on Learning Representations (ICLR), arXiv preprint arXiv:1412.6980, 2015.
- [103] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” in Advances in Neural Information Processing Systems (NIPS), pp. 1097–1105, 2012.
- [104] M. Y. Liu, T. Breuel, and J. Kautz, “Unsupervised image-to-image translation networks,” in Advances in Neural Information Processing Systems (NIPS), pp. 700–708, 2017.
- [105] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in Proc. International Conference on Learning Representations (ICLR), arXiv preprint arXiv:1409.1556, 2015.
- [106] X. Mao, Q. Li, H. Xie, R. Y. K. Lau, Z. Wang, and S. Paul Smolley, “Least squares generative adversarial networks,” in Proc. IEEE International Conference on Computer Vision (ICCV), pp. 2794–2802, 2017.
- [107] D. Ulyanov, A. Vedaldi, and V. Lempitsky, “Improved texture networks: Maximizing quality and diversity in feed-forward stylization and texture synthesis,” in Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6924–6932, 2017.
- [108] X. Huang and S. Belongie, “Arbitrary style transfer in real-time with adaptive instance normalization,” in Proc. IEEE International Conference on Computer Vision (ICCV), pp. 1501–1510, 2017.
- [109] L. Bottou, “Large-scale machine learning with stochastic gradient descent,” in Proc. International Conference on Computational Statistic (COMPSTAT), pp. 177–186, 2010.
- [110] S. Bianco, R. Cadene, L. Celona, and P. Napoletano, “Benchmark analysis of representative deep neural network architectures,” IEEE Access, vol. 6, pp. 64270–64277, 2018.
- [111] K. Chang, H. X. Bai, H. Zhou, C. Su, W. L. Bi, et al., “Residual convolutional neural network for the determination of IDH status in low-and high-grade gliomas from MR imaging,” Clin. Cancer Research, vol. 24, no. 5, pp. 1073–1081, 2018.
- [112] M. A. Mazurowski, M. Buda, A. Saha, and M. R. Bashir, “Deep learning in radiology: An overview of the concepts and a survey of the state of the art with focus on MRI,” J. Magn. Reson. Imaging, vol. 49, no. 4, pp. 939–954, 2019.
- [113] Q. McNemar, “Note on the sampling error of the difference between correlated proportions or percentages,” Psychometrika, vol. 12, no. 2, pp. 153–157, 1947.
- [114] S. Holm, “A simple sequentially rejective multiple test procedure,” Scand. J. Statist., pp. 65–70, 1979.
- [115] R. Geirhos, P. Rubisch, C. Michaelis, M. Bethge, F. A. Wichmann, and W. Brendel, “ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness,” in Proc. International Conference on Learning Representations (ICLR), arXiv preprint arXiv:1811.12231, 2019.
- [116] F. Iandola, M. Moskewicz, S. Karayev, R. Girshick, T. Darrell, and K. Keutzer, “Densenet: Implementing efficient convnet descriptor pyramids,” arXiv preprint arXiv:1404.1869, 2014.
- [117] D. Arthur and S. Vassilvitskii, “K-means++: The advantages of careful seeding,” in Proc. ACM-SIAM Symposium on Discrete Algorithms (SODA), pp. 1027–1035, 2007.
- [118] S. Ioffe and C. Szegedy, “Batch normalization: Accelerating deep network training by reducing internal covariate shift,” in Proc. International Conference on Learning Representations (ICLR) arXiv preprint arXiv:1502.03167, 2015.
- [119] S. Vandenhende, B. De Brabandere, D. Neven, and L. Van Gool, “A three-player GAN: Generating hard samples to improve classification networks,” in Proc. International Conference on Machine Vision Applications (MVA) arXiv preprint arXiv:1903.03496, 2019.
- [120] V. Gulshan, L. Peng, M. Coram, M. C. Stumpe, D. Wu, et al., “Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs,” JAMA, vol. 316, no. 22, pp. 2402–2410, 2016.
- [121] N. D. Arvold, E. Q. Lee, M. P. Mehta, K. Margolin, B. M. Alexander, et al., “Updates in the management of brain metastases,” Neuro Oncol., vol. 18, no. 8, pp. 1043–1065, 2016.
- [122] O. Bailo, D. Ham, and Y. Min Shin, “Red blood cell image generation for data augmentation using conditional generative adversarial networks,” in Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2019.
- [123] J. Redmon and A. Farhadi, “YOLOv3: An incremental improvement,” arXiv preprint arXiv:1804.02767, 2018.
- [124] S. Ren, K. He, R. Girshick, and J. Sun, “Faster R-CNN: Towards real-time object detection with region proposal networks,” in Advances in Neural Information Processing Systems (NIPS), pp. 91–99, 2015.
- [125] T. G. Dietterich, “Ensemble learning,” The Handbook of Brain Theory and Neural Networks, vol. 2, pp. 110–125, 2002.
- [126] Z. Xu, X. Wang, H. Shin, D. Yang, H. Roth, et al., “Correlation via synthesis: End-to-end nodule image generation and radiogenomic map learning based on generative adversarial network,” arXiv preprint arXiv:1907.03728, 2019.
- [127] R. L. Siegel, K. D. Miller, and A. Jemal, “Cancer statistics, 2019,” CA-Cancer J. Clin., vol. 69, no. 1, pp. 7–34, 2019.
- [128] S. G. Armato III, G. McLennan, L. Bidaut, M. F. McNitt-Gray, C. R. Meyer, et al., “The lung image database consortium (LIDC) and image database resource initiative (IDRI): A completed reference database of lung nodules on CT scans,” Med. Phys., vol. 38, no. 2, pp. 915–931, 2011.
- [129] A. A. A. Setio, A. Traverso, T. De Bel, M. S. N. Berens, C. van den Bogaard, et al., “Validation, comparison, and combination of algorithms for automatic detection of pulmonary nodules in computed tomography images: The LUNA16 challenge,” Med. Image Anal., vol. 42, pp. 1–13, 2017.
- [130] Ö. Çiçek, A. Abdulkadir, S. S. Lienkamp, T. Brox, and O. Ronneberger, “3D U-Net: Learning dense volumetric segmentation from sparse annotation,” in Proc. International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), pp. 424–432, 2016.
- [131] X. Ouyang, Y. Cheng, Y. Jiang, C. L. Li, and P. Zhou, “Pedestrian-synthesis-GAN: Generating pedestrian data in real scene and beyond,” arXiv preprint arXiv: 1804.02047, 2018.
- [132] M. Niemeijer, M. Loog, M. D. Abramoff, M. A. Viergever, M. Prokop, and B. van Ginneken, “On combining computer-aided detection systems,” IEEE Trans. Med. Imaging, vol. 30, no. 2, pp. 215–223, 2011.
- [133] I. E. Allen and C. A. Seaman, “Likert scales and data analyses,” Qual. Prog., vol. 40, no. 7, pp. 64–65, 2007.
- [134] N. Wu, J. Phang, J. Park, Y. Shen, Z. Huang, et al., “Deep neural networks improve radiologists’ performance in breast cancer screening,” in Proc. International Conference on Medical Imaging with Deep Learning (MIDL), arXiv preprint arXiv:1907.08612, 2019.
- [135] S. M. McKinney, M. Sieniek, V. Godbole, J. Godwin, N. Antropova, et al., “International evaluation of an AI system for breast cancer screening,” Nature, vol. 577, no. 7788, pp. 89–94, 2020.
- [136] A. Adadi and M. Berrada, “Peeking inside the black-box: A survey on Explainable Artificial Intelligence (XAI),” IEEE Access, vol. 6, pp. 52138–52160, 2018.
- [137] M. D. Abràmoff, P. T. Lavin, M. Birch, N. Shah, and J. C. Folk, “Pivotal trial of an autonomous AI-based diagnostic system for detection of diabetic retinopathy in primary care offices,” NPJ Digit. Med., vol. 1, no. 1, p. 39, 2018.
- [138] T. Nair, D. Precup, D. L. Arnold, and T. Arbel, “Exploring uncertainty measures in deep networks for multiple sclerosis lesion detection and segmentation,” Med. Image Anal., vol. 59, p. 101557, 2020.
- [139] G. R. Jankharia, “Commentary-radiology in India: The next decade,” Indian J. Radiol. Imaging, vol. 18, no. 3, p. 189, 2008.
- [140] S. Vollmer, B. A. Mateen, G. Bohner, F. J. Király, R. Ghani, et al., “Machine learning and AI research for patient benefit: 20 critical questions on transparency, replicability, ethics and effectiveness,” arXiv preprint arXiv:1812.10404, 2018.
- [141] D. O’Connor, N. V. Potler, M. Kovacs, T. Xu, L. Ai, et al., “The healthy brain network serial scanning initiative: A resource for evaluating inter-individual differences and their reliabilities across scan conditions and sessions,” Gigascience, vol. 6, no. 2, p. giw011, 2017.
- [142] M. E. Vandenberghe, M. L. J. Scott, P. W. Scorer, M. Söderberg, D. Balcerzak, and C. Barker, “Relevance of deep learning to facilitate the diagnosis of HER2 status in breast cancer,” Sci. Rep., vol. 7, p. 45938, 2017.
- [143] X. Li, Y. Wang, and D. Li, “Medical data stream distribution pattern association rule mining algorithm based on density estimation,” IEEE Access, vol. 7, pp. 141319–141329, 2019.
- [144] M. Agn, I. Law, P. M. af Rosenschöld, and K. Van Leemput, “A generative model for segmentation of tumor and organs-at-risk for radiation therapy planning of glioblastoma patients,” in Proc. Medical Imaging: Image Processing, vol. 9784, p. 97841D, 2016.
- [145] K. R. Abi-Aad, B. J. Anderies, M. E. Welz, and B. R. Bendok, “Machine learning as a potential solution for shift during stereotactic brain surgery,” Neurosurgery, vol. 82, no. 5, pp. E102–E103, 2018.
- [146] Q. Yang, P. Yan, Y. Zhang, et al., “Low-dose CT image denoising using a generative adversarial network with Wasserstein distance and perceptual loss,” IEEE Trans. Med. Imaging, vol. 37, no. 6, pp. 1348–1357, 2018.
- [147] J. M. M. Rumbold and B. Pierscionek, “The effect of the general data protection regulation on medical research,” J. Med. Internet Res., vol. 19, no. 2, p. e47, 2017.
- [148] L. H. Sobin, M. K. Gospodarowicz, and C. Wittekind, TNM classification of malignant tumours. John Wiley & Sons, 2011.
- [149] K. Nawata, A. Matsumoto, R. Kajihara, and M. Kimura, “Evaluation of the distribution and factors affecting blood pressure using medical checkup data in Japan,” Health, vol. 9, no. 1, pp. 124–137, 2016.
- [150] R. P. Mansour, “Visual charting method for creating electronic medical documents,” Apr. 16 2019. US Patent 10,262,106.
- [151] M. Ghafoorian, A. Mehrtash, T. Kapur, N. Karssemeijer, E. Marchiori, et al., “Transfer learning for domain adaptation in MRI: Application in brain lesion segmentation,” in Proc. International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), pp. 516–524, 2017.
- [152] A. Chen, Z. Zhang, Q. Li, W. Jiang, Q. Zheng, et al., “Feasibility study for implementation of the AI-powered Internet+ Primary Care Model (AiPCM) across hospitals and clinics in Gongcheng county, Guangxi, China,” Lancet, vol. 394, p. S44, 2019.
- [153] A. Laplante-Lévesque, H. Abrams, M. Bülow, T. Lunner, J. Nelson, et al., “Hearing device manufacturers call for interoperability and standardization of internet and audiology,” Am. J. Audiol., vol. 25, no. 3S, pp. 260–263, 2016.
- [154] J. Morley, M. Taddeo, and L. Floridi, “Google Health and the NHS: Overcoming the trust deficit,” Lancet Digit. Health, vol. 1, no. 8, p. e389, 2019.
- [155] G. Rossini, S. Parrini, T. Castroflorio, A. Deregibus, and C. L. Debernardi, “Diagnostic accuracy and measurement sensitivity of digital models for orthodontic purposes: A systematic review,” Am. J. Orthod. Dentofacial Orthop., vol. 149, no. 2, pp. 161–170, 2016.
- [156] K. Huang, H. Cheng, Y. Zhang, et al., “Medical knowledge constrained semantic breast ultrasound image segmentation,” in Proc. International Conference on Pattern Recognition (ICPR), pp. 1193–1198, 2018.
- [157] C. Krittanawong, “The rise of artificial intelligence and the uncertain future for physicians,” Eur. J. Intern. Med., vol. 48, pp. e13–e14, 2018.
- [158] H. Li, G. Jiang, J. Zhang, W. R., W. Z., et al., “Fully convolutional network ensembles for white matter hyperintensities segmentation in MR images,” NeuroImage, vol. 183, pp. 650–665, 2018.
- [159] A. Jain, S. Ratnoo, and D. Kumar, “Addressing class imbalance problem in medical diagnosis: A genetic algorithm approach,” in Proc. International Conference on Information, Communication, Instrumentation and Control (ICICIC), pp. 1–8, 2017.
- [160] S. A. Wartman and C. D. Combs, “Reimagining medical education in the age of AI,” AMA J. Ethics, vol. 21, no. 2, pp. 146–152, 2019.
- [161] L. Rundo, C. Han, J. Zhang, R. Hataya, Y. Nagano, et al., “CNN-based prostate zonal segmentation on T2-weighted MR images: A cross-dataset study,” in Neural Approaches to Dynamics of Signal Exchanges, Smart Innovation, Systems and Technologies, pp. 269–280, Springer, 2019.
- [162] E. H. P. Pooch, P. L. Ballester, and R. C. Barros, “Can we trust deep learning models diagnosis? The impact of domain shift in chest radiograph classification,” arXiv preprint arXiv:1909.01940, 2019.
- [163] P. Pérez, M. Gangnet, and A. Blake, “Poisson image editing,” ACM TOG, vol. 22, no. 3, pp. 313–318, 2003.
- [164] C. H. Lin, C. Chang, Y. Chen, D. Juan, W. Wei, and H. Chen, “COCO-GAN: Generation by parts via conditional coordinating,” in Proc. International Conference on Computer Vision (ICCV), pp. 4512–4521, 2019.
- [165] S. Chen, K. Ma, and Y. Zheng, “Med3D: Transfer learning for 3D medical image analysis,” arXiv preprint arXiv:1904.00625, 2019.
- [166] R. K. Samala, H. Chan, L. Hadjiiski, M. A. Helvie, J. Wei, and K. Cha, “Mass detection in digital breast tomosynthesis: Deep convolutional neural network with transfer learning from mammography,” Med. Phys., vol. 43, no. 12, pp. 6654–6666, 2016.
Appendix A Scientific Production
A.1 Related Publications/Presentations
Journal papers
-
•
C. Han, L. Rundo, R. Araki, Y. Nagano, Y. Furukawa, G. Mauri, H. Nakayama, H. Hayashi, Combining Noise-to-Image and Image-to-Image GANs: Brain MR Image Augmentation for Tumor Detection, IEEE Access, October 2019 (Project 2).
-
•
C. Han, K. Murao, S. Satoh, H. Nakayama, Learning More with Less: GAN-based Medical Image Augmentation, Medical Imaging Technology, Japanese Society of Medical Imaging Technology, June 2019 (Tutorial Paper).
Book chapter
-
•
C. Han, L. Rundo, R. Araki, Y. Furukawa, G. Mauri, H. Nakayama, H. Hayashi, Infinite Brain MR Images: PGGAN-based Data Augmentation for Tumor Detection, In A. Esposito, M. Faundez-Zanuy, F. C. Morabito, E. Pasero (eds.) Neural Approaches to Dynamics of Signal Exchanges, Springer, September 2019 (Project 2).
Conference proceedings
-
•
C. Han, L. Rundo, K. Murao, T. Nemoto, H. Nakayama, Bridging the gap between AI and healthcare sides: towards developing clinically relevant AI-powered diagnosis systems, In IFIP International Conference on Artificial Intelligence Applications and Innovations (AIAI), pp. 320–333, June 2020
(Discussion Paper). -
•
C. Han, K. Murao, T. Noguchi, Y. Kawata, F. Uchiyama, L. Rundo, H. Nakayama, S. Satoh, Learning More with Less: Conditional PGGAN-based Data Augmentation for Brain Metastases Detection Using Highly-Rough Annotation on MR Images, In ACM International Conference on Information and Knowledge Management (CIKM), Beijing, China, November 2019 (Project 3).
-
•
C. Han, Y. Kitamura, A. Kudo, A. Ichinose, L. Rundo, Y. Furukawa, K. Umemoto, H. Nakayama, Y. Li, Synthesizing Diverse Lung Nodules Wherever Massively: 3D Multi-Conditional GAN-based CT Image Augmentation for Object Detection, In International Conference on 3D Vision (3DV), Québec City, Canada, September 2019 (Project 4).
-
•
C. Han, H. Hayashi, L. Rundo, R. Araki, Y. Furukawa, W. Shimoda, S. Muramatsu, G. Mauri, H. Nakayama, GAN-based Synthetic Brain MR Image Generation, In IEEE International Symposium on Biomedical Imaging (ISBI), Washington, D.C., The United States, April 2018 (Project 1).
A.2 Other Publications/Presentations
Journal paper
-
•
L. Rundo*, C. Han*, Y. Nagano, J. Zhang, R. Hataya, C. Militello, A. Tangherloni, M. S. Nobile, C. Ferretti, D. Besozzi, M. C. Gilardi, S. Vitabile, G. Mauri, H. Nakayama, P. Cazzaniga, USE-Net: incorporating Squeeze-and-Excitation blocks into U-Net for prostate zonal segmentation of multi-institutional MRI datasets, Neurocomputing, July 2019 (* denotes co-first authors).
Book chapters
-
•
L. Rundo, C. Han, J. Zhang, R. Hataya, Y. Nagano, C. Militello, C. Ferretti, M. S. Nobile, A. Tangherloni, M. C. Gilardi, S. Vitabile, H. Nakayama, G. Mauri, CNN-based Prostate Zonal Segmentation on T2-weighted MR Images: A Cross-dataset Study, A. Esposito, M. Faundez-Zanuy, F. C. Morabito, E. Pasero (eds.) Neural Approaches to Dynamics of Signal Exchanges, Springer, September 2019.
-
•
C. Han, K. Tsuge, H. Iba, Application of Learning Classifier Systems to Gene Expression Analysis in Synthetic Biology, In S. Patnaik, X. Yang, and K. Nakamatsu (eds.) Nature Inspired Computing and Optimization: Theory and Applications, Springer, March 2017.
Conference proceedings
-
•
K. Murao, Y. Ninomiya, C. Han, K. Aida, S. Satoh, Cloud platform for deep learning-based CAD via collaboration between Japanese medical societies and institutes of informatics, In SPIE Medical Imaging (oral presentation), Houston, The United States, February 2020.
-
•
C. Han, L. Rundo, K. Murao, Z. Á. Milacski, K. Umemoto, H. Nakayama, S. Satoh, GAN-based Multiple Adjacent Brain MRI Slice Reconstruction for Unsupervised Alzheimer’s Disease Diagnosis, In Computational Intelligence methods for Bioinformatics and Biostatistics (CIBB), Bergamo, Italy, September 2019.
-
•
C. Han, K. Tsuge, H. Iba, Optimization of Artificial Operon Construction by Consultation Algorithms Utilizing LCS, In IEEE Congress on Evolutionary Computation (CEC), Vancouver, Canada, July 2016.