This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Knowledge-based automated planning with three-dimensional generative adversarial networks

Aaron Babier Department of Mechanical & Industrial Engineering, University of Toronto Rafid Mahmood Department of Mechanical & Industrial Engineering, University of Toronto Andrea McNiven Radiation Medicine Program, Princess Margaret Cancer Centre Adam Diamant Schulich School of Business, York University Timothy C. Y. Chan Department of Mechanical & Industrial Engineering, University of Toronto
({ababier,rmahmood}@mie.utoronto.ca, andrea.mcniven@rmp.uhn.ca, adiamant@schulich.yorku.ca, tcychan@mie.utoronto.ca)
Abstract

Purpose: To develop a knowledge-based automated planning pipeline that generates treatment plans using deep neural network architectures for predicting 3D doses.

Methods: Our knowledge-based automated planning (KBAP) pipeline consisted of a generative adversarial network (GAN) to predict dose from a CT image followed by two optimization models to learn objective function weights and generate fluence-based plans, respectively. We investigated three different GAN models. The first two models predicted dose for each axial slice independently. One predicted dose as a red-green-blue (three-channel) color map, while the other predicted a scalar value (one-channel) for dose directly. The third GAN model predicted scalar doses for the full 3D CT image at once, considering correlations between adjacent CT slices. For all models, we also investigated the impact of scaling the GAN predictions before optimization. Each GAN model was trained on 130 previously delivered oropharyngeal treatment plans. Performance was tested on 87 out-of-sample plans, by evaluating using clinical planning criteria and compared to their corresponding clinical plans. KBP prediction quality was assessed using dose-volume histogram (DVH) differences from the corresponding KBAP plans.

Results: The best performing KBAP plans were generated with the 3D GAN, which predicted one channel dose values followed by scaling. These plans satisfied close to 77% of all clinical criteria, compared to the clinical plans, which satisfied 64% of all criteria. Additionally, these KBAP plans satisfied the same criteria as the clinical plans 84% more frequently compared to the 2D GAN model using three channel dose prediction. The 3D GAN predictions were also more similar to the final plan compared to the other approaches, as it better captured the vertical dosimetric relationship between adjacent axial slices. The deliverable plans were also more similar to their predictions than plans generated using any other GAN-based approach and they better captured implicit constraints associated with physical deliverability.

Conclusion: We developed the first knowledge-based automated planning framework using a 3D generative adversarial network for prediction. Our results based on 217 oropharyngeal cancer treatment plans demonstrated superior performance in satisfying clinical criteria and generated more realistic predictions compared to the previous state-of-the-art.

1 Introduction

The conventional radiation therapy treatment planning process consists of an iterative, back-and-forth procedure between a treatment planner and an oncologist. The duration of a single iteration, compounded by the number of iterations that may take place, means that it can take several days for a treatment plan to be completed [Das et al., 2009]. Automated treatment planning systems have the potential to replace this conventional approach with a more efficient operational paradigm that reduces plan generation lead time [Purdie et al., 2014]. Hospitals that adopt these techniques should also be better equipped to efficiently produce high-quality treatment plans for complicated sites [Ziemer et al., 2017] and serve the growing demand for radiation therapy [Atun et al., 2015].

Knowledge-based automated planning (KBAP) is a two-step approach to automated radiation therapy treatment planning that first predicts a clinically acceptable dose before using optimization to convert the prediction into a deliverable plan [McIntosh and Purdie, 2017, Wu et al., 2017, Babier et al., 2018b, Fan et al., 2018]. The prediction component of the pipeline, referred to as knowledge-based planning (KBP), typically uses a library of historical treatment plans to learn the dose characteristics of previously delivered plans. It is essential that this prediction model be accurate as the the final plan quality strongly correlates with prediction quality [Babier et al., 2018b].

Many KBP approaches have been introduced in the literature with the objective of either predicting a dose distribution or a dose volume histogram (DVH) [Zhu et al., 2011, Yuan et al., 2012, Appenzoller et al., 2012, Yang et al., 2013, Shiraishi et al., 2015, Shiraishi and Moore, 2016, Younge et al., 2018, Babier et al., 2018a, Fan et al., 2018]. While the majority of these methods use hand-tailored or low dimensional features for prediction, recent advances in machine learning have spurred the development of KBP methods that predict full dose distributions using automatically generated, high-dimensional features [McIntosh and Purdie, 2016, Nguyen et al., 2017, Mahmood et al., 2018, Fan et al., 2018]. The most recent work in this space has focused on neural network-based KBP methods, including convolutional neural networks [Nguyen et al., 2017], generative adversarial neural networks (GANs) [Mahmood et al., 2018], and recurrent neural networks [Fan et al., 2018]. These models are trained on a library of historical plans and then predict 2D dose distributions for axial slices of a CT image. A 3D dose distribution is formed by concatenation.

In this paper, we develop the first 3D GAN-based KBP method, which takes as input a 3D image and simultaneously predicts the full 3D dose distribution. We embed our prediction model in a KBAP pipeline for oropharynx treatment planning [Babier et al., 2018a]. Unlike previously developed KBP methods, our approach uses a patient’s entire 3D CT image as input and learns to construct spatial features without human intervention. In doing so, it learns to produce the entire 3D dose distribution rather than separate 2D dose distributions for each axial slice. Further, we improve upon a previously developed 2D approach by specializing the GAN to the radiation therapy context: instead of generating an image of the dose, the GAN predicts the dose to be delivered to each voxel directly. We also investigate the effect of scaling the predictions before optimization.

We apply our approach to a dataset of CT images from 217 patients with oropharyngeal cancer that have undergone radiation therapy. Approximately 60% of these images are used to train our 3D GAN, which is then used to predict the dose distribution for the remaining out-of-sample patients. The predictions are used as input into an optimization model to generate treatment plans [Babier et al., 2018b]. We demonstrate, over several computational experiments, that our KBAP-specific modifications of (i) predicting the full 3D dose distribution; (ii) representing dose to each voxel as a scalar versus a red-green-blue (RGB) color map; and (iii) scaling the predictions before optimization, results in better out-of-sample performance than state-of-the-art benchmarks.

2 Methods and Materials

We use contoured CT images and dose distributions from clinically accepted treatment plans to train three GAN models in the KBAP pipeline. Each GAN was trained to predict the dose distribution given a contoured CT image. For testing, we passed out-of-sample CT images through each of the GANs to generate dose distributions. These predictions were then passed through an optimization pipeline [Babier et al., 2018b] to generate the final fluence-based treatment plans. Figure 1 shows a high-level overview of this automated planning pipeline.

Refer to caption
Figure 1: Overview of the knowledge-based automated treatment planning pipeline.

2.1 Prediction Using Generative Adversarial Networks

A GAN consists of two neural networks known as a generator and discriminator[Goodfellow et al., 2014] We focus on conditional GANs which, in addition to a Gaussian input, learn to generate different outputs conditioned on known problem-specific characteristics (e.g., CT images) [Isola et al., 2017]. Specifically, let 𝐳p𝐳\mathbf{z}\sim p_{\mathbf{z}} denote a sample from a Gaussian input. The generator network takes as input 𝐳\mathbf{z} and a CT image 𝐜\mathbf{c} and outputs a predicted dose distribution 𝐱=G(𝐳,𝐜)\mathbf{x}=G(\mathbf{z},\mathbf{c}). The discriminator network then takes a CT image and the predicted dose distribution as input and outputs a “belief” regarding whether the dose distribution is the actual clinical dose (as opposed to artificially produced by the generator). That is, D(𝐱,𝐜)[0,1]D(\mathbf{x},\mathbf{c})\in[0,1] where D(𝐱,𝐜)=1D(\mathbf{x},\mathbf{c})=1 suggests the discriminator is confident the dose distribution is the clinically delivered dose. Both networks are trained iteratively with a single loss function (D,G)\mathcal{L}(D,G). Letting 𝐱pdata\mathbf{x}\sim p_{\text{data}} denote the distribution of real delivered plans, we write the training problem as:

minGmaxD(D,G)\displaystyle\min_{G}\max_{D}~\mathcal{L}(D,G)

where

(D,G)\displaystyle\mathcal{L}(D,G) =𝔼𝐳p𝐳[log(1D(G(𝐳,𝐜),𝐜))]+𝔼𝐱pdata[logD(𝐱,𝐜)]+λ𝔼𝐱pdata,𝐳p𝐳[𝐱G(𝐳,𝐜)1].\displaystyle=\mathbb{E}_{\mathbf{z}\sim p_{\mathbf{z}}}\left[\log(1-D(G(\mathbf{z},\mathbf{c}),\mathbf{c}))\right]+\mathbb{E}_{\mathbf{x}\sim p_{\text{data}}}\left[\log D(\mathbf{x},\mathbf{c})\right]+\lambda\mathbb{E}_{\mathbf{x}\sim p_{\text{data}},\mathbf{z}\sim p_{\mathbf{z}}}\left[\|\mathbf{x}-G(\mathbf{z},\mathbf{c})\|_{1}\right].

The above formulation represents the objective for the most common class of conditional GANs used in problems associated with image generation [Isola et al., 2017, Zhu et al., 2017]. By minimizing the first term in (D,G)\mathcal{L}(D,G), the generator learns to construct dose distributions such that D(G(𝐳,𝐜),𝐜)=1D(G(\mathbf{z},\mathbf{c}),\mathbf{c})=1. That is, the generator attempts to fool the discriminator into believing that the generated dose is from the real clinical distribution. The discriminator adversarially maximizes the second term in (D,G)\mathcal{L}(D,G) to output D(G(𝐳,𝐜),𝐜)=0D(G(\mathbf{z},\mathbf{c}),\mathbf{c})=0 for 𝐳p𝐳\mathbf{z}\sim p_{\mathbf{z}} and D(𝐱,𝐜)=1D(\mathbf{x},\mathbf{c})=1 for 𝐱pdata\mathbf{x}\sim p_{\text{data}}, i.e., the discriminator attempts to correctly distinguish between artificially generated versus clinically delivered plans. The final term in (D,G)\mathcal{L}(D,G) is an l1l_{1} loss function which forces the generated samples to better resemble the ground truth dataset (i.e., the clinically delivered dose distribution). The hyperparameter λ\lambda balances the tradeoff between minimizing the GAN loss (first two terms) and having images resemble deliverable plans.

We constructed generator and discriminator networks derived from the pix2pix architecture proposed in the canonical Style Transfer GAN [Isola et al., 2017]. The generator possesses a U-net architecture that passes a contoured CT image through consecutive convolution layers, a bottleneck layer, and then several deconvolution layers. The U-net employs skip connections, i.e., the output of each convolution layer is concatenated to the input of a corresponding deconvolution layer. This allows the generator to easily pass “high-dimensional” information (e.g., structural outlines) between the inputted CT image and the outputted dose. The discriminator takes as input a dose distribution and CT image and passes them through several consecutive convolution layers until outputting a single scalar value between zero and one. We refer the reader to the original Style Transfer GAN work for full details on the number and size of the convolution and deconvolution layers in the pix2pix architecture [Isola et al., 2017].

We trained three GAN variants each with slightly modified architectures; we refer to them as 2D-RGB, 2D-dose, and 3D-dose. The “2D” or “3D” designation refers to whether dose is being predicted for 2D slices individually or the full 3D image at once. The “-RGB” or “-dose” designation refers to the input/output channels of the GAN, i.e., whether the output of the generator is a 3-dimensional vector of numbers representing the red-green-blue scale, or a 1-dimensional scalar that represents dose directly (i.e., grayscale). The 2D-RGB GAN is a direct replica of the original pix2pix network and has been previously implemented and tested for KBAP [Mahmood et al., 2018]. It takes as input a single axial slice of the contoured CT image, i.e., a 128×128128\times 128 pixel image with three channels for color, and outputs a three channel 128×128128\times 128 pixel image that represents dose. Two-dimensional convolution and deconvolution filters are used for the generator and discriminator. The 2D-dose GAN has an almost identical architecture except the output of the generator is a one channel 128×128128\times 128 image. In both 2D GANs, a 3D dose distribution is obtained by concatenating all outputted axial slices for a single patient. In contrast, the 3D-dose GAN takes an entire contoured CT image (128×128×128128\times 128\times 128 voxels) and generate the corresponding 3D dose distribution. Thus, it uses three-dimensional convolution and deconvolution filters [Hermoza and Sipiran, 2017] with a one channel output to represent dose. Apart from filter dimensions, the 2D-dose and 3D-dose GANs have the same architecture including the same number of layers and filter sizes.

The generator and discriminator networks were trained iteratively using gradient descent. After training was complete, the discriminator was disconnected and the generator was used on out-of-sample CT images. Figure 2 summarizes the difference between how the GANs were trained and tested. In our experiments, we used the loss function given by (D,G)\mathcal{L}(D,G) with λ=90\lambda=90, and trained the networks using the Adam optimizer [Kingma and Ba, 2014] with learning rate α=0.0002\alpha=0.0002 and β1=0.5\beta_{1}=0.5 and β2=0.999\beta_{2}=0.999. These hyperparameters are the default Adam settings and have been used in many style transfer problems [Isola et al., 2017]. We also performed a parameter sweep for λ\lambda and found the default setting sufficient. We stopped training when the standard GAN loss function 𝔼𝐱pdata[logD(𝐱,𝐜)]+𝔼𝐳p𝐳[log(1D(G(𝐳,𝐜),𝐜))]\mathbb{E}_{\mathbf{x}\sim p_{\text{data}}}\left[\log D(\mathbf{x},\mathbf{c})\right]+\mathbb{E}_{\mathbf{z}\sim p_{\mathbf{z}}}\left[\log(1-D(G(\mathbf{z},\mathbf{c}),\mathbf{c}))\right] was roughly equal to the regularization term λ𝔼𝐱pdata,𝐳p𝐳[𝐱G(𝐳,𝐜)1]\lambda\mathbb{E}_{\mathbf{x}\sim p_{\text{data}},\mathbf{z}\sim p_{\mathbf{z}}}\left[\|\mathbf{x}-G(\mathbf{z},\mathbf{c})\|_{1}\right]; if the loss from the l1l_{1} penalty fell too low, the GAN began to memorize the dataset. Therefore, the 2D-dose and 2D-RGB GANs were trained for 5050 epochs and the 3D-dose GAN was trained for 200200 epochs. It is natural for the 3D-dose GAN to require more epochs to converge as it contains significantly more parameters and generates the entire dose distribution. The code for all experiments is provided at http://github.com/rafidrm/gancer.

Refer to caption
Figure 2: Overview of the GAN training and testing phases.

2.2 Training the GAN

We obtained plans for 217 oropharyngeal cancer treatments delivered at a single institution with 6 MV, step-and-shoot, intensity-modulated radiation therapy. Plans were prescribed 70 Gy and 56 Gy in 35 fractions to the gross disease (PTV70) and elective target volumes (PTV56), respectively; in 130 plans there was also a prescription of 63 Gy to the intermediate risk target volume (PTV63). The organs-at-risk (OARs) included the brainstem, spinal cord, right parotid, left parotid, larynx, esophagus, mandible, and the limPostNeck, which is an artificial structure for sparing the posterior neck. Patient geometry was discretized into voxels of size 44 mm ×\times 44 mm ×\times 22 mm.

The CT images and dose distributions for all 217 treatment plans were converted into a suitable format for use by the GAN architectures. The CT images were reconstructed so that each voxel had RGB channels, which were assigned values according to Table 1, and converted into 128 axial slices of 128×128128\times 128 voxels. The images were separated into a training set of 130 random samples (a total of 15,657 pairs of CT image slices and dose distributions) and a testing set of the remaining 87 samples for out-of-sample evaluation.

Table 1: Colors assigned to each voxel. Voxels that were classified as both OAR and target were assigned nonzero green and red channel values, respectively.
Structure Red Channel Green Channel Blue Channel
Brainstem 0 125 CT Grayscale
Spinal Cord 0 147 CT Grayscale
Right Parotid 0 190 CT Grayscale
Left Parotid 0 190 CT Grayscale
Larynx 0 233 CT Grayscale
Esophagus 0 212 CT Grayscale
Mandible 0 168 CT Grayscale
limPostNeck 0 255 CT Grayscale
PTV70 255 0 CT Grayscale
PTV63 205 0 CT Grayscale
PTV56 155 0 CT Grayscale
Unclassified 0 0 CT Grayscale
Empty Space 0 0 0

2.3 Creating Plans Using Inverse Planning

During out-of-sample testing, predictions produced by the generator were used as input into an optimization model with two stages. In the first stage, given a predicted dose distribution, the objective weights for a standard inverse planning model were estimated using a parameter estimation technique that has been previously validated in oropharynx [Babier et al., 2018b]. In the second stage, the estimated weights were used in an inverse planning optimization model to generate treatment plans. The objective minimized the sum of 65 functions: seven per OAR and three per target. In our experiments, objectives for the OARs included the mean dose, max dose, and the average dose above 0.25, 0.50, 0.75, 0.90, and 0.975 of the maximum predicted dose to the OAR. Objectives for the target included the maximum dose, average dose below prescription, and average dose above prescription. The complexity of all generated treatment plans was constrained to a sum-of-positive-gradients (SPG) value of 55 [Craft et al., 2007]. SPG was used since it is a convex surrogate for the physical deliverability of a plan and the parameter 55 was chosen as it is two standard deviations above the average SPG [Babier et al., 2018a]. The dose influence matrices required for the optimization model were derived with A Computational Environment for Radiotherapy Research [Deasy et al., 2003]. Each of the KBP-generated plans were delivered from nine equidistant coplanar beams at angles 0, 40, …, 320. We used Gurobi 7.5 to solve the optimization model.

We also generated plans using a scaling procedure that multiplicatively increased the entire predicted dose distribution by the smallest amount to satisfy all target dose criteria. The scaled predictions were then input into the optimization model. Note that this scaling does not affect the fairness of the analysis because the final KBAP plans must satisfy the same constraints (e.g., SPG) as plans generated from the unscaled predictions. To study the full effect of the scaling step, we generated three additional populations of scaled final plans alongside the initial three unscaled plans corresponding to 2D-RGB, 2D-dose, and 3D-dose. We refer to the three scaled KBAP plans as 2D-RGB, 2D-dose, and 3D-dose respectively.

2.4 Performance Analysis

We conducted two primary analyses. First, we evaluated the quality of the KBAP plans by computing the fraction of clinical planning criteria that were satisfied. We compared these results against the performance of the clinical plans. Second, we evaluated the quality of the KBP predictions by comparing the predicted dose distributions to clinical dose distributions, and the KBAP plan DVHs to their respective predicted DVHs.

KBAP Plan Quality: The quality of the final KBAP plans were measured by evaluating how often they satisfied the clinical criteria presented in Table 2. For each class of clinical criteria, i.e., OARs, targets, and all regions-of-interest (ROIs), which includes both OARs and targets, we generated confusion matrices to compare the KBAP plans with the clinical plans. We then analyzed the organ-specific criteria that was achieved by each clinical plan to determine whether the corresponding KBAP plan also passed that criterion.

Table 2: The planning criteria used for evaluation: 𝒟99\mathcal{D}_{99} is the dose to a fractional volume of 0.990.99, 𝒟mean\mathcal{D}_{mean} is the mean dose to a structure, and 𝒟max\mathcal{D}_{max} is the max dose to a structure.
Structure Criteria
Brainstem 𝒟max\mathcal{D}_{max}\leq 54 Gy
Spinal Cord 𝒟max\mathcal{D}_{max}\leq 48 Gy
Right Parotid 𝒟mean\mathcal{D}_{mean}\leq 26 Gy
Left Parotid 𝒟mean\mathcal{D}_{mean}\leq 26 Gy
Larynx 𝒟mean\mathcal{D}_{mean}\leq 45 Gy
Esophagus 𝒟mean\mathcal{D}_{mean}\leq 45 Gy
Mandible 𝒟max\mathcal{D}_{max}\leq 73.5 Gy
PTV70 𝒟99\mathcal{D}_{99}\ \geq 66.5 Gy
PTV63 𝒟99\mathcal{D}_{99}\ \geq 59.9 Gy
PTV56 𝒟99\mathcal{D}_{99}\ \geq 53.2 Gy

KBP Prediction Quality: We measured KBP prediction quality by evaluating how similar the KBP predictions were to their corresponding KBAP plans. For every ROI and every patient, we calculated the average absolute error between the KBP predicted DVH and the KBAP plan DVH. Differences between 2D-RGB and 2D-dose, 2D-dose and 2D-dose, and 2D-dose and 3D-dose were evaluated; each comparison attempts to isolate the impact of each of the three KBAP-specific modifications. We then evaluated the difference between 2D-RGB and 3D-dose to quantify the joint contribution of all three modifications. The distribution of the errors over the population of plans was visualized using a separate boxplot for each set of comparisons. Finally, we visualized the predicted dose distributions to determine whether the 3D model produced smoother predictions across the longitudinal axis than a 2D model.

3 Results

KBAP Plan Quality

In Table LABEL:clinCritConfusion, we present confusion matrices comparing the quality of the final KBAP plans with the clinical plans. The columns represent KBAP performance using each of the six KBP approaches while the rows indicate the clinical plans and the performance targets. Overall, 3D-dose plans best replicated the performance of the clinical plans since they agreed most on what criteria passed and failed (i.e., Pass/Pass and Fail/Fail). Where they differed, 3D-dose plans satisfied five times as many criteria (Fail/Pass) as the clinical plans (Pass/Fail). We also observed that scaling made a substantial difference as scaled plans outperformed their unscaled counterparts. For example, scaled 3D plans satisfied 99.5% of all target criteria whereas only 52.3% were satisfied for unscaled 3D plans. Finally, 2D-dose and 3D-dose performed the best satisfying 77.0% and 76.4% of all ROI criteria, respectively.

Table 3: For each KBAP approach, the percentage of clinical criteria that passed and failed compared to the corresponding clinical plans.
Unscaled Scaled
2D-RGB 2D-dose 3D-dose 2D-RGB 2D-dose 3D-dose
Pass Fail Pass Fail Pass Fail Pass Fail Pass Fail Pass Fail
Clinical OARs Pass 63.4 3.2 64.9 1.7 65.8 0.8 60.5 6.1 63.1 3.5 63.4 3.2
Fail 6.2 27.2 7.9 25.5 7.9 25.5 4.7 28.7 5.9 27.5 4.6 28.8
Targets Pass 45.5 23.2 46.9 21.9 43.8 25.0 60.3 8.5 67.9 0.9 68.3 0.4
Fail 9.8 21.4 9.4 21.9 8.5 22.8 21.4 9.8 30.4 0.9 31.2 0.0
All ROIs Pass 58.5 8.7 60.0 7.2 59.7 7.5 60.5 6.7 64.4 2.8 64.7 2.4
Fail 7.2 25.6 8.3 24.5 8.1 24.7 9.3 23.5 12.6 20.2 11.9 20.9

Table LABEL:clinCrit summarizes the performance of the KBAP plans, focusing only on the criteria that the corresponding clinical plans also passed. Both 3D-dose and 3D-dose performed the best across all OAR and target criteria. In particular, 3D-dose achieved the highest passing rate for all target criteria. It also achieved the highest passing rate for all OAR criteria except the larynx and mandible. However, for some regions such as the brainstem, esophagus, and PTV63, multiple KBAP approaches yielded the top result.

Table 4: The percentage of criteria satisfied in each KBAP plan conditional on the clinical plan also achieving them. For each criteria, the highest value in each row is bolded.
Unscaled Scaled
2D-RGB 2D-dose 3D-dose 2D-RGB 2D-dose 3D-dose
OARs Brainstem 100.0 100.0 100.0 100.0 100.0 100.0
Spinal Cord 100.0 100.0 100.0 98.9 98.9 100.0
Right Parotid 94.1 88.2 94.1 94.1 88.2 94.1
Left Parotid 63.6 81.8 81.8 54.5 81.8 81.8
Larynx 91.8 89.8 98.8 87.8 87.8 91.8
Esophagus 100.0 100.0 100.0 100.0 100.0 100.0
Mandible 84.8 98.5 98.5 65.2 84.8 81.8
Targets PTV70 48.3 82.8 79.3 75.9 100.0 100.0
PTV63 100.0 100.0 94.0 100.0 100.0 100.0
PTV56 52.2 15.2 10.9 89.1 95.7 97.8

In Table LABEL:clinCritPlans, we recorded the proportion of KBAP plans that satisfied all of the same criteria as the corresponding clinical plans. Note that the subtle difference between these results and the results from Table LABEL:clinCrit is that here we only record a pass if the KBAP plan meets every single criteria that the clinical plan meets. Table LABEL:clinCrit, on the other hand, considers each criterion independently. We found that plans generated from the scaled predictions performed better than their unscaled counterparts in terms of satisfying all ROI criteria; we observed the biggest improvement between 3D-dose and 3D-dose (34.5 percentage points). Like all scaled plans, the improvement of 3D-dose, which satisfied the most target criteria (98.9%), was the result of much better target coverage at the expense of less OAR sparing compared to the 3D-dose plans, which satisfied the most OAR criteria (94.3%). Dose encoded as a grayscale image led to an increase in plan quality best shown by the difference of 27.6 percentage points separating 2D-RGB and 2D-dose. Finally, it was clear that the 3D GAN architecture resulted in large improvements in prediction quality as compared to the 2D approach. That is, 3D-dose satisfied criteria more frequently (by 2.3 percentage points) than 2D-dose. Overall, 3D-dose achieved the same criteria as the clinical plans in 78.2% of cases, which was more than any other approach.

Table 5: The percentage of plans in each KBAP population that satisfied all clinical criteria that were satisfied by the clinical plans. The highest percentage of satisfied criteria is bolded in each row.
Unscaled Scaled
2D-RGB 2D-dose 3D-dose 2D-RGB 2D-dose 3D-dose
OARs 80.5 88.5 94.3 62.1 78.2 79.3
Targets 51.7 52.9 47.1 78.2 97.7 98.9
All ROIs 42.5 47.1 43.7 48.3 75.9 78.2

KBP Prediction Quality

In Figure 3, we present box plots comparing the effect of each modification on prediction error, i.e., the average absolute error between the KBP predicted DVH and the KBAP plan DVH. Each subplot represents a pairwise comparison of the prediction error between KBAP pipelines with different GAN architectures for prediction. The larger the skew towards one side, the more accurate that GAN architecture is. By comparing 2D-RGB to 2D-dose, plotted in Figure 3(a), we found that dose encoded as a scalar value resulted in predictions that were closer (median error difference of 0.19 Gy) to deliverable plans as compared to predictions where dose was encoded as a RGB color map. We also found that the prediction errors were lower among scaled plans than unscaled plans (Figure 3(b)). That is, by comparing 2D-dose and 2D-dose, we found that half of all predictions generated via scaling were about 0.31 Gy more accurate than those generated without; the PTV56 was the only structure where the scaled KBAP plans were less accurate (average median error of 1.00 Gy). For both modifications, although the median differences were modest, the upper limit of the differences could be upwards of 2 Gy. Additionally, we found that the median error of 3D-dose was 0.07 Gy lower than 2D-dose (Figure 3(c)). Finally, in Figure 3(d), we evaluated the difference between our best set of KBAP plans, 3D-dose, and the plans generated by the previous state-of-the-art in GAN-based KBAP, 2D-RGB. The median error of 3D-dose was 0.63 Gy smaller than 2D-RGB, which suggests that the predictions made by 3D-dose were closer to the final dose distributions than those generated by 2D-RGB.

Figure 4 presents predicted dose distributions corresponding to the 2D-dose and a 3D-dose KBP methods. Note that the two models predicted distinct dose distributions that differed along the vertical (i.e., longitudinal) axis. In particular, the 3D GAN better learned the vertical relationship between adjacent axial slices and thus, generated smoother dose predictions across the longitudinal axis. In contrast, the 2D GAN predictions included more “streaky” and unrealistic discontinuities in the dose distribution, particularly around axial slices that were adjacent to the target boundaries. For example, the dose falloff was impossibly steep in the plans generated using 2D predictions (Figure 4(c)), while plans generated using 3D predictions had more realistic dose gradients (Figure 4(d)).

4 Discussion

Refer to caption
(a)  2D-RGB vs 2D-dose plans
Refer to caption
(b)  2D-dose vs 2D-dose plans
Refer to caption
(c)  2D-dose vs 3D-dose plans
Refer to caption
(d)  2D-RGB vs 3D-dose plans
Figure 3: The distribution of DVH errors between KBAP plans with respect to DVHs from their respective KBP predictions. The boxes indicate median and interquartile range (IQR). Whiskers extend to the minimum of 1.5 times the IQR and the most extreme outlier.
Refer to caption
(a)  CT image
Refer to caption
(b)  Clinical plan
Refer to caption
(c)  2D-dose prediction
Refer to caption
(d)  3D-dose prediction
Refer to caption
Figure 4: The dose distributions for a sample patient over a single CT image (a) of their clinical plan (b), 2D-dose prediction (c), and 3D-dose prediction (d).

Although historically, KBP methods have predicted DVHs using hand-tailored features, there is widespread interest in automatically predicting dose distributions [McIntosh and Purdie, 2016, Nguyen et al., 2017, Mahmood et al., 2018, Fan et al., 2018]. In this paper, we extend the literature by building a KBAP pipeline that automatically generates treatment plans from CT images. The pipeline consists of two major components: prediction and optimization. The prediction stage uses a generative adversarial network to predict dose distributions from a CT image. The optimization stage consists of two optimization models, a parameter estimation model that learns objective function weights from the predicted dose distribution and an inverse planning model that produces the final fluence-based treatment plans. We demonstrate that our approach produces treatment plans that are better than the previous state-of-the-art in KBAP and generates full dose distributions rather than summary statistics [McIntosh and Purdie, 2017, Fan et al., 2018].

Our framework includes three enhancements that improved performance: 1) representing dose to each voxel as a scalar value, 2) scaling the KBP prediction from the GAN prior to optimization, and 3) predicting the full 3D dose distribution from 3D CT images.

Representing dose to each voxel as a scalar. We hypothesized that predicting dose encoded in single value (i.e., grayscale) would be easier than predicting dose encoded as a 3-channel RGB value. Our experiments confirmed this result. In particular, 2D-dose plans satisfied the same criteria as the clinical plans 74% more often across all ROIs as compared to 2D-RGB plans. The improvement due to this modification is likely due to the fact that it is much easier to predict a single value rather than a triplet.

Scaling the predictions before optimization. Scaling significantly enhanced the final KBAP plan quality. Scaled KBAP plans satisfied the same criteria as clinical plans 52% more often than unscaled plans; scaled plans also satisfied 10% more criteria than the unscaled plans overall. The idea of scaling predictions prior to optimization is novel in the KBAP literature and is a general tool that can be applied to any other KBP method. There are two points worth highlighting. First, scaling is done automatically, just like how our approach automates high-dimensional feature selection. Thus, our KBAP pipeline remains automated. Second, we believe that scaling works because it corrects small inaccuracies that may arise when learning the absolute dose delivered. That is, the GAN seems to be more effective at learning how dose varies among different tissues rather than learning the exact dose that should be delivered to a tissue (otherwise, scaling would not make any difference).

Predicting the full 3D dose distribution from 3D CT images. Whereas previous work predicted dose to each voxel or axial slice independently and then stitched the predictions together to form the full 3D dose distribution [Shiraishi and Moore, 2016, Nguyen et al., 2017, Mahmood et al., 2018, Fan et al., 2018], our 3D GAN was designed to take an entire contoured 3D CT image as input and generate the corresponding 3D dose distribution as output. Because of this enhancement, the 3D GANs better learned the vertical relationship between adjacent axial slices. Indeed, we observed that the 3D GAN predicted more realistic dose distributions than the 2D GAN (e.g., Figure 4) with smoother dose predictions across the longitudinal axis.

In our experiments, we used clinical criteria as the primary performance measure to evaluate the plan quality of several GAN architectures. Since it is generally impossible to develop plans that simultaneously achieve all clinical criteria–in our dataset of 217 clinically delivered plans, only 68.4% of criteria were achieved–our primary goal was to achieve as many criteria as possible. Secondarily, we were interested in generating plans that met the same criteria that the original clinical plans achieved; presumably, these represent the criteria that clinicians originally believed to be the most important. We believe that an automated planning method that produces dose distributions that satisfy the same criteria as treatment plans that have already been delivered is more likely to be clinically implemented.

There are several reasons why we believe GANs are a good choice for KBP. First, they have a history of performing well in applications that involve medical images; specifically in the detection of brain lesions [Alex et al., 2017] and image augmentation for liver lesion classification [Frid-Adar et al., 2018]. Second, we found that all of our GAN models performed well inside a KBAP pipeline without significant parameter tuning and architecture modification, both of which are essential and potentially time consuming steps in conventional GAN implementations. We attribute this success to the application; the prediction of dose distributions is akin to the prediction of relatively smooth and uniform images with the same orientation. Third, in the KBAP pipeline, the GAN produces images that are used as input for an optimization model to obtain treatment plans via a traditional inverse planning procedure. Thus, the GAN learns a simpler style mapping as compared to conventional applications, and the optimization phase acts as a safety net to correct potential errors. Finally, it is interesting to note that the method to train a GAN conceptually mimics the iterative process between a treatment planner and an oncologist. The generator behaves as a treatment planner by proposing deliverable dose distributions while the discriminator behaves as an oncologist by determining whether the proposed dose distribution is suitable.

While other pipelines that predict dose directly have used voxel-based dose mimicking to construct the final plans [McIntosh et al., 2017], we chose to do inverse planning using DVH-based objectives following prediction because it is in line with current clinical practice and is the most common approach used in the academic KBAP literature [Wu et al., 2017, Babier et al., 2018a, Fan et al., 2018]. We also emphasize that our method does not use hand-tailored feature engineering (e.g., features derived from overlap-volume histograms). Thus, as compared to existing KBAP methods, we expect our pipeline to be easier to implement in practice and can result in more predictable results if custom treatment plans are desired. For example, institutions with specific clinical guidelines can train a GAN on images they deem indicative of an ideal treatment plan. In addition, in the future it may be possible that several medical centers combine data to form a large training set, which should further improve performance.

A limitation of our approach is that the prediction and optimization steps are separate stages in the overall pipeline. In theory, an integrated model that does both prediction and treatment plan optimization simultaneously should produce even better results. A second limitation is that our approach requires a clean, well-structured and high-quality dataset, where all images need to have a consistent size (in terms of number of pixels), coloring convention, and orientation. Finally, as with any neural network-based approach, GAN predictions suffer from a lack of interpretability. It is not straightforward to understand why the GAN makes certain predictions, effectively rendering it a black box. Consequently, a treatment planner may have more difficulty using this approach to understand when and why prediction errors occur.

5 Conclusion

We developed the first knowledge-based automated planning framework using a 3D generative adversarial network for prediction. Our results based on 217 oropharyngeal cancer treatment plans demonstrated superior performance in satisfying clinical criteria and generated more realistic predictions compared to the previous state-of-the-art. Our three unique contributions to the KBAP architecture of one-channel prediction, post-prediction scaling, and direct prediction of the full 3D dose allowed us to generate high-quality treatments without manual intervention.

6 Acknowledgments

This research was supported in part by the Natural Sciences and Engineering Research Council of Canada.

References

  • Alex et al. [2017] V. Alex, M. S. KP, S. S. Chennamsetty, and G. Krishnamurthi. Generative adversarial networks for brain lesion detection. In Medical Imaging 2017: Image Processing, volume 10133, page 101330G. International Society for Optics and Photonics, 2017.
  • Appenzoller et al. [2012] L. M. Appenzoller, J. M. Michalski, W. L. Thorstad, S. Mutic, and K. L. Moore. Predicting dose-volume histograms for organs-at-risk in IMRT planning. Med Phys, 39(12):7446–7461, 2012.
  • Atun et al. [2015] R. Atun, D. A. Jaffray, M. B. Barton, F. Bray, M. Baumann, B. Vikram, T. P. Hanna, F. M. Knaul, Y. Lievens, T. Y. M. Lui, M. Milosevic, B. O’Sullivan, D. L. Rodin, E. Rosenblatt, J. Van Dyk, M. L. Yap, E. Zubizarreta, and M. Gospodarowicz. Expanding global access to radiotherapy. Lancet Oncol, 16(10):1153–86, Sep 2015.
  • Babier et al. [2018a] A. Babier, J. J. Boutilier, A. L. McNiven, and T. C. Y. Chan. Knowledge-based automated planning for oropharyngeal cancer. Med Phys, 45(7):2875–2883, Jul 2018a.
  • Babier et al. [2018b] A. Babier, J. J. Boutilier, M. B. Sharpe, A. L. McNiven, and T. C. Y. Chan. Inverse optimization of objective function weights for treatment planning using clinical dose-volume histograms. Phys Med Biol, 63(10):105004, May 2018b.
  • Craft et al. [2007] D. Craft, P. Suss, and T. Bortfeld. The tradeoff between treatment plan quality and required number of monitor units in intensity-modulated radiotherapy. Int J Radiat Oncol Biol Phys, 67(5):1596–605, 2007.
  • Das et al. [2009] I. J. Das, V. Moskvin, and P. A. Johnstone. Analysis of treatment planning time among systems and planners for intensity-modulated radiation therapy. J Am Coll Radiol, 6(7):514–7, Jul 2009.
  • Deasy et al. [2003] J. O. Deasy, A. I. Blanco, and V. H. Clark. CERR: a computational environment for radiotherapy research. Med Phys, 30(5):979–85, 2003.
  • Fan et al. [2018] J. Fan, J. Wang, Z. Chen, C. Hu, Z. Zhang, and W. Hu. Automatic treatment planning based on three-dimensional dose distribution predicted from deep learning technique. Med Phys, Nov 2018. doi: 10.1002/mp.13271.
  • Frid-Adar et al. [2018] M. Frid-Adar, I. Diamant, E. Klang, M. Amitai, J. Goldberger, and H. Greenspan. Gan-based synthetic medical image augmentation for increased cnn performance in liver lesion classification. arXiv:1803.01229, 2018.
  • Goodfellow et al. [2014] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio. Generative adversarial nets. Advances in neural information processing systems, pages 2672–2680, 2014.
  • Hermoza and Sipiran [2017] R. Hermoza and I. Sipiran. 3D reconstruction of incomplete archaeological objects using a generative adversary network. arXiv:1711.06363, 2017.
  • Isola et al. [2017] P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros. Image-to-image translation with conditional adversarial networks. arXiv:1611.07004, 2017.
  • Kingma and Ba [2014] D. P. Kingma and J. Ba. Adam: A method for stochastic optimization. arXiv:1412.6980, 2014.
  • Mahmood et al. [2018] R. Mahmood, A. Babier, A. Mcniven, A. Diamant, and T.C.Y. Chan. Automated treatment planning in radiation therapy using generative adversarial networks. Proceedings of Machine Learning Research, 85:1–15, 2018.
  • McIntosh and Purdie [2016] C. McIntosh and T. G. Purdie. Contextual atlas regression forests: Multiple-atlas-based automated dose prediction in radiation therapy. IEEE Trans Med Imaging, 35(4):1000–12, Apr 2016.
  • McIntosh and Purdie [2017] C. McIntosh and T. G. Purdie. Voxel-based dose prediction with multi-patient atlas selection for automated radiotherapy treatment planning. Phys Med Biol, 62(2):415–431, Jan 2017.
  • McIntosh et al. [2017] C. McIntosh, M. Welch, A. McNiven, D. A. Jaffray, and T. G. Purdie. Fully automated treatment planning for head and neck radiotherapy using a voxel-based dose prediction and dose mimicking method. Phys Med Biol, 62(15):5926–5944, 2017.
  • Nguyen et al. [2017] D. Nguyen, T. Long, X. Jia, W. Lu, X. Gu, Z. Iqbal, and S. Jiang. Dose prediction with u-net: A feasibility study for predicting dose distributions from contours using deep learning on prostate imrt patients. arXiv:1709.09233, 2017.
  • Purdie et al. [2014] T. G. Purdie, R. E. Dinniwell, A. Fyles, and M. B. Sharpe. Automation and intensity modulated radiation therapy for individualized high-quality tangent breast treatment plans. Int J Radiat Oncol Biol Phys, 90(3):688–95, 2014.
  • Shiraishi and Moore [2016] S. Shiraishi and K. L. Moore. Knowledge-based prediction of three-dimensional dose distributions for external beam radiotherapy. Med Phys, 43(1):378, Jan 2016. doi: 10.1118/1.4938583.
  • Shiraishi et al. [2015] S. Shiraishi, J. Tan, L. A. Olsen, and K. L. Moore. Knowledge-based prediction of plan quality metrics in intracranial stereotactic radiosurgery. Med Phys, 42(2):908, 2015.
  • Wu et al. [2017] B. Wu, M. Kusters, M. Kunze-Busch, T. Dijkema, T. McNutt, G. Sanguineti, K. Bzdusek, A. Dritschilo, and D. Pang. Cross-institutional knowledge-based planning (KBP) implementation and its performance comparison to auto-planning engine (APE). Radiother Oncol, 123(1):57–62, 2017.
  • Yang et al. [2013] T. Yang, E. C. Ford, B. Wu, M. Pinkawa, B. van Triest, P. Campbell, D. Y. Song, and T. R. McNutt. An overlap-volume-histogram based method for rectal dose prediction and automated treatment planning in the external beam prostate radiotherapy following hydrogel injection. Med Phys, 40(1):011709, 2013.
  • Younge et al. [2018] K. C. Younge, R. B. Marsh, D. Owen, H. Geng, Y. Xiao, D. E. Spratt, J. Foy, K. Suresh, Q. J. Wu, F. Yin, S. Ryu, and M. M. Matuszak. Improving quality and consistency in NRG oncology radiation therapy oncology group 0631 for spine radiosurgery via knowledge-based planning. Int J Radiat Oncol Biol Phys, 100(4):1067–1074, Mar 2018.
  • Yuan et al. [2012] L. Yuan, Y. Ge, W. R. Lee, F. F. Yin, J. P. Kirkpatrick, and Q. J. Wu. Quantitative analysis of the factors which affect the interpatient organ-at-risk dose sparing variation in IMRT plans. Med Phys, 39(11):6868–78, 2012.
  • Zhu et al. [2017] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros. Unpaired image-to-image translation using cycle-consistent adversarial networks. arXiv:1703.10593, 2017.
  • Zhu et al. [2011] X. Zhu, Y. Ge, T. Li, D. Thongphiew, F. Yin, and Q. J. Wu. A planning quality evaluation tool for prostate adaptive IMRT based on machine learning. Med Phys, 38(2):719–26, 2011.
  • Ziemer et al. [2017] B. P. Ziemer, S. Shiraishi, J. A. Hattangadi-Gluth, P. Sanghvi, and K. L. Moore. Fully automated, comprehensive knowledge-based planning for stereotactic radiosurgery: Preclinical validation through blinded physician review. Pract Radiat Oncol, 7(6):e569–e578, 2017. doi: 10.1016/j.prro.2017.04.011.