Deep segmentation networks predict survival of non-small cell lung cancer

Non-small-cell lung cancer (NSCLC) represents approximately 80–85% of lung cancer diagnoses and is the leading cause of cancer-related death worldwide. Recent studies indicate that image-based radiomics features from positron emission tomography/computed tomography (PET/CT) images have predictive power for NSCLC outcomes. To this end, easily calculated functional features such as the maximum and the mean of standard uptake value (SUV) and total lesion glycolysis (TLG) are most commonly used for NSCLC prognostication, but their prognostic value remains controversial. Meanwhile, convolutional neural networks (CNN) are rapidly emerging as a new method for cancer image analysis, with significantly enhanced predictive power compared to hand-crafted radiomics features. Here we show that CNNs trained to perform the tumor segmentation task, with no other information than physician contours, identify a rich set of survival-related image features with remarkable prognostic value. In a retrospective study on pre-treatment PET-CT images of 96 NSCLC patients before stereotactic-body radiotherapy (SBRT), we found that the CNN segmentation algorithm (U-Net) trained for tumor segmentation in PET and CT images, contained features having strong correlation with 2- and 5-year overall and disease-specific survivals. The U-Net algorithm has not seen any other clinical information (e.g. survival, age, smoking history, etc.) than the images and the corresponding tumor contours provided by physicians. In addition, we observed the same trend by validating the U-Net features against an extramural data set provided by Stanford Cancer Institute. Furthermore, through visualization of the U-Net, we also found convincing evidence that the regions of metastasis and recurrence appear to match with the regions where the U-Net features identified patterns that predicted higher likelihoods of death. We anticipate our findings will be a starting point for more sophisticated non-intrusive patient specific cancer prognosis determination. For example, the deep learned PET/CT features can not only predict survival but also visualize high-risk regions within or adjacent to the primary tumor and hence potentially impact therapeutic outcomes by optimal selection of therapeutic strategy or first-line therapy adjustment.


Introduction
According to the World Health Organization (WHO), lung cancer remains the leading cause of cancer-related deaths worldwide, with 2.1 million new cases diagnosed and 1.8 million deaths in 2018 1 .NSCLC accounts for 80-85% of lung cancer diagnoses 2 and the five-year survival rate of NSCLC remains low (23%) compared to other leading cancer sites such as colorectal (64.5%), breast (89.6%), and prostate (98.2%) 3 .Historically, the tumor, nodes, and metastases (TNM) staging system has served as the major prognostic factor in predicting therapeutic outcomes, but it does not differentiate responders and non-responders within the same stage 4 .The maximum and the mean of standard uptake values (SUV MAX and SUV MEAN ) have been reported for their correlation with survival [5][6][7] , but are of limited clinical value due to their unsatisfactory predictive power and lack of robustness 8,9 .Other prognostic markers have also been studied, including TLG, which incorporates metabolic tumor volume (MTV) and metabolic activity (TLG = MTV × SUV MEAN ).Reports [10][11][12] suggest that TLG may have better prognostic power than SUV MAX or SUV MEAN .These metrics, however, are not optimal and do not provide a comprehensive image-based analysis of tumors 13 .More recently, radiomics approaches, which employ semi-automated analysis based on a few hand-crafted imaging features describing intratumoral heterogeneity, demonstrated higher prognostic power 14,15 .However, these features still have limited predictive power ranging between 0.5 and 0.79 in terms of the area under the ROC (AUC) [15][16][17] .Recent literature on Figure 1.Schematic diagram of the survival prediction framework.The proposed framework consists of two major components: the U-Net segmentation network and the survival prediction model.The U-Net is trained with PET/CT images and corresponding physician contours but without survival-related information.The "dimensional bottleneck" at the middle of the U-Net produces latent variables summarizing image features (55,296 features from CT + 55,296 features from PET), which we hypothesize to be potentially relevant to cancer survival.These features are then clustered by k-medoids clustering approach in an unsupervised manner.Next, the LASSO method is used to select medoid features from the clusters based on their associations with survival.Last, a logistic regression model is trained for survival prediction, and survival prediction is performed when a new patient arrives with features extracted from the same U-Net.
CNNs demonstrates their strong potential for cancer prognostication 18,19 , however the clinical implications of deep learning remain questioned due to the limited interpretability of CNNs.
Here, we propose an interpretable and highly accurate framework to solve this problem by capitalizing on the unprecedented success of deep convolutional neural networks (CNN).More specifically, we investigate U-Net 20 , a convolutional encoderdecoder network that has demonstrated exceptional performance in tumor detection and segmentation tasks.Illustrated in Fig, 1a, U-Net takes a three-dimensional (3D) volume image as an input, processes it through a "bottleneck layer" where the image features are compressed, and reconstructs the image into a binary segmentation map indicating a pixel-wise tumor classification result.Here, we focused on the information encoded at the bottleneck layer, which contains rich visual characteristics of the tumor and hypothesized that the encoded information at this layer might be relevant to the tumor malignancy, and thus cancer survival, which is the central hypothesis of this paper.

Results
In prior studies 21,22 , we analyzed PET/CT images of 96 NSCLC patients that were obtained within 3 months prior to SBRT, whose summary statistics are illustrated in Extended Data Fig. 1 and Extended Data Table 1.For each volume image, the region of interest (ROI) with a dimension of 96 mm × 96 mm × 48 mm was set around each SBRT treated tumor location and the image was cropped to the ROI volume.Two separate U-Net models were trained to perform tumor segmentation in PET and CT images, respectively.Each of the models was trained with 38 ROI images and the corresponding physician contours, but no other information such as survival time was provided.Segmentation performance was tested on 22 independent ROI images that were not included in training, and the average Sørensen-Dice similarity coefficients (DSC) were 0.861 ± 0.037 (mean ± std) and 0.828 ± 0.087 for CT and PET, respectively.After training, each U-Net model learned to encode 55,296 features at the bottleneck layer for each patient, resulting in a total of 110,592 features per patient.
These features are an intermediate throughput of U-Net, and are then decoded to generate an automated segmentation in the network.It is likely that these features summarize some rich structural and functional geometry of the intratumoral and peritumoral area, some of which might be relevant to cancer survival.We test this proposition by conducting validation studies and examining the statistical prognostic power of those features.One challenge here is that the number of features (d = 110, 592) is substantially larger than the number of cases, making a statistical analysis prone to overfitting.To this end, we first reduce the number of features via an unsupervised feature selection process in which the survival information is hidden.The k-medoids clustering method 23 is employed to serve this purpose, as the method is known to be able to select representative features from a large pool of inter-correlated features in similar literature 24 .For the feature similarity measure in k-medoids clustering, we use the Pearson correlation distance defined as 1 − R, where R is the Pearson correlation coefficient.The optimal number of clusters is determined by the Silhouette method 25 .Finally, given the optimal k clusters, the medoids of the clusters are selected to be the representative features of the clusters and form an unsupervised feature pool.We choose the final set of features via the least absolute shrinkage and selection operator (LASSO) 26 on the preliminary feature pool obtained from the k-medoids method.LASSO attempts to select a few features among the cluster centers that have a strong relationship with the survival outcomes using the fixed regularization parameter.The selected features then serve as regressors in the logistic regression model to predict the cancer survival outcome.Since the performance of k-medoids clustering is also dependent on a random initialization of the cluster centers, we tested 10 times for each k and computed the mean of the summation of the inner cluster distances to ensure consistency of the clustering outcome.We then computed the curve of the Silhouette value with respect to the number of clusters in total to determine the optimal number of clusters.The medoids of the feature clusters are selected to form a reduced feature set.
We then examined the selected features and their prognostic power via cross-validation on the NSCLC data set collected at the University of Iowa Hospitals and Clinics (UIHC).As summarized in Extended Data Fig 1 and Extended Data Table 1, the UIHC data set is comprised of primary and follow-up PET/CT images of total 96 NSCLC cases with their survival status and other clinical meta data.Information on the cause of death is also available so that the deceased cases can be further broken down to overall and disease-specific deaths.On this data set, we aim to predict four survival categories, namely, 2-year overall survival (2OS), 5-year overall survival (5OS), 2-year disease-specific survival (2DS), and 5-year disease-specific survival (5DS).Total N = 96 cases qualify for 2OS category, 74 for 5OS, 92 for 2DS, and 45 for 5DS, depending on the survival status and the cause of death.Other clinical meta data, such as sex, age, smoking history, and tumor, lymph nodes, and metastasis (TNM) staging, exist in the data set and may provide an improved prognostic power when added as regressors.However, we exclude all other parameters but U-Net encoded image features, in order to focus the analysis on the image features only.Only primary images are encoded via U-Net and used for prediction.Follow-up images are reserved for comparison and discussion later in this paper.Each cross validation experiment is comprised of unsupervised feature selection using k-medoids clustering, LASSO-based feature selection, and training of the logistic regression model.These tasks are performed independently from the other cross-validation experiments.In each experiment, we measure the accuracy, sensitivity, specificity, and the area under the receiver operating characteristic curve (AUC) per each survival category.We compute the average and the standard deviation of these performance metrics across the cross-validation experiments to derive the estimated performance metrics and their 95% confidence intervals (95% CI).
In such cross-validation experiments, estimated AUC of the proposed prediction model is 0.88 (95% CI: 0.80-0.96)for the prediction of 2OS.For other survival criteria, the estimated AUCs are similar, namely 0.89 (95% CI: 0.85-0.93)for 5OS, 0.86 (95% CI: 0.81-0.91)for 2DS, and 0.88 (95% CI: 0.81-0.95)for 5DS.Note that the estimated AUC values of the conventional TLG and other radiomics markers 27 range between 0.60 and 0.83 on the same data set.A more recent deep learning method reported in Hosny et al. 28 produces AUCs between 0.70∼0.73 on the same data set.A graphical illustration of the result as well as the full set of performance metrics are available in Fig. 2 and Extended Data Table 2, respectively.
Moreover, we further validated the result on an extramural dataset provided by the Stanford Cancer Institute.The Stanford In this setting, the estimated AUC is 0.87 (95% CI: 0.80-0.94)for 2 year OS, and 0.90 (95% CI: 0.82-0.98)for 5 year OS.More detailed results are illustrated in Fig. 3 and Extended Data Table 3.Meanwhile, we visualized the features learned by U-Net to develop an intuitive understanding of what those prognostic markers represent.The image features learned by U-Net are essentially artificial neurons in deep neural networks.In principle, we can visualize a neuron by showing multitudes of image patterns and observing which image pattern activates the neuron the most.Practically, we employ an optimization-based approach 29 where the objective is to maximize an individual neuron's activation value by manipulating the input image pattern: where q(•|W, b) is the U-Net encoder with the trained model parameters W and b, and X is the input image pattern.Displayed in Fig. 4 and Extended Data Fig. 2 are visualizations of features captured by artificial neurons.We also visualized which regions in the patient images predicted low survival probability.We employed a guided gradient backpropagation approach 30 .The main idea of the guided backpropagation algorithm is to compute where P is the probability of death and x i, j,k is a voxel value at the position (i, j, k) in the patient image.The gradient ∂ P ∂ x i, j,k can be interpreted as the change of the death probability when the voxel x i, j,k changes to a different value.If the voxel was not so significant in predicting death, the gradient value would be small, where as if the voxel played an important role in predicting high probability of death, the gradient value would be greater.Displayed in Fig. 5 and Extended Data Figs.3-5 are heat maps representing the gradient.Heated regions (red) are the areas that lowered the probability of survival whereas the other areas (blue) are the ones that had negligible effect on the survival.

Discussion
As illustrated Fig. 2 and Extended Data Table 2, the contrast in the predictive performance between the U-Net features and the conventional imaging features (SUV MEAN , SUV MAX and TLG) was evident, quantitatively proving the strong prognostic power of the U-Net features.There also was a noticeable enhancement of prognostic performance when the U-Net features were compared with a recent deep learning approach as in Hosny et al. 28 The same trend could be observed in Fig. 3 and Extended Data Table 3 where we validated the U-Net features against an extramural data set, confirming the enhanced prognostic performance of the U-Net features.Here, it is worth reemphasizing that the U-Net was trained without any survival-related information, and, thus it is highly unlikely that the U-Net-learned features were overfitted to the survival data or biased towards Note these templates are learned in an unsupervised manner, without any survival-related information provided, despite which these were later discovered to be survival-related.Note that the templates captured by U-Net are characterized by their sensical and interpretable geometric structures .For example, C37399 appears to be a template looking for a tumor-like shape at the top-right corner and a tube-like structure at the bottom-left.In addition, C08680 appears to look for a textural feature of the tumor.(c) In contrast, image templates learned by direct fitting of a CNN model to the survival data 28 .Note the features in (c) are less interpretable compared to the U-Net features.them.Nonetheless, while these U-Net features were identified independently from the survival data, the U-Net features demonstrated strong evidence of correlation and thus prognostic power for NSCLC survival as discussed above.
Here, a justification should be necessary on the rationale behind taking a "detour" by training a segmentation network first, then extracting prognostic image features, and training a survival model, as opposed to directly training a CNN model to the survival data as in other literature. 28,31,32 T this end, we make the following arguments.First, a CNN model trained on a segmentation task is more robust to overfitting and is more generalizable, as the image features themselves are developed from an unsupervised training.A direct prediction of survival from an image tends to be less robust and less generalizable, as the millions of parameters in a CNN can easily misconstrue the trend, and this process is difficult to control.Secondly, segmentation-trained features contain more structural, geometric features that are human-interpretable, whereas direct-trained features tend to be biased towards less-intuitive and fuzzy texture patterns.It is, in fact, well-understood now that CNNs tend to be biased towards texture rather than shape, while the human cognitive process works in the opposite. 33For example, Geirhos et al. 34 conducted an experiment where the texture of a cat in an image was replaced by the texture of an elephant-CNNs started to discern the image as an elephant while humans still categorized the image as a cat.From this, we can infer that the direct training of a CNN for survival prediction would result in a development of less sensical features lacking geometric information.In contrast, segmentation-trained features are incentivized to focus on geometric structures and shapes rather than textures, and thus are more sensical.
The above arguments are indeed supported by the visualization results in Fig. 4 and Extended Data Fig. 2, where the benchmark CNN method appears to be capturing some noisy texture patterns.In contrast, many of the U-Net features appear to be capturing more sensical and interpretable geometric shapes from the image, such as tumor-like blobs (e.g.C00048, C25988, P39051, P47258) or heterogeneity of the tumor (e.g.C08680).Interestingly, some of the U-Net features, for example C01777 and C37399, were looking for tube-like structures nearby the tumor-like blobs, which might be capturing blood vessels and lymphatics in the peritumoral area.This is consistent with the widely accepted clinical knowledge that tumors can show enhanced growth in the presence of nearby vessels and lymphatics as they carry nutrition to supply the tumoral growth.
Meanwhile, we further visualized the proposed prediction model by creating a heat map highlighting regions that predicted low survival probability.The map generated by the guided back-propagation algorithm confirmed that the prediction model was looking at clinically sound regions for making predictions by responding more to tumor and surrounding tumor regions than other regions.In addition to the results presented above, this is yet more evident that the prediction model has produced generalizable features and rules for making prognostic prediction, and is thus not overfitted.Moreover, through comparison with post-SBRT CT images and clinical records of the patients, we observe that the heat map has the potential to identify regions of progressions or recurrence.For the case illustrated in Fig. 6 as an example, a small nodule had been found initially at the progression region, but was not treated by SBRT as the region was not originally identified as tumor by PET.However, this region was identified as a high risk region on heat map by the U-Net as well as the primary tumor region of SBRT.Tumor recurrence was found on post-SBRT CT images and matched the highlighted risk region on pre-SBRT U-Net generated heat map.From such observations, the heat map visualization has the potential to identify regions at high risk for tumor progression or recurrence that could be utilized for the purpose of assisting patient-tailored treatment planning in the future.For this reason, we believe this indicates more rigorous risk map developments and requires quantitative follow-up validations, which will be our subsequent project.
In summary, we discovered that the U-Net segmentation algorithm trained for automated tumor segmentation on PET/CT, codifies rich structural and functional geometry at the bottleneck layer.These codified features, in turn, could be used for survival prediction in cancer patients even though the U-Net was trained without any survival-related information.The survival model based on such U-Net features demonstrated significantly higher predictive power compared to conventional PET-based, metabolic burden metrics such as TLG or relatively recent hand-crafted radiomics approaches.The validity of this discovery was further confirmed by the validation on an extramural data set provided by the Stanford Cancer Institute.Furthermore, we visualized the survival-related U-Net features and observed that they were indeed depicting intratumoral and/or peritumoral structures that had been previously acknowledged as potentially relevant to cancer survival.Our approach awaits a further validation against a larger number of observations and in a larger variety of cancer types.However, there was not enough clinical evidence to conclude that the visualization of the U-Net features may identify potential regions of recurrence and metastasis and, thus, a follow-up study is suggested.Our findings may suggest a new starting point for quantitative image-based cancer prognosis with a great deal of important new knowledge to be discovered.

Methods
Subjects PET-CT images of NSCLC patients who received SBRT at the University of Iowa Hospitals and Clinics were investigated in this study.The images were obtained using a dual PET-CT scanner (Biograph 40 PET/CT, Siemens Medical Solutions USA, Inc., Hoffman Estates, IL).CT image used for SBRT planning were co-registered with CT images of PET-CT datasets using a deformable image registration (DIR).Using the deformation map (vectors), PET images were resampled with the primary CT images of SBRT plan.Post-SBRT follow-up images were used only for qualitative validation of the visualization results.The gross tumor volume (GTV) for each of the images was delineated by radiation oncologists on both

7/19
CT and PET images with the guidance of the corresponding images in the other modality.All tumor contouring was completed using VelocityAI (Varian Medical System, Inc., Palo Alto, CA).
A total of 96 cases (Male=44; Feamle=52) were investigated in this study.Patients' group stages vary in sub-categories including 41 in stage I, 10 in stage II, 16 in stage III, 29 in stage IV.Meanwhile, the histology was confirmed by a thoracic pathologist based off visual interrogation of the biopsied or surgically resected specimen.Histologies among the 96 patients include 48 adenocarcinoma, 41 squamous cell, 1 adenosquamous, 3 metastases from previous NSCLC, and 1 without biopsy.For overall survival from SBRT, the 2-year survival rate is 54% and the 5-year survival rate is 6%.For overall survival from diagnosis, the 2-year survival rate is 66% and the 5-year survival rate is 28%.For disease-specific survival from SBRT, the 2-year survival rate is 61% and the 5-year survival rate is 23%.For disease-specific survival from diagnosis, the 2-year disease specific survival rate is 73% and the 5-year disease specific survival rate is 51%.Among a total 96 patients, the qualified portion utilized in this research in each survival sub-category has been visualized and discussed in Extended Data Fig. 1 respectively.
In this study, the total of 96 utilized NSCLC patients were retrospectively analyzed after approval from the University of Iowa Institutional Review Board (IRB: 200503706; Name: The utility of imaging in cancer: staging, assessment of treatment response, and surveillance).All data collection, experimental procedures, and methods applied are in accordance with the relevant guidelines and regulations.All patients consented for the use of their clinical information and medical images.All participants enrolled in this study signed an informed consent developed and approved by the University of Iowa Institutional Review Board.All scans were in digital imaging and communications in medicine (DICOM) format and de-codified.

Data Processing
Each pair of the co-registered PET-CT images were resampled with an isotropic spacing in all three dimensions.Each image was then cropped into a fixed size of 96 × 96 × 48 voxels where each voxel represents 1 cubic millimeter (mm 3 ) after resampling.Within the resampled voxels, intensity values were clipped to the range of [-500, 200] for all CTs and [0.01, 20] for all PETs, to remove outliers.More details are described in our previous work 21,22 .U-Net Features In our previous work 22 , two independent 3D U-Net networks were constructed and trained for automated tumor segmentation in PET and CT images, respectively.The U-Net networks comprise two major components: the encoder network and the decoder network (Fig. 1).The encoder network takes a 96 × 96 × 48 volume image as an input.The first convolution layer produces 32 features attached to each voxel, representing low-level visual cues such as lines, edges, and blobs.These features are then down-sampled by half in all three dimensions and the down-sampled volume is fed into the second convolution layer, which then produces 64 features per each voxel.This is repeated three more times increasing the number of features to 128, 256, and finally 512, while the volume size is reduced by half in all three dimensions each time.The final 6 × 6 × 3 × 512(= 55, 296) features that the encoder network produces are an abstract, high-level summary of the input image, which is then decoded by the symmetric decoder to produce the binary segmentation map (1: tumor, 0: none).The convolutional kernels are of a size 3 × 3 × 3 across all layers and the max-pooling layers of a 2 × 2 × 2 window size with a stride of 2 were used for down-sampling.
To train the PET-CT segmentation U-Net networks, we utilized co-registered PET-CT scan pairs from 60 patients with primary NSCLC.For each PET-CT image, the slice image size is 512 × 512 and the number of slices varies from 112 to 293.The tumor contour on each of the PET and CT scans were labeled by physicians as groundtruth.In data preprocessing, all PET-CT images are resampled with an isotropic spacing of 1 × 1 × 1 in voxels and then cropped at a fixed size of 3D volumes (96 × 96 × 48) centered on the mass gravity of each tumor.
All 60 PET-CT scan pairs were split into a training data set with 38 patients and testing data set with 22 patients.Data augmentation was performed using simple translation, rotation and flip operations and the augmented training set has over 3000 3D PET-CT scan pairs respectively.
The 3D-UNet was trained using open source TensorFlow package and ran on NVIDIA GeForce GTX 1080 Ti GPU with 11GB of memory.The Adam optimization method was utilized with a mini-batch size of 1 and for 20 epochs.To prevent overfitting, the weight decay and early-stop techniques were adopted to obtain the best performance on the test set where the DSC value was computed.
After the U-Nets had been trained, PET and CT images of each patient are fed into the U-Nets to produce 55,296 features per imaging modality.These 55,296 features extracted from the U-Net encoder were used for the analyses throughout the paper.The schematic diagram with respect to the methodologies applied in this research for the feature analysis has been illustrated in where X and Y denote the two distinct features; Cov() denotes the covariance of the two features and Var()

Figure 2 .
Figure 2. Prognostic performance of the U-Net features.There are four survival categories being tested: 2-year overall survival (2-yr.OS), 5-year overall survival (5-yr.OS), 2-year disease-specific survival (2-yr.DS), and 5-year disease-specific survival (5-yr.DS).The U-Net features are compared against the conventional TLG metric, the 17 radiomics features defined in Oikonomou et al. 27 and the benchmark CNN prediction model in Hosny et al. 28 .The box plots represent the average performance scores as indicted by the central mark and 25th and 75th percentiles across 6-fold cross validation experiments.(a) Overall prediction accuracy (proportion of the correct prediction over the entire data set).(b) Sensitivity (correct prediction of death over all death cases).(c) Specificity (correct prediction of survival across all survival cases).(d) AUC of the receiver operating characteristics (ROC) curve.

Figure 3 .
Figure 3. Prognostic performance on an extramural data set.The extramural data set provided by Stanford only includes CT images and not PET images.Additionally, the disease-specific survival information is not provided.Therefore, in this experiment, two survival categories are being tested: 2-year overall survival (2-yr.OS) and 5-year overall survival (5-yr.OS).The U-Net features are compared against the 6 CT-based radiomics features (Radiomic) defined in Oikonomou et al. 27 and the benchmark CNN prediction model (Benchmark) in Hosny et al. 28 .The box plots represent the average performance scores as indicted by the central mark and 25th and 75th percentiles across experiments tested on extramural Stanford data set.(a) Overall prediction accuracy (proportion of the correct prediction over the entire data set).(b) Sensitivity (correct prediction of death over all death cases).(c) Specificity (correct prediction of survival over all survival cases).(d) AUC of the receiver operating characteristics (ROC) curve.

Figure 4 .
Figure 4. Survival-related features captured by the U-Nets.During training, CNNs essentially learn "templates/patterns" from training images and apply these templates to analyze and understand images.(a) Image templates that the U-Nets have captured for the segmentation task, in CT and (b) in PET.Note these templates are learned in an unsupervised manner, without any survival-related information provided, despite which these were later discovered to be survival-related.Note that the templates captured by U-Net are characterized by their sensical and interpretable geometric structures .For example, C37399 appears to be a template looking for a tumor-like shape at the top-right corner and a tube-like structure at the bottom-left.In addition, C08680 appears to look for a textural feature of the tumor.(c) In contrast, image templates learned by direct fitting of a CNN model to the survival data28 .Note the features in (c) are less interpretable compared to the U-Net features.

Figure 5 . 19 Figure 6 .
Figure 5. Visualization of the U-Net features.Regions that predicted death of the patients obtained via a guided backpropagation method 30 .Trivially, tumoral regions are highlighted in red in the heatmap.However, some of the heated regions outside of the tumoral volume matched with the actual locations of recurrences and metastases when they were compared with the post-therapeutic images and clinical records, rendering a great potential as a practical, clinical tool for patient-tailored treatment planning in the future.(a) Patient deceased in 0.29 years after the acquisition of the images.(b) Deceased after 4.58 years.(c) Deceased after 7.11 years.

Fig. 1 .
k-medoids feature clustering In this research, several k-medoids clustering experiments were conducted on the training dataset for the cross-validation experiments.Pearson's correlation distance was employed as the distance metric for clustering the features as expressed in: D(X, Y) = 1 − Cov(X, Y) Var(X)Var(Y)

Table 1 .
the variance of the National Institutes of Health/National Cancer Institute grant 1R21CA209874.Dr. Wu's work has been funded in part by the National Institutes of Health grants 1R21CA209874 and R01EB020665.Drs.Baek, Wu, Kim, Smith, Allen, Buatti and Mr. He are the co-inventors of the provisional patent (U.S. App.No. 62/811,326.Systems And Methods For Image Segmentation And Survival Prediction Using Convolutional Neural Networks) based upon the work of this manuscript.Dr. Plichta, Dr. Seyedin, Ms. Gannon, and Ms. Cabel declare no potential conflict of interest.Summary statistics of the data set * Patients with no staging information.