-
PDF
- Split View
-
Views
-
Cite
Cite
Zhiqin Wang, Ruiqing Li, Minghui Wang, Ao Li, GPDBN: deep bilinear network integrating both genomic data and pathological images for breast cancer prognosis prediction, Bioinformatics, Volume 37, Issue 18, September 2021, Pages 2963–2970, https://doi.org/10.1093/bioinformatics/btab185
Close - Share Icon Share
Abstract
Breast cancer is a very heterogeneous disease and there is an urgent need to design computational methods that can accurately predict the prognosis of breast cancer for appropriate therapeutic regime. Recently, deep learning-based methods have achieved great success in prognosis prediction, but many of them directly combine features from different modalities that may ignore the complex inter-modality relations. In addition, existing deep learning-based methods do not take intra-modality relations into consideration that are also beneficial to prognosis prediction. Therefore, it is of great importance to develop a deep learning-based method that can take advantage of the complementary information between intra-modality and inter-modality by integrating data from different modalities for more accurate prognosis prediction of breast cancer.
We present a novel unified framework named genomic and pathological deep bilinear network (GPDBN) for prognosis prediction of breast cancer by effectively integrating both genomic data and pathological images. In GPDBN, an inter-modality bilinear feature encoding module is proposed to model complex inter-modality relations for fully exploiting intrinsic relationship of the features across different modalities. Meanwhile, intra-modality relations that are also beneficial to prognosis prediction, are captured by two intra-modality bilinear feature encoding modules. Moreover, to take advantage of the complementary information between inter-modality and intra-modality relations, GPDBN further combines the inter- and intra-modality bilinear features by using a multi-layer deep neural network for final prognosis prediction. Comprehensive experiment results demonstrate that the proposed GPDBN significantly improves the performance of breast cancer prognosis prediction and compares favorably with existing methods.
GPDBN is freely available at https://github.com/isfj/GPDBN.
Supplementary data are available at Bioinformatics online.
1 Introduction
The most common malignancy among females is breast cancer, which is one of the leading causes of cancer-related deaths in the world (Hortobagyi et al., 2005). As reported by WHO, more than 1.3 million new cases of breast cancer are diagnosed, and the death toll is as high as 458 000 each year (Nguyen et al., 2013). As a very heterogeneous disease, breast cancer has been reported with distinct prognoses on the basis of morphological and molecular stratifications (Huang et al., 2019), which represents the major hurdle for accurate diagnosis and curative therapy. Therefore, there is an urgent need to design computational methods to accurately predict the prognosis of breast cancer, which may further assist clinicians to prescribe the most appropriate therapeutic regime (Sun et al., 2018a,b).
There have been many studies in predicting the prognosis of breast cancer that assign breast cancer patients into a poor and a good prognosis group based on genomic data and/or pathological images (Gevaert et al., 2006; Sun et al., 2018a,b; Yu et al., 2016). Genomic data, especially gene expression profiles (Xu et al., 2012) obtained from high-throughput platforms, has been widely adopted in breast cancer prognosis prediction (Moon et al., 2017; Sun et al., 2018a,b; Wang et al., 2005). The initial studies aim to discover genes that can separate patients with good prognosis from those with poor prognosis. After that, there are plenty of studies generate a multigene predictor based on a hypothesis developed from in vivo or in vitro experiments and then apply it to breast cancer samples (Reis-Filho and Pusztai, 2011). For example, Van De Vijver et al. (2002) recognize a 70-gene prognostic signature from 117 patients with breast cancer by correlating candidate genes with the outcome and determine the optimal set of genes by using leave-one-out cross validation. Wang et al. (2005) further identify a 76-gene signature representing a strong prognostic factor that can be applied to identify patients who have a favorable prognosis. These studies undoubtedly have not only contributed to our understanding of the heterogeneity and complexity of breast cancer behavior but also provided computational methods to distinguish between patients with good and poor prognosis.
Besides aforementioned prognosis prediction methods using genomic data, the continued importance of pathological analysis of tumors should be emphasized, which has been confirmed to provide independent prognostic information of breast cancer (Reis-Filho and Pusztai, 2011). With the emergence of digital whole-slide images, comprehensive computational methods for pathological image analysis have demonstrated promising capability to improve efficiency, accuracy and consistency compared with human examination (Cheng et al., 2017; Courtiol et al., 2019). With rapid growth of computing resources, many computational pathological methods have been proposed for predicting the prognosis of breast cancer (Xu et al., 2016) and a considerable number of other cancers such as lung (Yu et al., 2016; Zhu et al., 2016) and kidney (Cheng et al., 2018) cancers. In this way, hundreds of pathological image features, characterizing cell size, shape, distribution and texture of nuclei, can be extracted from pathological images, which can provide important prognostic information for further investigation.
In addition to aforementioned studies using either genomic data or pathological images, there is an increasing interest in combining both of them that may contribute to more accurate cancer prognosis prediction (Cheng et al., 2017; Ning et al., 2020; Shao et al., 2020). With the fast development of computer-aided technology, machine learning methods are widely applied to predict cancer prognosis. For example, by using SVM with a Gaussian radial basis kernel, Yuan et al. (2012) integrate both genomic data and pathological images to improve the performance of breast cancer prognosis prediction. Sun et al. (2018a,b) develop a multiple kernel learning approach named GPMKL, which employs heterogeneous features extracted from genomic data and pathological images. Besides breast cancer, Cheng et al. (2017) predict the survival outcomes of renal cell carcinoma patients by combining quantitative image features derived from pathological images and eigengenes derived from gene expression profiles. These studies clearly show that genomic data and pathological images are complementary to each other and can acquire better performance of prognosis of patients when employed together.
Recently, deep learning-based approaches for the integration of data from different modalities have been proposed (Ngiam et al., 2011) and successfully applied in cancer prognosis prediction (Cheerla and Gevaert, 2019; Mobadersany et al., 2018; Yao et al., 2017), which are highly flexible and can interpret the complexity of data in a non-linear manner (Huang et al., 2019). For example, Mobadersany et al. (2018) develop a unified framework named genomic survival convolutional neural network (GSCNN) to predict time-to-event outcomes of patients diagnosed with glioma. To derive highly predictive prognostic features, GSCNN incorporate features embedded within pathological images with genomic features by feeding them into a multi-layer neural network that can add additional non-linear transformations to extracted features. Although the prediction accuracy of GSCNN exceeds human experts according to the current clinical standard, the direct combination of image and genomic features in GSCNN is suboptimal as it may ignore the intrinsic relationship of the features across different modalities (Shao et al., 2019; Yao et al., 2017). Alternatively, Yao et al. (2017) introduce an efficient correlation loss function for training deep learning-based survival model by explicitly maximizing the correlation among image and genomic features, and report quite promising performances for lung and brain cancer. Also, Cheerla and Gevaert (2019) exploit similarities between image and genomic features from the same patient via maximizing a cosine similarity-based loss function. The success of these studies highlights the importance of developing sophisticated deep learning-based methods for more accurate breast cancer prognosis prediction by efficiently leveraging both genomic data and pathological images.
In this study, we propose a novel unified framework named genomic and pathological deep bilinear network (GPDBN) for integrating genomic data and pathological images to predict breast cancer prognosis. Our GPDBN framework includes an inter-modality bilinear feature encoding module (Inter-BFEM) and two intra-modality bilinear feature encoding modules (Intra-BFEMs) to achieve effective information between and within the genomic data and pathological images. Given pre-processed genomic data and pathological images, Inter-BFEM generates inter-modality bilinear features to model complex inter-modality relations for fully exploiting intrinsic relationship of the features across different modalities. In addition, Intra-BFEMs calculate the intra-modality bilinear features within each modality for capturing intra-modality relations. We argue that the intra-modality relations play an important role for breast cancer prognosis prediction, as they can provide valuable information complementary to that included in inter-modality relations (Gao et al., 2019). Therefore, GPDBN further combines the intra- and inter-modality bilinear features by using a multi-layer deep neural network. The experimental results verify that our proposed GPDBN framework enhances the performance of breast cancer prognosis prediction and outperforms existing approaches using both genomic data and pathological images.
2 Materials and methods
2.1 Dataset and pre-process
Breast cancer patient samples adopted in this study include matched digital whole-slide images and gene expression profiles (totally 20 436 genes), which are acquired from the Cancer Genome Atlas (TCGA) data portal (Zhu et al., 2014). By following previous work (Cheng et al., 2017), patients with missing follow-up are excluded and finally 345 patients are enrolled in this study. These patients are further categorized into longer-term or shorter-term survivors by the criterion of 5-year survival (Gevaert et al., 2006; Nguyen et al., 2013). Accordingly, shorter-term survivors are labeled as 1 (i.e. poor prognosis) while the longer-term survivors are labeled as 0 (i.e. good prognosis). In consistent with previous prognosis prediction studies (Ching et al., 2018; Huang et al., 2019), 5-fold cross-validation is conducted on the dataset with 80% of the data in each fold used for training and 20% for testing.
Similar to Ding et al. (2016), the genes with missing values (NA) in more than 10% patients are deleted. After that, gene expression profiles with 19 006 genes are normalized with z-score by standardizing the distribution and then discretized into under-expression (-1), over-expression (1) and baseline (0) with z-score threshold of -1 and 1 by following Gevaert et al. (2006). For pathological images, the whole-slide images with 40 magnification are tiled into overlapping 1000 1000 pixels by adopting bftools in the open microscopy environment (Sun et al., 2018a,b; Yu et al., 2016). Next, we select the 10 densest images of each image series ranked by image density defined as the percentage of non-white (all of the red, green and blue values are below 200 in the 24-bit RGB color space) pixels by following previous study (Yu et al., 2016). Moreover, according to Sun et al. (2018a,b), by using CellProfiler we obtain 2343 image features from pathological images, which contain nucleus size, shape, distribution of pixel intensity and texture of nuclei. The aforementioned normalization and discretization steps are also applied to image features to mitigate the heterogeneity gap between different modalities. After that, the pre-processed genomic and pathological image features are fed into the FSelector package (Cheng et al., 2012) implemented by R, respectively, which performs automatic feature selection from the input by ranking informative features. Finally, from the results of FSelector we choose the top 32 genomic and pathological image features, respectively, for further study (Sun et al., 2018a,b; Yu et al., 2016).
2.2 GPDBN
The workflow of GPDBN is shown in Figure 1.Inter-modality bilinear features are firstly generated by Inter-BFEM to model complex inter-modality relations for fully exploiting intrinsic relationship of the features derived from genomic data and pathological images. Meanwhile, intra-modality bilinear features within each modality that are beneficial to prognosis prediction are also extracted by Intra-BFEMs. Finally, the intra- and inter-modality bilinear features are combined by multi-layer deep neural network to take advantage of the complementary information between intra-modality and inter-modality relations. The following subsections describe GPDBN in detail.
2.2.1 Inter-BFEM
2.2.2 Intra-BFEMs
2.2.3 Prognosis prediction
Details of multi-layer deep neural network
| . | Layer . | Details . | Output size . |
|---|---|---|---|
| Multi-layer deep neural network | Input combined features | concatenate the output of Inter-BFEM and Intra-BFEMs as input | (124) |
| Fully connected layer (+ReLU) | 500 neurons | (500) | |
| Dropout layer | P = 0.3 | (500) | |
| Fully connected layer (+ReLU) | 256 neurons | (256) | |
| Dropout layer | P = 0.3 | (256) | |
| Fully connected layer (+ReLU) | 128 neurons | (128) | |
| Dropout layer | P = 0.1 | (128) | |
| Fully connected layer (+ReLU) | 32 neurons | (32) | |
| Dropout layer | P = 0.1 | (32) | |
| Softmax output unit | 2 neurons | (2) |
| . | Layer . | Details . | Output size . |
|---|---|---|---|
| Multi-layer deep neural network | Input combined features | concatenate the output of Inter-BFEM and Intra-BFEMs as input | (124) |
| Fully connected layer (+ReLU) | 500 neurons | (500) | |
| Dropout layer | P = 0.3 | (500) | |
| Fully connected layer (+ReLU) | 256 neurons | (256) | |
| Dropout layer | P = 0.3 | (256) | |
| Fully connected layer (+ReLU) | 128 neurons | (128) | |
| Dropout layer | P = 0.1 | (128) | |
| Fully connected layer (+ReLU) | 32 neurons | (32) | |
| Dropout layer | P = 0.1 | (32) | |
| Softmax output unit | 2 neurons | (2) |
Details of multi-layer deep neural network
| . | Layer . | Details . | Output size . |
|---|---|---|---|
| Multi-layer deep neural network | Input combined features | concatenate the output of Inter-BFEM and Intra-BFEMs as input | (124) |
| Fully connected layer (+ReLU) | 500 neurons | (500) | |
| Dropout layer | P = 0.3 | (500) | |
| Fully connected layer (+ReLU) | 256 neurons | (256) | |
| Dropout layer | P = 0.3 | (256) | |
| Fully connected layer (+ReLU) | 128 neurons | (128) | |
| Dropout layer | P = 0.1 | (128) | |
| Fully connected layer (+ReLU) | 32 neurons | (32) | |
| Dropout layer | P = 0.1 | (32) | |
| Softmax output unit | 2 neurons | (2) |
| . | Layer . | Details . | Output size . |
|---|---|---|---|
| Multi-layer deep neural network | Input combined features | concatenate the output of Inter-BFEM and Intra-BFEMs as input | (124) |
| Fully connected layer (+ReLU) | 500 neurons | (500) | |
| Dropout layer | P = 0.3 | (500) | |
| Fully connected layer (+ReLU) | 256 neurons | (256) | |
| Dropout layer | P = 0.3 | (256) | |
| Fully connected layer (+ReLU) | 128 neurons | (128) | |
| Dropout layer | P = 0.1 | (128) | |
| Fully connected layer (+ReLU) | 32 neurons | (32) | |
| Dropout layer | P = 0.1 | (32) | |
| Softmax output unit | 2 neurons | (2) |
2.3 Training
2.4 Evaluation metrics
3 Results
3.1 Performance evaluation of GPDBN
To verify the effectiveness of our proposed GPDBN in contributing to prognosis prediction of breast cancer, we conduct ablation study comparing different model configurations in 5-fold cross validation. First, we train four unimodal models to evaluate the performance of Intra-BFEMs: (i) BaselineG: a multi-layer deep neural network using genomic features as input, (ii) Intra-BFEMG: an Intra-BFEM using genomic features as input, followed by a multi-layer deep neural network, (iii) BaselineP: a multi-layer deep neural network using pathological image features as input, (iv) Intra-BFEMP: an Intra-BFEM using pathological image features as input, followed by a multi-layer deep neural network. After that, we train three multimodal models to evaluate the performance of Inter-BFEM: (i) BaselineGP: a multi-layer deep neural network using directly concatenated genomic and pathological image features as input, (ii) Inter-BFEM*: an Inter-BFEM using genomic and pathological image features as input, followed by a multi-layer deep neural network, (iii) GPDBN: our proposed deep bilinear network using genomic and pathological image features as input. It is of note that all aforementioned multi-layer deep neural networks have the same hyperparameters as described in Table 1.
Table 2 shows the C-index and AUC values of different models. We can find that Intra-BFEMG and Intra-BFEMP achieve consistently better overall performance than BaselineG and BaselineP, respectively. For example, the C-index and AUC values of Intra-BFEMG are 0.695 and 0.779, while the corresponding C-index and AUC values of BaselineG are 0.674 and 0.734, respectively. It is of note that the average C-index and AUC values over 5-fold cross validation (Supplementary Table S1) show similar performance. Besides, we plot ROC curves to compare the predictive performance in Figure 2, which suggest that our proposed Intra-BFEMs have great efficiency on both genomic data and pathological images. Furthermore, we calculate the corresponding Sn, Acc, Pre and F1 of all compared methods with Sp equal to 90.0% or 95.0% and the histograms of these methods are shown in Figure 3. From these results, we can see that when Sp is equal to 95.0%, by adopting Intra-BFEM Sn is improved by 6.2% and 5.1% on pathological images and genomic data, respectively. Meanwhile, compared with BaselineP and BaselineG, the Pre values of Intra-BFEMP and Intra-BFEMG have an improvement of 16.7% and 3.2% respectively. The above results verify the advantage of using Intra-BFEMs for prognosis prediction of breast cancer in our proposed GPDBN framework by capturing the intra-modality relations within genomic data and pathological images.
The values of Sn, Acc, Pre and F1 of GPDBN for breast cancer prognosis prediction at stringent levels of Sp = 90.0% (left) and Sp = 95.0% (right)
C-index and AUC values of GPDBN for breast cancer prognosis prediction
| Data . | Method . | C-index . | AUC . |
|---|---|---|---|
| Genomic data | BaselineG | 0.674 | 0.734 |
| Intra-BFEMG | 0.695 | 0.779 | |
| Pathological images | BaselineP | 0.569 | 0.571 |
| Intra-BFEMP | 0.578 | 0.585 | |
| Genomic data + Pathological images | BaselineGP | 0.703 | 0.775 |
| Inter-BFEM* | 0.708 | 0.793 | |
| GPDBN | 0.723 | 0.817 |
| Data . | Method . | C-index . | AUC . |
|---|---|---|---|
| Genomic data | BaselineG | 0.674 | 0.734 |
| Intra-BFEMG | 0.695 | 0.779 | |
| Pathological images | BaselineP | 0.569 | 0.571 |
| Intra-BFEMP | 0.578 | 0.585 | |
| Genomic data + Pathological images | BaselineGP | 0.703 | 0.775 |
| Inter-BFEM* | 0.708 | 0.793 | |
| GPDBN | 0.723 | 0.817 |
The bold values show the performances of proposed GPDBN method.
C-index and AUC values of GPDBN for breast cancer prognosis prediction
| Data . | Method . | C-index . | AUC . |
|---|---|---|---|
| Genomic data | BaselineG | 0.674 | 0.734 |
| Intra-BFEMG | 0.695 | 0.779 | |
| Pathological images | BaselineP | 0.569 | 0.571 |
| Intra-BFEMP | 0.578 | 0.585 | |
| Genomic data + Pathological images | BaselineGP | 0.703 | 0.775 |
| Inter-BFEM* | 0.708 | 0.793 | |
| GPDBN | 0.723 | 0.817 |
| Data . | Method . | C-index . | AUC . |
|---|---|---|---|
| Genomic data | BaselineG | 0.674 | 0.734 |
| Intra-BFEMG | 0.695 | 0.779 | |
| Pathological images | BaselineP | 0.569 | 0.571 |
| Intra-BFEMP | 0.578 | 0.585 | |
| Genomic data + Pathological images | BaselineGP | 0.703 | 0.775 |
| Inter-BFEM* | 0.708 | 0.793 | |
| GPDBN | 0.723 | 0.817 |
The bold values show the performances of proposed GPDBN method.
As indicated in Table 2, Intra-BFEMG outperforms Intra-BFEMP with an improvement of more than 11.2% in C-index, which reflects the importance of genomic data for prognosis prediction. Although the prognostic performance on pathological images is worse than genomic data, we can also see its effect in improving performance when integrated with genomic data. For example, the C-index value of BaselineGP is 0.703, which has 13.4% and 2.9% improvement over BaselineP and BaselineG, respectively. Meanwhile, we also find that Inter-BFEM* can obtain higher C-index and AUC values than BaselineGP. For example, the AUC value achieved by Inter-BFEM* is 0.793, compared with 0.775 obtained by BaselineGP. This demonstrates the power of Inter-BFEM in capturing complex inter-modality relations for prognosis prediction. More importantly, by adopting Intra-BFEM and Inter-BFEM simultaneously, GPDBN obtains the best performance with remarkable improvements on both C-index and AUC values in comparison with other methods. For example, the C-index value achieved by GPDBN is 0.723, which has an improvement of 2.0% and 1.5% over BaselineGP and Inter-BFEM*, respectively. Furthermore, we calculate the Sn, Acc, Pre and F1 of all compared methods, and the histograms these methods are displayed in Figure 4. From these results, we can find that GPDBN consistently yields the best performance on all metrics. Taken together, these results demonstrate that our proposed GPDBN can effectively integrate genomic data and pathological images by taking advantage of the complementary information between inter-modality and intra-modality relations obtained by Inter-BFEM and Intra-BFEM, respectively.
The values of Sn, Acc, Pre and F1 of different methods for breast cancer prognosis prediction at stringent levels of Sp = 90.0% (left) and Sp = 95.0% (right)
3.2 Comparison with existing methods
To further evaluate the performance of our proposed GPDBN, we compare it with several existing methods, namely BoostCI, PCRM, superPC, En-Cox, LASSO-Cox, MDNNMD and DeepCorrSurv, and the C-index values of different methods are listed in Table 3. It is obvious that all these methods have satisfying performance by using genomic data. For example, the C-index values obtained by BoostCI and LASSO-Cox are 0.677 and 0.690, respectively, meanwhile deep learning-based MDNNMD and GPDBN have similar results of 0.691 and 0.695, respectively. These results further corroborate the conclusion that genomic data plays a crucial role in prognosis prediction of breast cancer. By using pathological images, we can find that deep learning-based approaches generally exhibit better performance than other methods evaluated in this study. For example, with a C-index value of 0.570, MDNNMD outperforms all other non-deep learning methods except superPC. At the same time, GPDBN achieves the best performance among all investigated methods and obtains a C-index improvement of 1.3% compared with MDNNMD, which indicates the effectiveness of GPDBN in breast cancer prognosis prediction.
Performance comparison of the proposed GPDBN and other methods using C-index value
| . | Genomic data . | Pathological images . | Genomic data + pathological images . |
|---|---|---|---|
| BoostCI | 0.677 | 0.435 | 0.681 |
| PCRM | 0.569 | 0.569 | 0.688 |
| superPC | 0.696 | 0.573 | 0.696 |
| En-Cox | 0.688 | 0.560 | 0.697 |
| LASSO-Cox | 0.690 | 0.564 | 0.700 |
| MDNNMD | 0.691 | 0.570 | 0.704 |
| DeepCorrSurv | NA | NA | 0.694 |
| GPDBN | 0.695 | 0.583 | 0.723 |
| . | Genomic data . | Pathological images . | Genomic data + pathological images . |
|---|---|---|---|
| BoostCI | 0.677 | 0.435 | 0.681 |
| PCRM | 0.569 | 0.569 | 0.688 |
| superPC | 0.696 | 0.573 | 0.696 |
| En-Cox | 0.688 | 0.560 | 0.697 |
| LASSO-Cox | 0.690 | 0.564 | 0.700 |
| MDNNMD | 0.691 | 0.570 | 0.704 |
| DeepCorrSurv | NA | NA | 0.694 |
| GPDBN | 0.695 | 0.583 | 0.723 |
Performance comparison of the proposed GPDBN and other methods using C-index value
| . | Genomic data . | Pathological images . | Genomic data + pathological images . |
|---|---|---|---|
| BoostCI | 0.677 | 0.435 | 0.681 |
| PCRM | 0.569 | 0.569 | 0.688 |
| superPC | 0.696 | 0.573 | 0.696 |
| En-Cox | 0.688 | 0.560 | 0.697 |
| LASSO-Cox | 0.690 | 0.564 | 0.700 |
| MDNNMD | 0.691 | 0.570 | 0.704 |
| DeepCorrSurv | NA | NA | 0.694 |
| GPDBN | 0.695 | 0.583 | 0.723 |
| . | Genomic data . | Pathological images . | Genomic data + pathological images . |
|---|---|---|---|
| BoostCI | 0.677 | 0.435 | 0.681 |
| PCRM | 0.569 | 0.569 | 0.688 |
| superPC | 0.696 | 0.573 | 0.696 |
| En-Cox | 0.688 | 0.560 | 0.697 |
| LASSO-Cox | 0.690 | 0.564 | 0.700 |
| MDNNMD | 0.691 | 0.570 | 0.704 |
| DeepCorrSurv | NA | NA | 0.694 |
| GPDBN | 0.695 | 0.583 | 0.723 |
From Table 3, we can also observe that compared with using single-modality data alone, the application of both genomic data and pathological images enhances the performance for most of the methods. For example, the C-index value of LASSO-Cox using both genomic data and pathological images is increased by 1.0% and 13.6% compared with those obtained by using only genomic data and pathological images, respectively. These suggest that pathological images can provide valuable predictive information additional to those provided by genomic data. Importantly, GPDBN outperforms all non-deep learning methods with remarkable improvement in C-index value, which indicates that sophisticated deep learning-based methods can be advantageous in integrating data from different modalities. At the same time, GPDBN achieves the highest C-index value of 0.723 among all the deep learning-based methods, which demonstrates that compared with MDNNMD and DeepCorrSurv, the proposed GPDBN can efficiently integrate genomic data and pathological images. In addition to C-index value, the ROC curves of different methods using both genomic data and pathological images are plotted in Figure 5, in which GPDBN also achieves the best AUC value of 0.817 among all investigated methods. Taken together, these results suggest that our proposed GPDBN can take advantage of not only inter- but also intra-modality relations to significantly improve the performance of breast cancer prognosis prediction. Moreover, we conduct survival analysis on TCGA breast cancer dataset, in which GPDBN is supervised with the Cox objective function (Chen et al., 2020a,b) and the detailed information is provided in Supplementary Methods and Supplementary Table S2. As shown in Supplementary Table S3, we can find that the average C-index value of GPDBN compares favorably with Pathomic Fusion (Chen et al., 2020a,b) for survival analysis.
ROC curves of GPDBN and other methods by employing both genomic data and pathological images
To further evaluate the performance of GPDBN, Kaplan-Meier curves of different methods are plotted and displayed in Figure 6. We can observe that for non-deep learning-based methods, LASSO-Cox provides better prognostic prediction with a log-rank test value of 6.117E-8 than other methods such as BoostCI (log-rank test value = 0.000010) and superPC (log-rank test value = 4.4754E-7). In comparison with these non-deep learning methods, deep learning-based DeepCorrSurv provides competitive performance (log-rank test value = 2.4049E-11) by considering intrinsic relationship of the features across different modalities in deep learning architecture. Moreover, by effectively integrating genomic data and pathological images, our proposed GPDBN can successfully distinguish shorter-term survivors from longer-term survivors and achieve superior performance (log-rank test value = 6.802E-15) over all other methods. In addition, Kaplan-Meier curves of GPDBN supervised with the Cox objective function and Pathomic Fusion are plotted in Supplementary Figures S1 and S2, and GPDBN provides better survival prediction with a log-rank test P value of 2.0452E-10 than Pathomic Fusion (log-rank test P value = 0.000117). These results further confirm the effectiveness of GPDBN in breast cancer prognosis prediction.
Performance comparison of the proposed GPDBN and other methods using Kaplan–Meier curve
3.3 Visualization of features
To visualize the superiority of the proposed method in prognosis prediction, a popular algorithm t-SNE is used to squeeze abstract features extracted by Intra-BFEM and GPDBFN into a 2-D space. As shown in Figure 7, original genomic and pathological image features are overlapped in mixture, while combined genomic and pathological image features show separate trends, which suggests that the integration of data from different modalities can contribute to more accurate prognosis prediction of breast cancer. Furthermore, we can observe that compared with using original genomic and pathological image features, using abstract features extracted by Intra-BFEM shows an obvious improvement in separating shorter-term survivors and longer-term survivors, which become even more evident when using combined features extracted by GPDBFN. These results show that original genomic and pathological image features can be transformed into meaningful representations by Intra-BFEM, respectively, and GPDBFN can generate better combined representation with stronger discriminant power in distinguishing shorter-term and longer-term survivors.
Visualization of original genomic, pathological image, combined features and the corresponding abstract features extracted by our proposed method. The red star represents shorter-term survivors, and the blue dot represents longer-term survivors
4 Discussion
In this study, we propose a novel deep learning-based GPDBN framework to effectively integrate both genomic data and pathological images for more accurate breast cancer prognosis prediction. Our findings suggest that prognosis prediction methods based on data from different modalities outperform those using single-modality data and our proposed Intra-BFEM and Inter-BFEM can efficiently capture complex relations not only within each modality but also between different modalities. Meanwhile, the comprehensive experimental results show that by using both intra- and inter-modality bilinear features as input, GPDBN compares favorably with many existing methods. Furthermore, the visualization results also indicate its powerful capability in distinguishing shorter-term survivors from longer-term survivors. The main contributions of this work are as follows: (i) we design Inter-BFEM to generate inter-modality bilinear features that can fully exploit intrinsic relationship of the features across different modalities, (ii) two efficient deep bilinear encoding modules named Intra-BFEMs are proposed to calculate intra-modality bilinear features for capturing relations within each modality that are also beneficial for breast cancer prognosis prediction, (iii) by taking advantage of the complementary information between intra-modality and inter-modality relations, the proposed GPDBN shows great strength in integrating data from different modalities, and achieves remarkable performance for predicting the prognosis of breast cancer.
Although GPDBN has enhanced the predictive performance of breast cancer prognosis, there is still considerable room for further expansion and improvement. Firstly, limited by available genomic data and pathological images, this work can further improve prognostic performance by introducing more breast cancer patients. Secondly, other genomic data (e.g. copy number, miRNA expression) have also been reported to be beneficial for cancer prognosis prediction (Huang et al., 2019). To integrate more modalities, computational approaches such as tensor fusion (Chen et al., 2020a,b; Zadeh et al., 2017) can be incorporated into the proposed GPDBN framework in future work. Meanwhile, other computationally efficient bilinear models (Fukui et al., 2016; Gao et al., 2016) can also be explored in the future to reduce memory consumption and computation times when more modalities are adopted. Thirdly, it is possible to further improve our framework by building a pancancer model for prognosis that also predicts prognosis of breast cancer (Cheerla and Gevaert, 2019). Finally, it is important to develop deep learning-based prognosis prediction method with improved interpretability (Ma et al., 2018). In conclusion, we present a novel deep bilinear network for breast cancer prognosis prediction, which has the potential to be extended to integrate data from different modalities for other predictive tasks and provides clues for further cancer prognosis research.
Funding
This work was supported by the National Natural Science Foundation of China [61971393, 61871361, 61571414, 61471331].
Conflict of Interest: none declared.






