Abstract

Imaging genetics provides unique insights into the pathological studies of complex brain diseases by integrating the characteristics of multi-level medical data. However, most current imaging genetics research performs incomplete data fusion. Also, there is a lack of effective deep learning methods to analyze neuroimaging and genetic data jointly. Therefore, this paper first constructs the brain region-gene networks to intuitively represent the association pattern of pathogenetic factors. Second, a novel feature information aggregation model is constructed to accurately describe the information aggregation process among brain region nodes and gene nodes. Finally, a deep learning method called feature information aggregation and diffusion generative adversarial network (FIAD-GAN) is proposed to efficiently classify samples and select features. We focus on improving the generator with the proposed convolution and deconvolution operations, with which the interpretability of the deep learning framework has been dramatically improved. The experimental results indicate that FIAD-GAN can not only achieve superior results in various disease classification tasks but also extract brain regions and genes closely related to AD. This work provides a novel method for intelligent clinical decisions. The relevant biomedical discoveries provide a reliable reference and technical basis for the clinical diagnosis, treatment and pathological analysis of disease.

Introduction

The brain is the control center of human physiological activities, whereas brain diseases with complex causes have now become the major health concern of modern people [1, 2]. For a long time, scholars have found that there exist complex functional correlations between brain activities and gene expressions, which can be reflected through neuroimaging and genetic data [3]. On the one hand, neuroimaging data such as magnetic resonance imaging (MRI) can intuitively show the operative condition of the brain with radiographic techniques. Image characteristics are also the most common diagnostic indices [4, 5]. On the other hand, the physiological activities of the brain are strictly controlled by genetic materials like DNA and RNA, whose abnormalities will have a profound impact on the development of brain diseases [6–9] Specifically, single nucleotide polymorphism (SNP) can more accurately reflect the disease-related mutation [10, 11]. In such a context, imaging genetics has emerged as a new field to bridge these two aspects, where scholars can thus fully understand the internal genetic causes of brain lesions [12, 13].

Data from different omics usually contain complementary biological information, which can provide a wider horizon to pathology studies. In existing imaging genetics research, though fair performance improvements have been accomplished, the practice of data fusion is either to simply splice the feature matrices of the multi-level data [14] or to design regularization terms according to the characteristics of multimodal data to improve the objective functions of model training [15, 16]. In short, the fusion of imaging genetic data is not intuitive enough, which not only reduces the interpretability of intelligent learning methods but makes it difficult to ensure that the extracted fusion features have clinical importance. Comparatively, graph-based data representation can flexibly express the non-Euclidean relationships among nodes, and thus attracts increasing attention [17–19]. For example, Yang et al. [20] proposed a graph embedding method to study the complex association between circular RNAs and diseases, designed a deep network to extract discriminant line features and achieved good classification results in real datasets. Song et al. [21] constructed functional graphs and structural graphs with functional MRI (fMRI) and diffusion tensor imaging data and fused the multimodal information into edges through an effective calibration mechanism. Therefore, the first contribution of this paper is that we first introduce the network-based data representation into the field of imaging genetics. Specifically, the brain region-gene network (BG-network) is constructed based on the fMRI and SNP data, which can bring two benefits. First, the nonlinear association patterns between brain regions and genes are explicitly expressed as the edges in the network. Second, based on the topological structure of the network, we define the structural information of each node, and creatively propose a structural information aggregation model to realize the iterative evolution of the BG-networks, and then obtain the key structural sub-networks that can accurately characterize the distinctive characteristics of Alzheimer’s disease (AD) patients.

With the development of computational technology, computer-aided diagnosis and treatment based on artificial intelligence have become a vital means for disease risk prediction and diagnosis, which has improved the development of intelligent medicine [22–24]. The machine learning method has become a popular method to extract distinctive features that are more conducive to the classification of brain diseases [25–29]. For example, Illan et al. [30] used Bayesian networks to simulate the dependency between affected areas of AD to assist the diagnosis based on MRI data. Abdul et al. [31] developed a convolutional neural network (CNN) classification framework for AD, and the proposed framework achieved satisfactory results on a variety of classification tasks. Though deep learning methods often outperform machine learning due to their strong representation and modeling abilities, the inherent black-box characteristics of the deep learning make it difficult for scholars to understand the concrete process of model training, which reduces the application scope of deep learning approaches. In this paper, we have made improvements suitable for multi-level data analysis to the basic framework of the generative adversarial network (GAN). In the generator and discriminator, we design novel bidirectional cross convolution (BDC-C) and diffusion deconvolution (D-DC) operations according to the established information aggregation model, which helps our method to accurately identify the pathogenic features of AD.

This paper adopts a modeling-before-learning research pipeline and studies AD based on imaging genetic data. First, based on the association analysis of multimodal number features, BG-networks are constructed to represent the association pattern between the brain regions and genes. Second, we construct a novel feature information aggregation model, which clearly describes the information dissemination process between nodes. Then, we build a feature information aggregation and diffusion generative adversarial network (FIAD-GAN) based on the model, where the specific convolution operation is defined, and the GAN structure and adversarial training can accurately extract distinctive features. Finally, we test the performance of FIAD-GAN in the task of disease classification and feature extraction based on the imaging genetic data of AD patients. The results not only show the advancement of FAID-GAN compared with other competitive methods but also reveal multimodal biomarkers that may attribute good prospects in clinical diagnosis and further pathological research.

Methods

Framework

In this study, we propose the feature information aggregation model and a FIAD-GAN method to classify and extract the pathogenesis of AD. Figure 1 shows the overall framework of FIAD-GAN. First, the BG-networks are constructed based on brain imaging genetic data. Second, the feature information aggregation model is built, and the generator consisting of convolutional and deconvolutional layers is established accordingly, where the BDC-C and D-DC operation are designed for feature enhancement and network reconstruction. Finally, the discriminator is built to classify the input network and calculate the loss function. After the training, FIAD-GAN can accurately identify patients and extract diseased brain regions, causative genes and the association pattern among those pathogenies.

Framework of the proposed FIAD-GAN.
Figure 1

Framework of the proposed FIAD-GAN.

Datasets and preprocessing

The data of 870 samples are collected from the Alzheimer’s Disease Neuroimaging Initiative (ADNI). The use of data is approved and authorized by ANDI, and the demographic characteristics are listed in Table 1. The data are collected and processed according to strict standards, which ensures homogeny. We have tested the difference in gender and age through the Chi-square test and t-test, and the respective P-values are 0.274 and 0.007. The former is greater than 0.05 and the latter is <0.05, indicating no significant difference. Each sample contains SNP data and fMRI data, and the preprocessing of data is based on Plink and DPARSF software [32, 33]. The main steps of SNP preprocessing include quality control on the deletion rate, heterozygosity and gender, SNP screening, and encoding (assuming that the C of a locus has minimum allele frequencies, the phenotypes of CC, GG and CG are, respectively, encoded as −1, 1 and 0). The main steps of fMRI data preprocessing include time layer alignment and head movement correction, spatial normalization and smoothing, delinearization and filtering, and covariate regression. The length of the series and sequences is determined as 70, which is a compromised result after evaluating the information richness and the balance of multimodal data. If the length is too short, there will be too less information contained in the series and sequences to afford effective analysis. If the length is too long, too few genes will be retained, making the modality balance hard to be kept. As a result, the sequences of 45 genes are obtained for each sample. Trimmed to the same length of 70, the time series of 116 brain regions were acquired based on the anatomical automatic labeling template, which is a commonly used template in imaging analysis [34].

Table 1

Basic information of the data

VariablesNCEMCILMCIAD
Sample number237197203233
Age (mean ± SD)73.03 ± 6.0673.43 ± 5.4574.83 ± 7.9271.55 ± 5.90
Gender (female/male)142/95108/89108/95151/82
VariablesNCEMCILMCIAD
Sample number237197203233
Age (mean ± SD)73.03 ± 6.0673.43 ± 5.4574.83 ± 7.9271.55 ± 5.90
Gender (female/male)142/95108/89108/95151/82

All data are from the ADNI databases.

Table 1

Basic information of the data

VariablesNCEMCILMCIAD
Sample number237197203233
Age (mean ± SD)73.03 ± 6.0673.43 ± 5.4574.83 ± 7.9271.55 ± 5.90
Gender (female/male)142/95108/89108/95151/82
VariablesNCEMCILMCIAD
Sample number237197203233
Age (mean ± SD)73.03 ± 6.0673.43 ± 5.4574.83 ± 7.9271.55 ± 5.90
Gender (female/male)142/95108/89108/95151/82

All data are from the ADNI databases.

Construction of BG-network

It has been confirmed that the expression of genes at the microscopic level can lead to macroscopic changes in brain regions, and these associations may be closely related to AD [35–37]. To better characterize the inter-modal associations, the BG-network is constructed based on the sequence of genes and brain regions. The Pearson correlation coefficient (PCC) between nodes is calculated as the connection strength of each edge, which is presented as:
(1)
where |${v}_a$| or |${v}_b$|is the time series or genetic sequence, |$len$| is the node sequence length. Accordingly, |$W\in{R}^{N\times N}$| is defined as the weight matrix of the BG-network, |$N$| represents the node number.

Feature information aggregation model of BG-network

In this section, the feature information aggregation model is designed for the key feature information extraction from BG-networks. The model can aggregate disease-specific information into key structural sub-networks by evolving the initial BG-network. Therefore, diseased brain regions and causative genes and their association patterns can be found according to the key structural sub-networks. The elements of the feature information aggregation model are as follows:

  • BG-network set: |$G=\{{G}_i|i=1,\dots, m\}$|⁠, |${G}_i$|denotes the BG-network of the |$i$|-th sample (NC or AD patient).

  • Edge weights: |${W}^{(i,n)}=\{{W}_{pq}^{(i,n)}|i=1,\dots, m,1\le p,q\le N\}$|⁠, |${W}_{pq}^{(i,0)}$| denotes the weights of edge |${E}_{pq}$| at the initial BG-network of |$i$|-th sample, which is obtained by calculating the PCC of nodes |$p$| and |$q$|⁠. |${W}_{pq}^{(i,n)}$| denotes the weights of edge |${E}_{pq}$| after the |$n$|-th feature information aggregation of the |$i$|-th sample, with larger weights indicating a stronger degree of connection between nodes |$p$| and |$q$|⁠.

  • Key structural sub-network: |${KG}^i=\{({E}_{pq},{W}_{pq}^{(i,n)})|1\le p,q\le N\}$|⁠, |${KG}^i$| denotes the key structural sub-network obtained from the |$i$|-th sample BG-network after |$n$| times of feature information aggregation.

According to the above elements, the key structural sub-network is finalized through feature information aggregation. The following equations represent the feature information aggregation model.

Equation (2) denotes the update of the edge weight |${W}^{(i,n)}$| during the |$n$|-th feature information aggregation process in the BG-network of |$i$|-th sample, where |$\beta \sum_{k=1}^n({W}_{pk}^{(i,n-1)}+{W}_{kq}^{(i,n-1)})$| denotes the feature information that edge |${E}_{pq}$| aggregates from its neighboring edges.

Equation (3) is the calculation of the key structure sub-network based on |${W}^{(i,n)}$|⁠.

Equation (4) represents the |$i$|-th sample classification, where |$K{G}^i$| represents the key structure sub-network of the |$i$|-th sample.

Equation (5) shows the process of feature extraction, where |$CG$| represents causative genes, |$DBR$| represents diseased brain regions and |$AP$| is the association patterns of brain regions and genes.

FIAD-GAN

GAN is becoming a popular research topic in deep learning because it can weigh the synergies and differences between data more accurately [38]. Combining GAN with the feature information aggregation model, FIAD-GAN framework is designed. The details are specified below.

Generator

As shown in Figure 2, the generator consists of a convolution part and a deconvolution part. The concrete description is as follows.

Flowchart of generator.
Figure 2

Flowchart of generator.

On the one hand, the weight matrix of the BG-network is updated with each iteration of the BDC-C operation, which leads to the topology change of the BG-network. In the existing deep learning methods based on network structure, cross convolution has been used to realize the feature information aggregation of network nodes in their neighborhood. However, the original approach only considers the one-way information propagation process from the source node to the target node, ignoring the bidirectional propagation of information on the same edge that may have different effects on the edge weight. Based on the established information aggregation model, BDC-C layers are designed, which decompose the information aggregation process between each pair of nodes into two directional edges in opposite directions, and therefore the key structural information can be captured more accurately. The following describes the specific |$n$|-order BDC-C operation. Based on the weight matrix |${W}^{(n-1)}$| of the |$n-1$| order BG-network, each edge aggregates the feature information from the |$n$|-order neighboring edges in the |$n$|-order BDC-C operation. As a result, the updated weight matrix |${W}^{(n)}$|is obtained by:
(6)
where |${b}^{(n)}$| is the bias of the |$n$|-order BDC-C operation, |${A}^{(n)}\in{R}^{N\times N}$| denotes the |$n$|-th order BDC-C kernel, |$r$| and |$c$| are the row and column of the matrix, respectively, |$ReLU(\bullet )$| is the activation functions and |${W}_{rc}^{(n)}$| is the updated value of |${W}_{rc}^{(n-1)}$| through the |$n$|-order BDC-C operation.
On the other hand, the D-DC operation is designed to reconstruct the generative BG-network from the key structural sub-network. In the D-DC process proposed in this paper, each edge only diffuses the feature information into the first-order domain, so the shape of the deconvolution kernel is also a cross shape rather than a traditional square. Such a design is also suitable for our network-based data representation. The following is the process of the |$n$|-order D-DC operation. The input of the |$n$|-order D-DC operation is the weight matrix |${W}^{\prime (n-1)}$| obtained from the |${W}^{(n)}$| after |$n$|-|$1$| times D-DC operation. Then, |${W}^{\prime (n-1)}$| is deconvoluted by |$n$|-order diffusion, with each edge absorbing feature information from its |$n$|-order neighborhood. The updated weight matrix |${W}^{\prime (n)}$|is calculated by:
(7)
where |${b}^{\prime (n)}$| is the bias of the |$n$|-order D-DC operation, |${B}^{(n)}\in{R}^{N\times N}$| is the |$n$|-order D-DC kernel and |${W}_{rc}^{\prime (n)}$|is the value of |${W}_{rc}^{\prime (n-1)}$|updated by the |$n$|-order D-DC operation. Generally, the layer number |$n$| can be any integer number greater than 1. In this paper, however, only two BDC-C layers and two D-DC layers are included. In general, stacking excessive convolution layers may cause over-smoothing and reduced accuracy. As experiments indicate, the two BDC-C layers can capture > 99% of the feature information in the BG-networks, so it is unnecessary to add more BDC-C layers that may bring extra calculation load.

Compared with the ordinary convolution and deconvolution in the traditional GAN-based approaches, the design of BDC-C and D-DC operations in this paper is based on more intuitive and reasonable mathematical modeling, and thus the interpretability and performance of the deep learning framework are improved. The former is used to aggregate the structural information in BG-network to obtain the key structural sub-network representing the association pattern between the multimodal data, from which we can better determine the discriminative features of AD patients. The latter is used for reconstructing the BG-network based on the key structural sub-network as the output of the generator.

Discriminator

The discriminator consists of BDC-C layers and fully connected (FC) layers and aims to discriminate whether the input BG-network is real or generated and classify the sample.

The purpose of the BDC-C layers is to enable edges to aggregate key feature information in their neighborhood, whose input is either the real BG-network or the generative BG-network reconstructed by the generator. The process of convolution is demonstrated using the n-order BDC-C as the example, which is defined as follows:
(8)
where |${c}^{(n)}$| denotes the bias of the |$n$|-order BDC-C operations, |${C}^{(n)}\in{R}^{N\times N}$| denotes the |$n$|-order BDC-C total kernel, |$r$| and |$c$| are the row and column number of the matrix, and |${W}_{rc}^{(n)}$| is the updated weight of |${W}_{rc}^{(n-1)}$| by the BDC-C operations.
The purpose of the FC layer is to extract the deeper feature information of the BG-network, and to determine the truth and class of the input BG-network. The equation of the |$L$|-layer FC layer is as follows:
(9)
where |${X}^0= Ran({W}^{(n)})$|⁠, |${W}^{(n)}$|is the output of the BDC-C layer, |$Ran(\cdotp )$| denotes the straightening operation by row, |${W}^l$| denotes the weight matrix of the |$l$|-th FC layer; |$Sigmoid(\bullet )$| and |$Softmax(\bullet )$| are the activation functions. The output of |$Sigmoid$| indicates whether the input network is real or fake. The output of |$Softmax$| indicates the sample belongs to which category.

Loss function

The loss function of FIAD-GAN has two parts: the difference loss value |${L}_1$| between generative and real data, and the adversarial loss value |${L}_2$| between the generator and the discriminator. |${L}_1$| is defined as:
(10)
where |${P}_{data}(x)$| is the real sample distribution, |${W}_{ij}^{x,(0)}$| is the element of the |$x$|-th BG-network weight matrix and |${W}_{ij}^{x,\prime (n)}$| denotes the weight value between nodes |$i$| and |$j$| in the weight matrix generated by the generator.
The adversarial loss |${L}_2$|⁠, which constitutes the adversarial training between generator and discriminator, is defined by the following equation:
(11)
where |$G(\bullet )$| and |$D(\bullet )$| are the representative functions of the generator and discriminator, |${P}_d(x)$| represents the initial sample distribution and |${P}_z(z)$| is the sample distribution of the generated BG-network reconstructed by the generator |$G$|⁠.
In summary, the total function loss of FIAD-GAN is defined as follows:
(12)

Feature extraction

After FIAD-GAN converges, the generator captures the key feature information of the BG-networks and the discriminator can accurately classify the samples. Accordingly, the diseased brain regions, causative genes and the association patterns of AD-related pathogenies can be further extracted.

First, based on the results of the FC layer, the predicted probability score of the sample is defined by:
(13)
where |${P}_i$| and |${R}_i$| are the predicted result and real label. |${P}_i=({P}_{i_{11}},{P}_{i_{12}})$| is a 2D vector, where |${P}_{i_{11}}$| and |${P}_{i_{12}}$| denote the probability that the sample is predicted to be a NC or an AD patient, respectively, |${T}_i$| is a vector of |${R}^{1\times 2}$|⁠, |${T}_i={(1,0)}^T$| represents the sample is normal, and |${T}_i={(0,1)}^T$| represents the sample is AD.
Second, based on the key structure sub-network of the sample and the predicted probability score, the key feature matrix |$FW$| is obtained by
(14)
where |$K{W}^i$| is the weight matrix of the key structural sub-network, i.e. the output of the BDC-C layers.
Similarly, the upper triangle values in |$FW$| are the importance scores of the corresponding edges. For a specific edge |${E}_{v_r{v}_c}$|⁠, the importance score |${Score}_{E_{v_r{v}_c}}$| can be defined as:
(15)
where |${Score}_{E_{v_r{v}_c}}$| denotes the importance score of edge |${E}_{v_r{v}_c}$|⁠.

The |${Score}_{v_r}(r=1,\dots, N)$| and |${Score}_{E_{v_r{v}_c}}(r,c=1,\dots, N)$| are sorted to obtain the set of node features |${Sub}_V$| and the set of edge features |${Sub}_E$|⁠. The optimal subsets of node and edge features were further selected by incremental search.

Evaluation indicators

Four traditional assessment indicators are applied for performance evaluation in this study, i.e. accuracy (ACC), sensitivity (SEN), specificity (SPE) and balance accuracy (BAC). The indicators are defined by:
(16)
(17)
(18)
(19)
where |$TP$|⁠, |$TN$|⁠,|$FP$|and|$FN$| denote true positive, true negative, false positive and false negative, respectively.

Results and discussion

Performance evaluation

Disease classification is the first and most important application of the method. With the ratio of 8:2, the training set and test set are derived from the original data and kept in the proportion of each class of sample. Taking the division of AD patients and normal controls (NCs) as an example, 189 of 237 NCs and 186 of 233 AD patients are divided into the training set, whereas the remaining data will form the test set. The classification performance of the competing methods was evaluated by the aforementioned four metrics. The baseline methods include RF, SVM, CNN and graph convolutional network (GCN), and two GAN methods whose discriminator is implemented respectively by CNN (CNN-GAN) and GCN (GCN-GAN). The structure of CNN-GAN and GCN-GAN was the same as FIAD-GAN. Therefore, they could be regarded as two variants of our method.

Table 2 shows the classification performance results on three tasks, and Figure 3 depicts the corresponding radar plots. It could be concluded that methods based on brain imaging genetic data generally outperform those based on a single modality of data. Also, the GAN-based methods outperformed the others. In addition, FIAD-GAN had the largest radar plot area among all tasks and all modalities. As the results indicate, the proposed algorithms of BDC-C and D-DC could fully use the complementary information in imaging genetic data, and accurately extract brain regions and genes highly related to diseases, providing an efficient new method for the diagnosis and pathogeny detection of AD.

Table 2

Classification performance results of different methods on three tasks under different modal data

MethodModalNC versus ADEMCI versus LMCILMCI versus AD
ACC (%)SEN (%)SPE (%)BAC (%)ACC (%)SEN (%)SPE (%)BAC (%)ACC (%)SEN (%)SPE (%)BAC (%)
RFfMRI60.6455.5670.9763.2660.0056.4572.2264.3459.3857.8165.6361.72
SNP59.5752.6370.2761.4557.5057.1458.8257.9858.3357.3860.0058.69
Dual64.8970.5958.1464.3663.7559.3276.1967.7663.5462.0765.7963.93
SVMfMRI65.9650.0080.0065.0062.5062.0763.6462.8564.5859.0274.2966.65
SNP63.8360.0070.5965.2960.0060.3459.0959.7261.4657.1469.7063.42
Dual68.0865.5269.2367.3771.2566.6780.7773.7267.7166.6769.2467.95
CNNfMRI74.4772.9276.0974.5067.5075.5158.0666.7970.8371.1570.4570.80
SNP69.1563.1677.7870.4766.2562.5075.0068.7568.7579.1758.3368.75
Dual75.5372.9278.2675.5970.0066.0477.7971.9172.9472.0073.9172.96
GCNfMRI74.4770.0079.5574.7768.7564.2979.1771.7370.8364.2980.0072.14
SNP72.3467.3178.5772.9467.5077.2755.5666.4166.6761.0275.6868.35
Dual76.6070.0084.0977.0572.5070.5975.8673.2373.9673.4774.4773.97
CNN-GANfMRI77.6674.4782.9878.7273.7572.9275.0073.9675.0084.9162.7973.85
SNP74.4771.4277.7874.6072.5067.9281.4874.7072.9267.9279.0773.50
Dual80.8583.3380.2383.8777.5072.9284.3878.6578.1378.5777.7878.17
GCN-GANfMRI78.7276.0981.2578.6775.0060.0081.8270.9176.0472.9279.1776.04
SNP75.5367.6580.0073.8273.7571.4377.4274.4273.9675.4772.0973.78
Dual81.9189.4780.0084.7478.7575.0084.3879.6979.1778.2680.0079.13
FIAD-GANfMRI87.2377.2790.2883.3885.0085.7184.2184.9684.3885.3783.6484.50
SNP84.4887.5081.4884.4983.7588.3778.3883.3883.3381.4084.9183.15
Dual93.6295.3592.1693.7591.2591.6790.6391.1591.9592.1191.8491.97
MethodModalNC versus ADEMCI versus LMCILMCI versus AD
ACC (%)SEN (%)SPE (%)BAC (%)ACC (%)SEN (%)SPE (%)BAC (%)ACC (%)SEN (%)SPE (%)BAC (%)
RFfMRI60.6455.5670.9763.2660.0056.4572.2264.3459.3857.8165.6361.72
SNP59.5752.6370.2761.4557.5057.1458.8257.9858.3357.3860.0058.69
Dual64.8970.5958.1464.3663.7559.3276.1967.7663.5462.0765.7963.93
SVMfMRI65.9650.0080.0065.0062.5062.0763.6462.8564.5859.0274.2966.65
SNP63.8360.0070.5965.2960.0060.3459.0959.7261.4657.1469.7063.42
Dual68.0865.5269.2367.3771.2566.6780.7773.7267.7166.6769.2467.95
CNNfMRI74.4772.9276.0974.5067.5075.5158.0666.7970.8371.1570.4570.80
SNP69.1563.1677.7870.4766.2562.5075.0068.7568.7579.1758.3368.75
Dual75.5372.9278.2675.5970.0066.0477.7971.9172.9472.0073.9172.96
GCNfMRI74.4770.0079.5574.7768.7564.2979.1771.7370.8364.2980.0072.14
SNP72.3467.3178.5772.9467.5077.2755.5666.4166.6761.0275.6868.35
Dual76.6070.0084.0977.0572.5070.5975.8673.2373.9673.4774.4773.97
CNN-GANfMRI77.6674.4782.9878.7273.7572.9275.0073.9675.0084.9162.7973.85
SNP74.4771.4277.7874.6072.5067.9281.4874.7072.9267.9279.0773.50
Dual80.8583.3380.2383.8777.5072.9284.3878.6578.1378.5777.7878.17
GCN-GANfMRI78.7276.0981.2578.6775.0060.0081.8270.9176.0472.9279.1776.04
SNP75.5367.6580.0073.8273.7571.4377.4274.4273.9675.4772.0973.78
Dual81.9189.4780.0084.7478.7575.0084.3879.6979.1778.2680.0079.13
FIAD-GANfMRI87.2377.2790.2883.3885.0085.7184.2184.9684.3885.3783.6484.50
SNP84.4887.5081.4884.4983.7588.3778.3883.3883.3381.4084.9183.15
Dual93.6295.3592.1693.7591.2591.6790.6391.1591.9592.1191.8491.97

The bold values indicate the highest results of each indicator in each group of experiments, and the proposed approach always performs the best.

Table 2

Classification performance results of different methods on three tasks under different modal data

MethodModalNC versus ADEMCI versus LMCILMCI versus AD
ACC (%)SEN (%)SPE (%)BAC (%)ACC (%)SEN (%)SPE (%)BAC (%)ACC (%)SEN (%)SPE (%)BAC (%)
RFfMRI60.6455.5670.9763.2660.0056.4572.2264.3459.3857.8165.6361.72
SNP59.5752.6370.2761.4557.5057.1458.8257.9858.3357.3860.0058.69
Dual64.8970.5958.1464.3663.7559.3276.1967.7663.5462.0765.7963.93
SVMfMRI65.9650.0080.0065.0062.5062.0763.6462.8564.5859.0274.2966.65
SNP63.8360.0070.5965.2960.0060.3459.0959.7261.4657.1469.7063.42
Dual68.0865.5269.2367.3771.2566.6780.7773.7267.7166.6769.2467.95
CNNfMRI74.4772.9276.0974.5067.5075.5158.0666.7970.8371.1570.4570.80
SNP69.1563.1677.7870.4766.2562.5075.0068.7568.7579.1758.3368.75
Dual75.5372.9278.2675.5970.0066.0477.7971.9172.9472.0073.9172.96
GCNfMRI74.4770.0079.5574.7768.7564.2979.1771.7370.8364.2980.0072.14
SNP72.3467.3178.5772.9467.5077.2755.5666.4166.6761.0275.6868.35
Dual76.6070.0084.0977.0572.5070.5975.8673.2373.9673.4774.4773.97
CNN-GANfMRI77.6674.4782.9878.7273.7572.9275.0073.9675.0084.9162.7973.85
SNP74.4771.4277.7874.6072.5067.9281.4874.7072.9267.9279.0773.50
Dual80.8583.3380.2383.8777.5072.9284.3878.6578.1378.5777.7878.17
GCN-GANfMRI78.7276.0981.2578.6775.0060.0081.8270.9176.0472.9279.1776.04
SNP75.5367.6580.0073.8273.7571.4377.4274.4273.9675.4772.0973.78
Dual81.9189.4780.0084.7478.7575.0084.3879.6979.1778.2680.0079.13
FIAD-GANfMRI87.2377.2790.2883.3885.0085.7184.2184.9684.3885.3783.6484.50
SNP84.4887.5081.4884.4983.7588.3778.3883.3883.3381.4084.9183.15
Dual93.6295.3592.1693.7591.2591.6790.6391.1591.9592.1191.8491.97
MethodModalNC versus ADEMCI versus LMCILMCI versus AD
ACC (%)SEN (%)SPE (%)BAC (%)ACC (%)SEN (%)SPE (%)BAC (%)ACC (%)SEN (%)SPE (%)BAC (%)
RFfMRI60.6455.5670.9763.2660.0056.4572.2264.3459.3857.8165.6361.72
SNP59.5752.6370.2761.4557.5057.1458.8257.9858.3357.3860.0058.69
Dual64.8970.5958.1464.3663.7559.3276.1967.7663.5462.0765.7963.93
SVMfMRI65.9650.0080.0065.0062.5062.0763.6462.8564.5859.0274.2966.65
SNP63.8360.0070.5965.2960.0060.3459.0959.7261.4657.1469.7063.42
Dual68.0865.5269.2367.3771.2566.6780.7773.7267.7166.6769.2467.95
CNNfMRI74.4772.9276.0974.5067.5075.5158.0666.7970.8371.1570.4570.80
SNP69.1563.1677.7870.4766.2562.5075.0068.7568.7579.1758.3368.75
Dual75.5372.9278.2675.5970.0066.0477.7971.9172.9472.0073.9172.96
GCNfMRI74.4770.0079.5574.7768.7564.2979.1771.7370.8364.2980.0072.14
SNP72.3467.3178.5772.9467.5077.2755.5666.4166.6761.0275.6868.35
Dual76.6070.0084.0977.0572.5070.5975.8673.2373.9673.4774.4773.97
CNN-GANfMRI77.6674.4782.9878.7273.7572.9275.0073.9675.0084.9162.7973.85
SNP74.4771.4277.7874.6072.5067.9281.4874.7072.9267.9279.0773.50
Dual80.8583.3380.2383.8777.5072.9284.3878.6578.1378.5777.7878.17
GCN-GANfMRI78.7276.0981.2578.6775.0060.0081.8270.9176.0472.9279.1776.04
SNP75.5367.6580.0073.8273.7571.4377.4274.4273.9675.4772.0973.78
Dual81.9189.4780.0084.7478.7575.0084.3879.6979.1778.2680.0079.13
FIAD-GANfMRI87.2377.2790.2883.3885.0085.7184.2184.9684.3885.3783.6484.50
SNP84.4887.5081.4884.4983.7588.3778.3883.3883.3381.4084.9183.15
Dual93.6295.3592.1693.7591.2591.6790.6391.1591.9592.1191.8491.97

The bold values indicate the highest results of each indicator in each group of experiments, and the proposed approach always performs the best.

The radar plot of FIAD-GAN method and other related methods, including RF, SVM, CNN, GCN, CNN-GAN and GCN-GAN. (A) AD versus NC task based on fMRI data; (B) EMCI versus LMCI task based on fMRI data; (C) LMCI versus AD task based on fMRI data; (D) AD versus NC based on SNP data tasks; (E) EMCI versus LMCI task based on SNP data; (F) LMCI versus AD task based on SNP data; (G) AD versus NC task based on fMRI and genetic data; (H) EMCI versus LMCI task based on fMRI and genetic data; (I) LMCI versus AD task based on fMRI and genetic data.
Figure 3

The radar plot of FIAD-GAN method and other related methods, including RF, SVM, CNN, GCN, CNN-GAN and GCN-GAN. (A) AD versus NC task based on fMRI data; (B) EMCI versus LMCI task based on fMRI data; (C) LMCI versus AD task based on fMRI data; (D) AD versus NC based on SNP data tasks; (E) EMCI versus LMCI task based on SNP data; (F) LMCI versus AD task based on SNP data; (G) AD versus NC task based on fMRI and genetic data; (H) EMCI versus LMCI task based on fMRI and genetic data; (I) LMCI versus AD task based on fMRI and genetic data.

Evaluation of network construction method

To verify the superiority of the BG-network, the brain networks were constructed through single fMRI data and the gene networks were constructed through single SNP data. Meanwhile, Kendall correlation coefficient (KCC), Spearman correlation coefficient (SCC) and PCC were selected to verify the validity of data fusion.

Table 3 presents the classification performance of FIAD-GAN under different network construction methods. In the classification task distinguishing AD and NC, the PCC network construction had the best classification performance using both unimodal and multi-modal data. The method had the best performance when using multimodal data. Similarly, it could be observed that PCC with multimodal data construction performed the best in the other two tasks, which indicated that the PCC was able to better detect the association information in multimodal data, which can improve the classification of diseases.

Table 3

Performance comparison of different network construction methods based on different modal data

DataMethodNC versus ADEMCI versus LMCILMCI versus AD
ACC (%)SEN (%)SPE (%)BAC (%)ACC (%)SEN (%)SPE (%)BAC (%)ACC (%)SEN (%)SPE (%)BAC (%)
fMRISCC78.7276.0981.2578.6775.0072.9278.1375.5277.0873.4780.8577.16
KCC79.7977.7881.6379.7176.2575.0078.1276.5678.1374.4781.6378.05
PCC87.2377.2790.2883.3885.0085.7184.2184.9684.3885.3783.6484.50
SNPSCC75.5376.674.4775.5372.5071.1575.0073.0873.9670.0078.2674.13
KCC77.6674.4780.8577.6673.7571.4277.4274.4276.0473.4778.7276.10
PCC84.4887.5081.4884.4983.7588.3778.3883.3883.3381.4084.9183.15
DualSCC82.9883.3382.6983.0178.7576.0982.3579.2279.1778.2680.0079.13
KCC84.0485.3783.0284.1980.0078.2682.3580.2180.2179.5580.7780.16
PCC93.6295.3592.1693.7591.2591.6790.6391.1591.9592.1191.8491.97
DataMethodNC versus ADEMCI versus LMCILMCI versus AD
ACC (%)SEN (%)SPE (%)BAC (%)ACC (%)SEN (%)SPE (%)BAC (%)ACC (%)SEN (%)SPE (%)BAC (%)
fMRISCC78.7276.0981.2578.6775.0072.9278.1375.5277.0873.4780.8577.16
KCC79.7977.7881.6379.7176.2575.0078.1276.5678.1374.4781.6378.05
PCC87.2377.2790.2883.3885.0085.7184.2184.9684.3885.3783.6484.50
SNPSCC75.5376.674.4775.5372.5071.1575.0073.0873.9670.0078.2674.13
KCC77.6674.4780.8577.6673.7571.4277.4274.4276.0473.4778.7276.10
PCC84.4887.5081.4884.4983.7588.3778.3883.3883.3381.4084.9183.15
DualSCC82.9883.3382.6983.0178.7576.0982.3579.2279.1778.2680.0079.13
KCC84.0485.3783.0284.1980.0078.2682.3580.2180.2179.5580.7780.16
PCC93.6295.3592.1693.7591.2591.6790.6391.1591.9592.1191.8491.97

The bold values indicate the highest results of each indicator under each classification task and each data modality, and the proposed approach always performs the best.

Table 3

Performance comparison of different network construction methods based on different modal data

DataMethodNC versus ADEMCI versus LMCILMCI versus AD
ACC (%)SEN (%)SPE (%)BAC (%)ACC (%)SEN (%)SPE (%)BAC (%)ACC (%)SEN (%)SPE (%)BAC (%)
fMRISCC78.7276.0981.2578.6775.0072.9278.1375.5277.0873.4780.8577.16
KCC79.7977.7881.6379.7176.2575.0078.1276.5678.1374.4781.6378.05
PCC87.2377.2790.2883.3885.0085.7184.2184.9684.3885.3783.6484.50
SNPSCC75.5376.674.4775.5372.5071.1575.0073.0873.9670.0078.2674.13
KCC77.6674.4780.8577.6673.7571.4277.4274.4276.0473.4778.7276.10
PCC84.4887.5081.4884.4983.7588.3778.3883.3883.3381.4084.9183.15
DualSCC82.9883.3382.6983.0178.7576.0982.3579.2279.1778.2680.0079.13
KCC84.0485.3783.0284.1980.0078.2682.3580.2180.2179.5580.7780.16
PCC93.6295.3592.1693.7591.2591.6790.6391.1591.9592.1191.8491.97
DataMethodNC versus ADEMCI versus LMCILMCI versus AD
ACC (%)SEN (%)SPE (%)BAC (%)ACC (%)SEN (%)SPE (%)BAC (%)ACC (%)SEN (%)SPE (%)BAC (%)
fMRISCC78.7276.0981.2578.6775.0072.9278.1375.5277.0873.4780.8577.16
KCC79.7977.7881.6379.7176.2575.0078.1276.5678.1374.4781.6378.05
PCC87.2377.2790.2883.3885.0085.7184.2184.9684.3885.3783.6484.50
SNPSCC75.5376.674.4775.5372.5071.1575.0073.0873.9670.0078.2674.13
KCC77.6674.4780.8577.6673.7571.4277.4274.4276.0473.4778.7276.10
PCC84.4887.5081.4884.4983.7588.3778.3883.3883.3381.4084.9183.15
DualSCC82.9883.3382.6983.0178.7576.0982.3579.2279.1778.2680.0079.13
KCC84.0485.3783.0284.1980.0078.2682.3580.2180.2179.5580.7780.16
PCC93.6295.3592.1693.7591.2591.6790.6391.1591.9592.1191.8491.97

The bold values indicate the highest results of each indicator under each classification task and each data modality, and the proposed approach always performs the best.

Effect of parameters on classification performance

On the one hand, when constructing the BG-network, the correlation between nodes in the network was calculated by PCC. Then, all edges whose correlation coefficients are less than a threshold |$\alpha$| were deleted. The correlation coefficients of the remaining edges were defined as the weights of the edges, by which the weight matrix of the network was constructed. The value of |$\alpha$| directly affected the density of edges in BG-network. On the other hand, in the feature information aggregation process, the proportion of feature information in the aggregating neighborhood of a node was |$\beta$|⁠. The value of |$\beta$| directly affected the change degree of feature information.

In summary, the best combination of |$\alpha$| and |$\beta$| could make FIAD-GAN reach the best performance. Letting |$\alpha, \beta =[\textrm{0.1,0.2,0.3}\dots, 1]$|⁠, the classification performance of the method was examined by ACC, as shown in Figure 4. According to the experimental results, when |$\alpha$||$=0.3$| and |$=0.6$|⁠, FIAD-GAN achieved the best ACC in the three classification tasks.

Classification accuracy under different parameter combinations in different classification tasks. (A) AD versus NC. (B) EMCI versus LMCI. (C) LMCI versus AD.
Figure 4

Classification accuracy under different parameter combinations in different classification tasks. (A) AD versus NC. (B) EMCI versus LMCI. (C) LMCI versus AD.

Validity of BDC-C

To verify that the designed BDC-C operation was vital to obtaining better classification performance, a visualization experiment was conducted in this section. The left part of Figure 5 shows the specific experimental steps. First, for a specific sample, we define the |$i$|-th row in the weight matrix as the feature information sequence of the |$i$|-th feature (a brain region or a gene). Second, for a specific feature |$i$|⁠, we average the feature information sequences of the |$i$|-th feature, respectively, in the AD and NC groups to obtain the standard feature information sequences of the feature in the two groups. Finally, the PCC is calculated between each feature information sequence and its corresponding standard sequence. The above process was carried out in the weight matrix before and after the BDC-C operation, respectively. The right part of Figure 5 shows the experimental results of some features. For better visualization, only the top four features extracted from each classification task were shown.

The structural information similarity of the most discriminative features.
Figure 5

The structural information similarity of the most discriminative features.

In Figure 5, the fluctuation in the red lines was significantly weaker than that of the blue lines. Even for the samples having the same disease status, there were fluctuations in the similarity of the feature information between the same features. With the help of the proposed BDC-C operation, the information noise in features was suppressed. The similarity between samples in the same category tended to stabilize, which could improve the classification performance. The results uncovered the rationale for the improvement of classification ability brought by the proposed BDC-C operation. Compared with the ordinary convolution, the BDC-C operation takes the non-Euclidean structure of the graph into account, facilitating the efficient capture of the potential correlation in multimodal data. At the same time, the impact on performance caused by excessive model parameters is suppressed, improving the classification performance of the model.

Comparison with advanced methods

This section compared FIAD-GAN with some advanced methods [34, 39–42], covering several of the most common deep learning methods like CNN or GAN. The results are presented in Table 4, which also reported the sample numbers and data modality. The results showed that, among the compared methods, FIAD-GAN maintained satisfactory performance on all three classification tasks.

Table 4

Comparison with advanced methods

MethodModalitySamplesACC (%)
3D-CNNMRI + SNP + EHR267 (NC)/128 (AD)86.00
C-GNNfMRI960 (NC)/592 (AD)85.80
SVMMRI74 (EMCI)/38 (LMCI)64.30
FSN + PFCfMRI29 (EMCI)/18 (LMCI)80.85
C-GNNfMRI638 (LMCI)/592 (AD)65.20
TLAMRI70 (LMCI)/75 (AD)76.73
FIAD-GANfMRI + SNP237 (NC)/233 (AD)93.62
197 (EMCI)/203 (LMCI)91.25
203 (LMCI)/233 (AD)91.95
MethodModalitySamplesACC (%)
3D-CNNMRI + SNP + EHR267 (NC)/128 (AD)86.00
C-GNNfMRI960 (NC)/592 (AD)85.80
SVMMRI74 (EMCI)/38 (LMCI)64.30
FSN + PFCfMRI29 (EMCI)/18 (LMCI)80.85
C-GNNfMRI638 (LMCI)/592 (AD)65.20
TLAMRI70 (LMCI)/75 (AD)76.73
FIAD-GANfMRI + SNP237 (NC)/233 (AD)93.62
197 (EMCI)/203 (LMCI)91.25
203 (LMCI)/233 (AD)91.95
Table 4

Comparison with advanced methods

MethodModalitySamplesACC (%)
3D-CNNMRI + SNP + EHR267 (NC)/128 (AD)86.00
C-GNNfMRI960 (NC)/592 (AD)85.80
SVMMRI74 (EMCI)/38 (LMCI)64.30
FSN + PFCfMRI29 (EMCI)/18 (LMCI)80.85
C-GNNfMRI638 (LMCI)/592 (AD)65.20
TLAMRI70 (LMCI)/75 (AD)76.73
FIAD-GANfMRI + SNP237 (NC)/233 (AD)93.62
197 (EMCI)/203 (LMCI)91.25
203 (LMCI)/233 (AD)91.95
MethodModalitySamplesACC (%)
3D-CNNMRI + SNP + EHR267 (NC)/128 (AD)86.00
C-GNNfMRI960 (NC)/592 (AD)85.80
SVMMRI74 (EMCI)/38 (LMCI)64.30
FSN + PFCfMRI29 (EMCI)/18 (LMCI)80.85
C-GNNfMRI638 (LMCI)/592 (AD)65.20
TLAMRI70 (LMCI)/75 (AD)76.73
FIAD-GANfMRI + SNP237 (NC)/233 (AD)93.62
197 (EMCI)/203 (LMCI)91.25
203 (LMCI)/233 (AD)91.95

Biomedical discoveries and discussion

In terms of biological discovery, this work provides a reliable reference and technical basis for the clinical diagnosis, treatment and pathological analysis of diseases. This section displayed and analyzed these discoveries from different aspects.

First, based on the key structural sub-network, the importance scores of each node (brain region or gene) and edge (brain region-gene pair) were calculated. Then, they were sorted in descending order to find the optimal feature subset (optimal node feature subset and optimal edge feature subset) using the incremental search method. Figure 6 shows a visualization of 15 of the most distinctive brain region-gene pairs.

Visualization of the most discriminative brain region-gene pairs.
Figure 6

Visualization of the most discriminative brain region-gene pairs.

These brain region-gene pairs had unique guiding significance for further biomedical research. With the development of gene sequencing technology, it is rather difficult to screen out the gene-to-brain connections related to diseases from the entire human genome. The findings of this paper enabled researchers to focus on the regulation mode of certain specific genes to specific diseased brain regions, which may significantly improve the accuracy and speed of pathological research on complex brain diseases.

Second, Figure 7A shows the importance scores of the top 10 most discriminative brain regions. Figure 7B shows the locations of these brain regions, where the node sizes reflect the importance score. The extracted brain regions were consistent with other medical findings.

Location and importance scores of the most discriminative brain regions. (A) Importance of diseased brain regions. (B) Locations of diseased brain regions.
Figure 7

Location and importance scores of the most discriminative brain regions. (A) Importance of diseased brain regions. (B) Locations of diseased brain regions.

The inferior frontal gyrus had long been believed to be related to the development and degradation of lingual ability, yet recently, with the gradual deepening of the research on the function of brain regions, more and more scholars have found that the inferior frontal gyrus has a great relationship with the decline of memory ability and cognitive ability. Hay et al. [43] found a significant negative correlation between individual subjective cognitive ability and cerebral blood flow in the inferior frontal gyrus, proving that the inferior frontal gyrus may be an important biomarker of AD-like disease. As an important component, the Triangular part of the inferior frontal gyrus (IFGtriang) was shown to play a determinant role in memory generation [44]. Also, a significant increase in functional connectivity was detected in the orbital part of the right inferior frontal gyrus (ORBinf) in patients with cognitive impairment [45]. Kandilarova et al. [4] found that IFGtriang and ORBinf have significant regulatory effects on emotional abnormalities through 3 T-MRI scan data. Coincidentally, IFGtriang and ORBinf also have high importance scores in this study.

Gyrus rectus (REC), once regarded as the non-functional area of the brain, has been found evidently related to multiple cognitive functions like personality and thus may be linked with neurodegenerative diseases like AD. For example, Knutson et al. [46] found that the disinhibited behavior of patients with REC injury will be more significant, which is also one of the important criteria for the behavior of AD patients. Cajanus et al. [5] further observed a statistically significant association between the degree of atrophy in the REC region and the index of disinhibition in patients with neurodegenerative diseases. Shin et al. [47] studied MRI cortical thickness changes in the progression from MCI to dementia, finding that the right REC region was associated with the AD course.

Hippocampus (HIP) is the main control area of human memory and emotion regulation, whose atrophy has become a common and recognized biological characterization in AD research [48]. Recent studies have provided concrete evidence of its association with diseases such as AD from various perspectives. For example, the study by Tobin et al. [49] showed that the degree of neurogenesis in the hippocampal region is related to cognitive status. Specifically, the decrease in neurogenesis may indicate the onset of AD. Altuna et al. [37] found multiple DNA methylation sites related to AD in HIP, supporting the view that genes regulate AD development by affecting brain lesions.

Figure 8 shows all the genes extracted in this study and their importance scores, most of which were also supported by previous results. For example, the gene DAB1 regulates the synaptic neurotransmission in the adult brain and is involved in memory formation and maintenance [50]. The study by Blume et al. [13] showed that abnormal expression of DAB1 affects synapse generation in HIP, thereby affecting the generation of neuronal circuits with memory and learning functions, which may be a new way of the gene DAB1 affecting AD development. Kunkle et al. [51] conducted a genome-wide association study analysis of African-American AD and found some AD-related genes including LRP1B. CNTNAP2 has long been linked to neurodevelopmental disorders such as autism spectrum disorder. Based on gene knockout experiments, Varea et al. [52] studied the role of the gene CNTNAP2 in neuronal development and maturation, indicating the potential mechanism of this gene in neurodevelopmental diseases. As the results show, FIAD-GAN can precisely identify AD-related factors. For other atypical pathogenic factors not yet discussed, we plan to expand our dataset in the follow-up work and cooperate with clinicians to find more proof of the pathogenetic mechanism behind these factors. The identified brain regions and genes were distinctive between patients and normal people, which provides potential biomarkers for the clinical diagnosis and treatment of AD and can directly guide the development of transcranial magnetic stimulation therapy and targeted drugs.

Importance scores of risk genes. Genes with greater importance were considered as causative genes.
Figure 8

Importance scores of risk genes. Genes with greater importance were considered as causative genes.

For the node features (brain regions and genes) extracted by all comparative methods, the discriminative ability was assessed through the standard t-test. Figure 9 displays the P-values of the top 30 features extracted by the six methods. As shown, the P-values for most of the distinguishing node features extracted by FIAD-GAN were <0.05. It could be concluded that the features extracted by FIAD-GAN had a good discriminating ability, which not only illustrated that the proposed method has better classification capability, but also proved that FIAD-GAN had a promising prospect in the application of clinical decisions.

P-values of the top 30 features extracted by different methods.
Figure 9

P-values of the top 30 features extracted by different methods.

Finally, the functional connectivity developmental pattern of NC to AD was demonstrated by visualizing the brain functional connectivity network of the sample, as shown in Figure 10. The mean connectivity weights of brain functional connectivity in the NC group and AD group were calculated by Pearson correlation analysis. For better demonstration, the weights below a certain threshold were reset to 0. According to the exhibited results, as the disease develops, the number of functional connections significantly decreased and the strength of the connections diminished, indicating that some of the functional brain connections were disrupted during the disease progression of AD. In addition, most brain regions corresponding to the discriminative brain functional connectivity were consistent with those extracted by the method in this study, which from another aspect demonstrated the rationality and validity of FIAD-GAN. Also, the brain regions corresponding to the functional connections corrupted in AD patients were those who needed to be particularly concerned in the clinical treatment.

Brain functional connection networks. (A) NC. (B) AD.
Figure 10

Brain functional connection networks. (A) NC. (B) AD.

Rich experimental results indicate that the effectiveness of FIAD-GAN in feature extraction is not derived from the experimental coincidence but from the design of methodology. Based on the premise that the brain region level abnormalities reflected by brain imaging are regulated by genes, the effective extraction of multi-modal features by the proposed method benefits from two key points. The first is network-based fusion feature representation. By abstracting pathogenetic factors as nodes and the inter-modal relationship as the edges, BG-networks implement the information fusion of two modalities and provide a premise for inter-modal feature learning. The second is the proposed FIAD-GAN method, where the convolution and deconvolution operations in the generator are designed according to the feature information aggregation model we established. FIAD-GAN can explore the aggregation and diffusion pattern of the feature information between brain regions and genes in the disease development process, and realize the fusion of the multi-modal information.

In addition, FIAD-GAN is a general method for multimodal feature extraction. In addition to AD, our method has been applied for classifying Parkinson’s disease, autism spectrum disorder and other diseases, where different modalities of data in public datasets such as PPMI or ABIDE are applied. We also conducted experiments based on the clinical data collected by the Xiangya Hospital, which also confirmed the effectiveness and universality of our method. Considering the diversity of diseases, it is hard to enumerate all existing datasets from the experimental perspective. However, from the perspective of a data analysis algorithm, it is promising that our work can be extended to a wider range of data.

Conclusions

The development of complex brain diseases like AD has profound genetic causes. To fully utilize the complementary information between fMRI data and genetic data, this paper proposes a fusion data representation using graph structure and proposes an effective GAN-based method to realize automatic disease classification and feature extraction. Therefore, we established the feature information aggregation model and designed the specific convolution operation. We design experiments based on fMRI and SNP data in ADNI, and the superiority of FIAD-GAN in terms of disease classification and feature selection is verified by rich experiments. Furthermore, we extracted brain regions and genes that are highly related to the development of AD, which may provide a reference for future disease research and clinical diagnosis.

Key Points
  • This paper integrated multi-level medical data and represent it with a BG-network to explore AD from multiple perspectives, and it is one of the handful studies where network-based data representation is introduced.

  • We established a more intuitive and reasonable feature information aggregation model to describe the pathological interaction among brain regions and genes.

  • A novel deep learning framework named FIAD-GAN is proposed based on mathematical modeling, where the BDC-C and D-DC operations in the generator are specially designed for better key structural information capture and BG-network reconstruction. As the results indicate, the interpretability of the GAN framework has been improved, and performance was achieved in the classification and feature extraction.

  • The proposed FIAD-GAN extracted salient features, which provide a reliable reference and technical basis for the clinical diagnosis, treatment and pathological analysis of AD-like diseases.

Data availability

The source code is available at: https://github.com/fmri123456/FIAD-GAN.git. Multi-modal data used in this study is obtained from the Alzheimer’s Disease Neuroimaging Initiative (https://adni.loni.usc.edu/).

Funding

National Natural Science Foundation of China (62072173), Natural Science Foundation of Hunan Province, China (2020JJ4432), Key Scientific Research Projects of Department of Education of Hunan Province (20A296), Key open project of Key Laboratory of Data Science and Intelligence Education (Hainan Normal University), Ministry of Education (DSIE202101), National Key R&D Program of China (2020YFB2104400), Medical humanities and Social Sciences project of Hunan Normal University, and Innovation & Entrepreneurship Training Program of Hunan Xiangjiang Artificial Intelligence Academy.

Author Biographies

Xia-an Bi is currently a professor in the Hunan Provincial Key Laboratory of Intelligent Computing and Language Information Processing, and College of Information Science and Engineering in Hunan Normal University. His current research interests include machine learning, brain science and artificial intelligence.

Yuhua Mao, Sheng Luo and Hao Wu are currently pursuing the MS or ME degree with the Department of Computing, School of Information Science and Engineering, Hunan Normal University, Changsha, China. Their research interests and expertise are in machine learning, big data analysis and artificial intelligence.

Lixia Zhang is currently a professor at the School of Information Science and Engineering, Hunan Normal University. Her current research interests include machine learning, large-scale graph algorithm research and artificial intelligence.

Xun Luo is currently an associate professor in the College of Information Science and Engineering in Hunan Normal University. His current research interests include big data analysis, algorithm design and deep learning.

Luyun Xu is currently an associate professor in the College of Business in Hunan Normal University. Her current research interests include knowledge network, brain science and artificial intelligence.

References

1.

Garre-Olmo
J
.
Epidemiology of Alzheimer's disease and other dementias
.
Rev Neurol
2018
;
66
:
377
86
.

2.

Dugger
BN
,
Dickson
DW
.
Pathology of neurodegenerative diseases
.
Cold Spring Harb Perspect Biol
2017
;
9
:
a028035
.

3.

Du
L
,
Zhang
J
,
Liu
F
, et al.
Identifying associations among genomic, proteomic and imaging biomarkers via adaptive sparse multi-view canonical correlation analysis
.
Med Image Anal
2021
;
70
:
102003
.

4.

Kandilarova
S
,
Stoyanov
D
,
Sirakov
N
, et al.
Reduced grey matter volume in frontal and temporal areas in depression: contributions from voxel-based morphometry study
.
Acta Neuropsychiatrica
2019
;
31
:
252
7
.

5.

Cajanus
A
,
Solje
E
,
Koikkalainen
J
, et al.
The association between distinct frontal brain volumes and behavioral symptoms in mild cognitive impairment, Alzheimer's disease, and frontotemporal dementia
.
Front Neurol
2019
;
10
:
1059
.

6.

Pan
X
,
Zhang
C
,
Wang
J
, et al.
Epigenome signature as an immunophenotype indicator prompts durable clinical immunotherapy benefits in lung adenocarcinoma
.
Brief Bioinform
2022
;
23
:
bbab481
.

7.

Cao
X
,
Liu
J
,
Guo
M
, et al.
HiSSI: high-order SNP-SNP interactions detection based on efficient significant pattern and differential evolution
.
BMC Med Genomics
2019
;
12
:
1
12
.

8.

Luo
X
,
Liu
Y
,
Dang
D
, et al.
3D genome of macaque fetal brain reveals evolutionary innovations during primate corticogenesis
.
Cell
2021
;
184
:
723
740.e21
.

9.

Chu
Y
,
Wang
X
,
Dai
Q
, et al.
MDA-GCNFTG: identifying miRNA-disease associations based on graph convolutional networks via graph sampling through the feature and topology graph
.
Brief Bioinform
2021
;
22
:
bbab165
.

10.

Gao
Y
,
Li
X
,
Shang
S
, et al.
LincSNP 3.0: an updated database for linking functional variants to human long non-coding RNAs, circular RNAs and their regulatory elements
.
Nucleic Acids Res
2021
;
49
:
D1244
50
.

11.

Yue
M
,
Zhou
D
,
Zhi
H
, et al.
MSDD: a manually curated database of experimentally supported associations among miRNAs, SNPs and human diseases
.
Nucleic Acids Res
2018
;
46
:
D181
5
.

12.

Chen
J
,
Li
X
,
Calhoun
VD
, et al.
Sparse deep neural networks on imaging genetics for schizophrenia case–control classification
.
Hum Brain Mapp
2021
;
42
:
2556
68
.

13.

Blume
M
,
Inoguchi
F
,
Sugiyama
T
, et al.
Dab1 contributes differently to the morphogenesis of the hippocampal subdivisions
.
Dev Growth Differ
2017
;
59
:
657
73
.

14.

Lei
B
,
Zhao
Y
,
Huang
Z
, et al.
Adaptive sparse learning using multi-template for neurodegenerative disease diagnosis
.
Med Image Anal
2020
;
61
:
101632
.

15.

Du
L
,
Huang
H
,
Yan
J
, et al.
Structured sparse canonical correlation analysis for brain imaging genetics: an improved GraphNet method
.
Bioinformatics
2016
;
32
:
1544
51
.

16.

Hao
X
,
Li
C
,
Du
L
, et al.
Mining outcome-relevant brain imaging genetic associations via three-way sparse canonical correlation analysis in Alzheimer’s disease
.
Sci Rep
2017
;
7
:
1
12
.

17.

Farahani
FV
,
Karwowski
W
,
Lighthall
NR
.
Application of graph theory for identifying connectivity patterns in human brain networks: a systematic review
.
Front Neurosci
2019
;
13
:
585
.

18.

Chen
X
,
Yin
J
,
Qu
J
, et al.
MDHGI: matrix decomposition and heterogeneous graph inference for miRNA-disease association prediction
.
PLoS Comput Biol
2018
;
14
:
e1006418
.

19.

Yi
H-C
,
You
Z-H
,
Huang
D-S
, et al.
Graph representation learning in bioinformatics: trends, methods and applications
.
Brief Bioinform
2022
;
23
:
bbab340
.

20.

Yang
J
,
Lei
X
.
Predicting circRNA-disease associations based on autoencoder and graph embedding
.
Inform Sci
2021
;
571
:
323
36
.

21.

Song
X
,
Zhou
F
,
Frangi
AF
, et al.
Graph convolution network with similarity awareness and adaptive calibration for disease-induced deterioration prediction
.
Med Image Anal
2021
;
69
:
101947
.

22.

Du
L
,
Liu
K
,
Zhu
L
, et al.
Identifying progressive imaging genetic patterns via multi-task sparse canonical correlation analysis: a longitudinal study of the ADNI cohort
.
Bioinformatics
2019
;
35
:
i474
83
.

23.

Chen
X
,
Zhao
Q
.
Developing novel computational techniques for medicine and pharmacy
.
Curr Top Med Chem
2018
;
18
:
947
8
.

24.

Gong
Y
,
Liao
B
,
Peng
D
, et al.
Accurate prediction and key feature recognition of immunoglobulin
.
Appl Sci
2021
;
11
:
6894
.

25.

Qiang
X
,
Zhou
C
,
Ye
X
, et al.
CPPred-FL: a sequence-based predictor for large-scale identification of cell-penetrating peptides by feature representation learning
.
Brief Bioinform
2020
;
21
:
11
23
.

26.

Rao
B
,
Zhou
C
,
Zhang
G
, et al.
ACPred-fuse: fusing multi-view information improves the prediction of anticancer peptides
.
Brief Bioinform
2020
;
21
:
1846
55
.

27.

Peng
L
,
Peng
M
,
Liao
B
, et al.
The advances and challenges of deep learning application in biological big data processing
.
Curr Bioinform
2018
;
13
:
352
9
.

28.

Chen
X
,
Li
T-H
,
Zhao
Y
, et al.
Deep-belief network for predicting potential miRNA-disease associations
.
Brief Bioinform
2021
;
22
:
bbaa186
.

29.

Feng
C
,
Elazab
A
,
Yang
P
, et al.
Deep learning framework for Alzheimer’s disease diagnosis via 3D-CNN and FSBi-LSTM
.
IEEE Access
2019
;
7
:
63605
18
.

30.

Illan
IA
,
Górriz
JM
,
Ramírez
J
, et al.
Spatial component analysis of MRI data for Alzheimer's disease diagnosis: a Bayesian network approach
.
Front Comput Neurosci
2014
;
8
:
156
.

31.

AbdulAzeem
Y
,
Bahgat
WM
,
Badawy
M
.
A CNN based framework for classification of Alzheimer’s disease
.
Neural Comput Applic
2021
;
33
:
10415
28
.

32.

Purcell
S
,
Neale
B
,
Todd-Brown
K
, et al.
PLINK: a tool set for whole-genome association and population-based linkage analyses
.
Am J Hum Genet
2007
;
81
:
559
75
.

33.

Karpiel
I
,
Klose
U
,
Drzazga
Z
.
Optimization of rs-fMRI parameters in the seed correlation analysis (SCA) in DPARSF toolbox: a preliminary study
.
J Neurosci Res
2019
;
97
:
433
43
.

34.

Wee
C-Y
,
Liu
C
,
Lee
A
, et al.
Cortical graph neural network for AD and MCI diagnosis and transfer learning across populations
.
Neuroimage
2019
;
23
:
101929
.

35.

Du
L
,
Liu
K
,
Yao
X
, et al.
Detecting genetic associations with brain imaging phenotypes in Alzheimer’s disease via a novel structured SCCA approach
.
Med Image Anal
2020
;
61
:
101656
.

36.

Du
L
,
Liu
F
,
Liu
K
, et al.
Associating multi-modal brain imaging phenotypes and genetic risk factors via a dirty multi-task learning method
.
IEEE Trans Med Imaging
2020
;
39
:
3416
28
.

37.

Altuna
M
,
Urdánoz-Casado
A
,
Sánchez-Ruiz de Gordoa
J
, et al.
DNA methylation signature of human hippocampus in Alzheimer’s disease is linked to neurogenesis
.
Clin Epigenet
2019
;
11
:
1
16
.

38.

Aggarwal
A
,
Mittal
M
,
Battineni
G
.
Generative adversarial network: an overview of theory and applications
.
Int J Inf Manage Data Insights
2021
;
1
:
100004
.

39.

Venugopalan
J
,
Tong
L
,
Hassanzadeh
HR
, et al.
Multimodal deep learning models for early detection of Alzheimer’s disease stage
.
Sci Rep
2021
;
11
:
1
13
.

40.

Mehmood
A
,
Yang
S
,
Feng
Z
, et al.
A transfer learning approach for early diagnosis of Alzheimer’s disease on MRI images
.
Neuroscience
2021
;
460
:
43
52
.

41.

Prasad
G
,
Joshi
SH
,
Nir
TM
, et al.
Brain connectivity and novel network measures for Alzheimer's disease classification
.
Neurobiol Aging
2015
;
36
:
S121
31
.

42.

Yang
P
,
Zhou
F
,
Ni
D
, et al.
Fused sparse network learning for longitudinal analysis of mild cognitive impairment
.
IEEE Trans Cybernet
2019
;
51
:
233
46
.

43.

Hays
CC
,
Zlatar
ZZ
,
Campbell
L
, et al.
Subjective cognitive decline modifies the relationship between cerebral blood flow and memory function in cognitively normal older adults
.
J Int Neuropsychol Soc
2018
;
24
:
213
23
.

44.

Varela-López
B
,
Cruz-Gómez
ÁJ
,
Lojo-Seoane
C
, et al.
Cognitive reserve, neurocognitive performance, and high-order resting-state networks in cognitively unimpaired aging
.
Neurobiol Aging
2022
;
117
:
151
64
.

45.

Wang
S
,
Sun
H
,
Hu
G
, et al.
Altered insular subregional connectivity associated with cognitions for distinguishing the spectrum of pre-clinical Alzheimer's disease
.
Front Aging Neurosci
2021
;
13
:
597455
.

46.

Knutson
KM
,
Dal Monte
O
,
Schintu
S
, et al.
Areas of brain damage underlying increased reports of behavioral disinhibition
.
J Neuropsychiatry Clin Neurosci
2015
;
27
:
193
8
.

47.

Shin
N-Y
,
Bang
M
,
Yoo
S-W
, et al.
Cortical thickness from MRI to predict conversion from mild cognitive impairment to dementia in Parkinson disease: a machine learning–based model
.
Radiology
2021
;
300
:
390
9
.

48.

Aschenbrenner
AJ
,
Gordon
BA
,
Benzinger
TL
, et al.
Influence of tau PET, amyloid PET, and hippocampal volume on cognition in Alzheimer disease
.
Neurology
2018
;
91
:
e859
66
.

49.

Tobin
MK
,
Musaraca
K
,
Disouky
A
, et al.
Human hippocampal neurogenesis persists in aged adults and Alzheimer’s disease patients
.
Cell Stem Cell
2019
;
24
:
974
982.e3
.

50.

Kim
H-R
,
Lee
T
,
Choi
JK
, et al.
Polymorphism in the MAGI2 gene modifies the effect of amyloid β on neurodegeneration
.
Alzheimer Dis Assoc Disord
2021
;
35
:
114
20
.

51.

Kunkle
BW
,
Jean-Francois
MN
,
Hamilton-Nelson
KL
, et al.
APOE-stratified genome-wide association analysis identifies novel Alzheimer disease candidate risk loci for African Americans
.
Alzheimers Dement
2021
;
17
:
e056383
.

52.

Varea
O
,
Martin-de-Saavedra
MD
,
Kopeikina
KJ
, et al.
Synaptic abnormalities and cytoplasmic glutamate receptor aggregates in contactin associated protein-like 2/Caspr2 knockout neurons
.
Proc Natl Acad Sci U S A
2015
;
112
:
6176
81
.

This article is published and distributed under the terms of the Oxford University Press, Standard Journals Publication Model (https://academic.oup.com/journals/pages/open_access/funder_policies/chorus/standard_publication_model)