
Brain community detection is an efficient method to represent the communities of brain networks. However, time-variable functions of the brain and the intricate brain community structure impose a great challenge on it. In this paper, a time-sequential graph adversarial learning (TGAL) framework is proposed to detect brain communities and characterize the structure of communities from brain networks. In the framework, a novel time-sequential graph neural network is designed as an encoder to extract efficient graph representations by spatio-temporal attention mechanism. Since it is difficult to capture the community structure, the measurable modularity loss is used to optimize by maximizing the modularity of the community. In addition, the framework employs an adversarial scheme to guide the learning of representation. The effectiveness of our model is shown through experiments on the real-world brain network datasets, and the great performance of brain community detection demonstrates the advantage of the proposed framework.
Citation: Changwei Gong, Bing Xue, Changhong Jing, Chun-Hui He, Guo-Cheng Wu, Baiying Lei, Shuqiang Wang. Time-sequential graph adversarial learning for brain modularity community detection[J]. Mathematical Biosciences and Engineering, 2022, 19(12): 13276-13293. doi: 10.3934/mbe.2022621
[1] | Xiaowen Jia, Jingxia Chen, Kexin Liu, Qian Wang, Jialing He . Multimodal depression detection based on an attention graph convolution and transformer. Mathematical Biosciences and Engineering, 2025, 22(3): 652-676. doi: 10.3934/mbe.2025024 |
[2] | Zhongwei Li, Wenqi Jiang, Xiaosheng Liu, Kai Tan, Xianji Jin, Ming Yang . GAN model using field fuzz mutation for in-vehicle CAN bus intrusion detection. Mathematical Biosciences and Engineering, 2022, 19(7): 6996-7018. doi: 10.3934/mbe.2022330 |
[3] | Jing Zhang, Ting Fan, Ding Lang, Yuguang Xu, Hong-an Li, Xuewen Li . Intelligent crowd sensing pickpocketing group identification using remote sensing data for secure smart cities. Mathematical Biosciences and Engineering, 2023, 20(8): 13777-13797. doi: 10.3934/mbe.2023613 |
[4] | Xing Hu, Minghui Yao, Dawei Zhang . Road crack segmentation using an attention residual U-Net with generative adversarial learning. Mathematical Biosciences and Engineering, 2021, 18(6): 9669-9684. doi: 10.3934/mbe.2021473 |
[5] | Bingyu Liu, Jiani Hu, Weihong Deng . Attention distraction with gradient sharpening for multi-task adversarial attack. Mathematical Biosciences and Engineering, 2023, 20(8): 13562-13580. doi: 10.3934/mbe.2023605 |
[6] | Jia Yu, Huiling Peng, Guoqiang Wang, Nianfeng Shi . A topical VAEGAN-IHMM approach for automatic story segmentation. Mathematical Biosciences and Engineering, 2024, 21(7): 6608-6630. doi: 10.3934/mbe.2024289 |
[7] | Qi Cui, Ruohan Meng, Zhili Zhou, Xingming Sun, Kaiwen Zhu . An anti-forensic scheme on computer graphic images and natural images using generative adversarial networks. Mathematical Biosciences and Engineering, 2019, 16(5): 4923-4935. doi: 10.3934/mbe.2019248 |
[8] | Paula Mercurio, Di Liu . Identifying transition states of chemical kinetic systems using network embedding techniques. Mathematical Biosciences and Engineering, 2021, 18(1): 868-887. doi: 10.3934/mbe.2021046 |
[9] | Shi Liu, Kaiyang Li, Yaoying Wang, Tianyou Zhu, Jiwei Li, Zhenyu Chen . Knowledge graph embedding by fusing multimodal content via cross-modal learning. Mathematical Biosciences and Engineering, 2023, 20(8): 14180-14200. doi: 10.3934/mbe.2023634 |
[10] | Sadia Anjum, Lal Hussain, Mushtaq Ali, Adeel Ahmed Abbasi, Tim Q. Duong . Automated multi-class brain tumor types detection by extracting RICA based features and employing machine learning techniques. Mathematical Biosciences and Engineering, 2021, 18(3): 2882-2908. doi: 10.3934/mbe.2021146 |
Brain community detection is an efficient method to represent the communities of brain networks. However, time-variable functions of the brain and the intricate brain community structure impose a great challenge on it. In this paper, a time-sequential graph adversarial learning (TGAL) framework is proposed to detect brain communities and characterize the structure of communities from brain networks. In the framework, a novel time-sequential graph neural network is designed as an encoder to extract efficient graph representations by spatio-temporal attention mechanism. Since it is difficult to capture the community structure, the measurable modularity loss is used to optimize by maximizing the modularity of the community. In addition, the framework employs an adversarial scheme to guide the learning of representation. The effectiveness of our model is shown through experiments on the real-world brain network datasets, and the great performance of brain community detection demonstrates the advantage of the proposed framework.
Neurological diseases are one of the most common diseases, causing a large number of patients and casualties worldwide every year and imposing a huge burden on the healthcare industry. Meanwhile, pharmacological magnetic resonance imaging (phMRI) is a technology that allows researchers to obtain noninvasive functional brain imaging of drug-induced variations in blood flow dynamics. As a derivative of functional magnetic resonance imaging (fMRI), phMRI will significantly promote the progression of neurological diseases in terms of pharmacokinetic and pharmacodynamic properties if the time course and neurological response of specific pharmacological stimulus can be analyzed for specific diseases and medications. As a result, phMRI processing analysis has a wide range of applications and a significant impact in the medical field. For example, the analysis of phMRI of addictive drugs can be used to study the addictive mechanisms of drugs in order to develop therapies for treating addiction [1]. And the work in this paper proposes the relevant technique for medical research in this field and engages in the processing and analysis of MRI data.
With the increasing number of studies contributing to the exploration and advancement of neuroscience, researchers are extracting and processing brain information more elaborately and profoundly [2], characterized by a large amount of complicated biological data obtained from large-scale brain neural systems. The vast amount of this enormous data is overwhelmingly in the form of network datay [3], representing the connections or interconnections of components in various large-scale neurobiological systems. In neuroscience, these data often span a variety of scales [4] (neurons, circuits, systems, and brains) or include different types of data (e.g., structural networks that express the anatomical connections of nerves, functional networks that represent the connections of distributed brain regions associated with neural activity). Brain networks [5] are composed of anatomical structures that segment different brain regions and connect them through functional networks that reveal complex patterns of neuronal communication. Such complex neural signal patterns, attributed to current advances in imaging technology and advanced medical image processing methods [6,7], can be studied with fMRI [8] and mentioned phMRI, yielding neural response activity thought to be associated with various behavioral and cognitive functions as well as brain diseases [9].
Complex networks are characterized by their modular structure [10]. This implies that the nodes of a network may be divided into modules or communities that are inwardly dense and outwardly sparse. The brain follows the organizing principle of modular organization [11], and the most widely encountered and biologically meaningful aspect of brain networks is their organization into distinct network communities or modules, meaning that tightly connected clusters within them are less connected to each other. Modular blocks have neuroscientific significance due to the fact that their boundaries distinguish functionally relevant neuron elements [12], define hubss [13] and crucial bridges that connect communities, channel and impede the flow of neural impulses and messages, and constrain the unrestricted spread of instabilities.
Due to brain networks being large and their connectivity patterns being complex, it is often impossible to identify modules simply by visually inspecting the network. Community identification methods have been used to significant effect in solving such problems [14]. Subgraphs, network modules, and communities have all been explored in depth in network topologies. They have been widely used in network neuroscience [15]. Network structure identification, also known as community detection, is the division of nodes in a network into groups, where nodes within communities are densely connected and nodes between communities are sparsely connected. Mining network structure reveals and understands complex network systems' organizational principles and operational roles. In addition, the development of new methods for mapping the structural and functional connectivity of the brain has led to the preparation of complete network maps of neuronal circuits and systems [16,17]. The structure of these brain networks can be examined and analyzed using various graph-theoretic methods. Thus, methods for finding modules or network clusters in brain networks have specialized applicability, revealing tightly coupled core building blocks or substructures, often corresponding to specific functional components [18]. With the recent intersection of neurobiological imaging and network science graph theory, new applications and analytical methods are being developed to analyze real-world biological networks.
Since the past few decades, data-driven models have received a great deal of attention in a broad range of fields, for example, it is used for image [19], disease prediction [20,21] and trajectory detection [22]. When cooperated with machine learning approaches, these models has led to a great deal of success in the area of medical image computing [23] when it comes to constructing pattern recognition frameworks. The models get the capacity to fulfill a high level of precision at a minimal computational cost. Building on its foundation, deep learning has become a method presently regarded as one of the most critical advances in machine learning overall. Deep learning is a novel technique for processing high-dimensional biological network data and learning low-dimensional graph representations of brain network structures [24,25,26]. Exemplary approaches include those based on generative adversarial techniques [27,28,29] and graph neural networks [30]. Generative adversarial networks (GANs) [31], which can bee seen as variational-inference [32] based generative model, are frequently unsupervised in training, and the freshly produced data (in theory) has the same distribution as actual data, allowing for robust, complicated data analysis. And it is important to note that the prevalent technique in medical image analysis is the use of GAN [35,36]. By pooling and convolution, the convolutional neural network (CNN) [33] method decreases the dimensionality of medical imaging data, allowing it to discover patterns in biomedical research effectively. Graph convolutional network (GCN) [34] is built to extract community characteristics as it directly analyzes network-structured data and succeeds CNN capabilities. However, existing approaches for processing dynamic network data to produce temporal graph representations for community discovery remain hard [37], particularly for small network sample datasets [38].
To resolve mentioned problems, we developed a novel time-sequential graph adversarial learning (TGAL) to accomplish the partitioning of brain regions in time-varying brain networks, detect different communities involving similar brain regions in brain networks and their evolution over time steps, and improve the robustness of the framework when dealing with small sample network data utilizing generative adversarial techniques. The main contributions of our work can be concluded as follows: 1) A time-sequential graph attention encoder is designed to learn more efficient graph representation, and more helpful graph embeddings are obtained to complete the clustering to detect more accurate dynamic communities. 2) An adversarial training scheme is proposed to assist the representation learning of brain networks. It alleviates some of the issues created by training with few data samples. 3) Due to the excellent classification performance of our frameworks with neuroscientific supports, the communities discovered in biological and medical imaging experiments may be employed as biomarkers.
Following is the outline for this paper: In Section 2, a graph learning framework with adversarial strategy is established, and its formulas and details of the composition of each module is given. In Section 3, a neurobiology networks dataset is constructed and described as following, a dynamic community detection experiment is designed to estimate the capability of node clustering in brain regions, Then, a visualization of the results is displayed. And a graph representation learning experiment is implemented to verify the representational learning capabilities of the framework. In Section 4, the discussion of the pros and cons of our framework and the future directions for improvement is presented. Finally, we summarize the work of this paper in Section 5.
The architecture of our TGAL is illustrated in Figure 1. Firstly, the raw brain functional image dataset is input to the module called "brain network construction" to construct network data with nodes and connections by time series and brain atlas. Secondly, on one hand, the network data is fed to an encoder, and on the other hand, the network data is given as positive samples to a discriminator in adversarial learning to generate augmented data. Then, the encoder provides the output graph embedding for not only decoding to reconstruct but calculating the loss function for community detection. The forward computation is then completed and the detected brain communities are outputted, backpropagation is finally carried out to update the model parameters.
Brain networks construction will be describe in the datasets preparation below. And the explanation of the notations is given in Table 1. In the auto-encoder, the encoder (E) adopts temporal graph attention networks to transform the time series of brain regions ({Xt}Tt=1) and brain functional connections (A) into the embeddings ({Zt}Tt=1). Additionally, in the adversarial regularizer, the encoder, which is referred to as the generator (G), and the discriminator (D), engage in a min-max adversarial game in order to learn more suitable embeddings. The community assignment matrix is optimized using the measurable soft modularity loss in order to detect communities. Consequently, the encoder is trained with two objectives: a typical auto-encoder reconstruction loss and a measured modularity loss for community detection.
notation | description |
E | encoder |
G | generator |
D | discriminator |
X | attributes of brain region(node) |
A | adjacency(connectivity) matrix of brain networks |
H | hidden representations of region(node) attributes |
Z | embeddings of region(node) attributes |
ϕ | parameters of discriminator |
ψ | parameters of generator |
E | expectation |
i,j | No.i and No.j brain region |
The purpose of the temporal graph autoencoder is to encode the properties of the dynamic brain network in a low-dimensional latent space. First, we create the encoder using two distinct network blocks: the topological attention block and the temporal attention block. Each block is composed of multiple layers of the respective stack. Both utilize self-attention techniques to achieve an effective time-sequential graph representation from its neighbors and historical context data.
The initial input for this layer is a set of brain network attributes {xi∈Rd,∀i∈V} where d is the dimension of time series. The output is a set of brain region representations {hi∈Rf,∀i∈V} where f is the dimension of captured topological properties.
Similar to graph attention networks (GAT) [39], our topological attention layer is concerned with the near neighbors of the brain region i by calculating attention weight from input brain region representations:
hi=σ(∑u∈NiαijWxj), | (2.1) |
αij=exp(σ(Aij⋅[Wxj‖Wxi]))∑w∈Niexp(σ(Awi⋅[Wxw‖Wxi])) | (2.2) |
Here Ni={j∈V:(i,j)∈E} is the set of near neighbor of region i which are linked by functional connection A; W∈Rd×f is a weight transformation matrix for each region representations; σ(⋅) is sigmoid activation function and ‖ is the concatenation operation. The learnt coefficients αij, which is computed by performing softmax on each neighbors, indicates the significance of brain region i to region j. Note that topological attention layer applies on brain region representation at a single time step, and multiple topological attention layer can calculate the entire time sequence in parallel.
Capturing the constantly changing patterns of brain networks in a global way is critical for dynamic community detection. When obtaining attributes at current time step, it is essential to consider the global temporal context. The essential concern is how to record the temporal variations in the organization of brain networks over several time steps. Temporal attention layers are intended to address this problem using the scaled dot-product attention [40]. Its queries, keys, and values are utilized to represent the properties of brain regions that provide input.
We define Hs={h1s,h2s,...,hTs}, a representation sequence of a brain region s at continuous time steps as input, where T is the number of time steps. And the output of the layer is Zs={z1s,z2s,...,zTs}, a new brain network representation sequence for region s at different time steps.
Using hts as the query, temporal attention layer evaluate its historical representations, inquiring the temporal context of the neighborhood around region s. Hence, temporal self-attention allows the discovery of relationships between time-varying representations of a brain region across several time steps. Formally, the temporal attention layer is computed as:
Zs=βs(HsWs), | (2.3) |
βijs=exp(eijs)∑Tk=1exp(eiks),eijs=(((HsWq)(HsWk)T)ij√F′) | (2.4) |
where βs∈RT×T is the attention coefficient matrix computed by the query-key dot product attention operation; Wq∈Rd×f, Wk∈Rd×f and Wv∈Rd×f are linear projections matrices which transform representations into a particular space.
The two attention blocks are calculated in sequence to obtain the final temporal representation, i, e., the output embeddings Z. And It is utilized to reconstruct the brain network topology in the decoder:
ˆA=σ(ZZT) | (2.5) |
ˆA is the reconstructed brain functional connection and σ(⋅) is still sigmoid function.
The classic reconstruction loss is defined by the form of cross entropy:
LRE=∑E[AijlogˆAij+(1−Aij)log(1−ˆAij)] | (2.6) |
In this adversarial model, the main objective is enforcing brain network embeddings Z to match the prior distribution. Other naive regularizers push the learned embeddings to conform to the Gaussian distribution rather than capture semantic diversity [41]. As a result, conventional techniques to network embedding cannot effectively profit from adversarial learning. Generative data augmentation is needed to explore the underlying features of the data to offset the negative impact of small sample data. Therefore, we derive the previous distribution of communities by counting different kinds of modules in the functional brain network that have been confirmed by neuroscience. The adversarial model serves as a discriminator by using a three-layer fully connected network to identify whether a latent code generated from the prior distribution pz′ or brain network data from the real-world dataset (X,A). The regularizer will eventually enhance the embedding during the minimax competition between the encoder and the discriminator in the training phase by generating augmented data X′ as input to the auto encoder.
The loss of the encoder(generator) LG and discriminator LD in the adversarial model, defined as follows:
LG=−Ex∼pdata logDϕ(Gψ(X,A)) | (2.7) |
LD=−Ez′∼pz′logDϕ(X′)−Ex∼pdata log(1−Dϕ(Gψ(X,A))) | (2.8) |
In this expression, z′ is a latent code sampled from the prior distribution pz′ of empirically confirmed brain communities; Dϕ(⋅) and Gψ(⋅) is the above-mentioned discriminator and encoder.
Formally, the objective of this adversarial learning model can be indicated as a minmax criterion:
LAL=minGmaxDΘ(D,G)=minGmaxD(Ez′∼pz′logD(X′;ϕ)+Ex∼pdata log(1−D(G(X,A);ϕ))) | (2.9) |
Modularity maximization is a technique for community discovery that is commonly used in the detection of brain modules. A partition is regarded high quality (and so has a higher Q score [24]) conceptually if the communities it forms are more dense internally than would be predicted by chance. Thus, the partition that gets the maximum value of Q is considered to be a good estimation of the community structure of a brain network. This intuition may be expressed as follows:
Q=12m∑ij[Aij−cij]δ(ωi,ωj) | (2.10) |
Here aij indicates the number of functional connection between region i and j; cij=kikj2m denotes the estimated number of connections based on a null model where ki=∑jAij is a degree of the region i and 2m=∑ijAij is overall amount of connections in the brain networks; δ(ωi,ωj)=1, if ωi=ωj, which means reigon i and reigon j are in the same community and 0 otherwise.
Inspired by [42], to develop a differentiable objective for optimizing the community assignment matrix P=softmax(Z)∈RN×C which represents a matrix of probabilities of brain region attribution to communities, the measurable modularity loss employed by our framework is defined as:
LMM=−1∑|Aij|tr(P⊤BP)⏟measurable modularity +λ(C∑i(N∑jPij−1C)2)⏟regularization | (2.11) |
where the modularity matrix B=A−ddT2m; C is the amount of communities and N is the number of regions in the brain networks. The regularization ensures that the model can identify communities of the predicted size.
Thus, the total loss for the encoder optimization in the train process to obtain better embeddings is sum of the above three loss terms, expressed as follows:
Ltotal=LRE+LG+LMM | (2.12) |
In this part, we assess the performance of TGAL in terms of both dynamic community detection and graph representation learning. We use metrics of graph theory and machine learning to measure the experimental results over dynamic community detection, evaluating the performance of our model in graph representation learning with binary classification metrics.
The neurobiological experiment dataset contains two types of data with different numbers: raw functional MRI images of 8 nicotine non-drug injected rats and 16 drug-injected rats, each with 800 time series. The injection was carried out in the middle moment of the fMRI processing on rats, and its injection time was ignored. The injected drug will stimulate the nervous system of the rat brain and cause some disturbance to the brain network.
● Brain networks construction: By preprocessing long-term functional MRI scans of experimental rats, we produced the dynamic brain network dataset necessary for the experiment. The first preprocessing was performed using the Statistical Parametric Mapping 8 (SPM8) tool in MATLAB. In order to account for head motion, functional signals were aligned and unwrapped, and the average motion-corrected picture was coregistered with the high-resolution anatomical T2 image. The functional data were then smoothed using an isotropic 3mm full-width at half-maximum (FWHM) Gaussian kernel. Upon the base of the Wister rat brain atlas, 150 functional brain regions have been characterized. We used magnitude-squared coherence to evaluate the spectrum correlation between regional time series, resulting in a 150×150 functional connection matrix for each time step, the elements of which displayed the strength of functional connectivity between all interactions of regions. Therefore, in the time-varying dynamic brain network, the BOLD sequence signal from fMRI is used as a nodal attribute for each brain region, the adjacency matrix is given by the preprocessed functional connectivity, and the entire time period is separated equally into 6 time steps.
Python Torch was used as the backend for our TGAL. Two NVIDIA GeForce RTX 3080 Ti sped up the training of the networks. During training, the training epoch was set at 1000 and the learning rate was set to 0.001. Adam was used as an optimizer with a weight decay of 0.01 to reduce overfitting. The encoder was trained using 2 topological and two temporal attention layers. We perform each experiment 10 times and calculate the mean result. We set the regularization value to 0.5 and the number of communities to 10 for all datasets and methods.
Our approach was compared against the following two kinds of baselines:
● GAE [44]: is recently the most common autoencoder-based unsupervised framework for graph data, in which the encoder is composed of two-layer graph convolutional networks to leverage topological information.
● ARGA [45]: is an adversarially regularized autoencoder method that employs graph autoencoder to learn the representations, regularizes the latent codes, and forces the latent codes to match a prior distribution; differing from ours, it used simple Gaussian distribution as the prior distribution.
For graph-level metrics, we report average community conductance C and modularity Q. For ground-truth label correlation analysis, we report normalized mutual information (NMI) between the community assignments and labels and pairwise F-1 score between all node pairs and their associated community pairs.
As indicated in Table 2, we conducted ablation research on community detection to evaluate the effectiveness of our proposed TGAL framework in the three terms of encoder composition, adversarial learning module and loss function. And three significant outcomes were achieved: 1) In a comparison of graph-level metrics, the k-means-based technique performed admirably for community conductance, while the modularity loss-based method performed better for community modularity. This is owing to the fact that the two methods have fundamentally different optimization goals; modularity loss is motivated by the desire to maximize modularity. 2) The strategy with adversarial regularizer performs well on average; this demonstrates that adversarial learning serves as an adjunct to graph representation learning. 3) Our technique performs poorly when the proposed encoder is replaced with a two-layer graph convolutional encoder; this suggests that the proposed encoder may learn better embeddings to improve performance by virtue of a powerful attention mechanism [43].
encoder | AL | loss | Metrics(%) | ||||||
GCN | TG | K-means | modularity | C | Q | NMI | F1 | ||
√ | × | × | √ | × | 68.4 | 16.7 | 55.5 | 28.5 | |
√ | × | √ | √ | × | 72.5 | 24.3 | 49.6 | 44.4 | |
√ | × | × | × | √ | 28.4 | 64.5 | 50.8 | 60.2 | |
√ | × | √ | × | √ | 22.8 | 58.8 | 47.5 | 59.8 | |
× | √ | √ | × | √ | 42.9 | 69.5 | 67.5 | 66.7 |
Figure 2 depicts the results of our dynamic community experiment. It illustrates the changes in the spatial distribution of the three major brain communities observed by our method over increasing time intervals. The algorithm for community detection modularizes the clustering of nodes represented by each brain region, and the top three communities with the largest number of nodes can be identified as evolving in each time step. And each time step corresponds to a short period in the description of the dataset as mentioned above. As can be observed from the figure, there was no significant change in the distribution of brain communities at time steps 1–3, and the spatial location of the brain communities did not change much in general during this period. But there was a significant change from time step 4–6. The three main brain communities have altered spatial distribution to some extent in each of the periods after time step 3. Because the experiment was administered drug injection after time step 3 caused the rats in the original data set to change the characteristics and topology of their brain networks. Thus, the experimental results are consistent with neuroscientific facts.
We divided the data from drug-injected and non-drug-injected rats into two groups and verified whether the models learned efficient graph representations by competing with state-of-the-art graph representation learning models for binary classification performance.
● DGI [46]: highlights the importance of cluster and representation learning in combination. We learn unsupervised graph representation with DGI and two algorithms both run SVM on the final representations as the classifer.
The performance of binary classification is evaluated using six quantitative representation learning metrics: To summarize: 1) Accuracy (ACC), 2) Area under receiver operating characteristic curve (AUC), 3) Precision (PRE), 4) Recall (REC), 5) Receiver operating characteristic (ROC), and 6) F1-score. Our suggested technique is being evaluated using leave-oneout cross-validation (LOOCV), since we only have a small quantity of data. One of the N individuals is omitted from the testing process, and the N−1 subjects that remain are used for training purposes only. Each technique's hyperparameters are adjusted to their optimal values via the greedy search algorithm.
As shown in Figure 3, our method achieves better results in terms of classification performance overall. In a way, it verifies that the representation obtained by our method is stronger in the unsupervised learning process. Figures 4 and 5 display the comparison of our model's second classification accuracy and precision compared to the DGI method at different time steps. Again, the overall superiority of our method can be seen, and it can be observed that the classification metrics at time steps 4–6 are generally better than those at time steps 1–3, which may be due to the fact that it is easier to classify when the alterations in the dynamic brain network start to appear after drug injection.
Considering all the factors that affect the performance of our framework in the experiment, one of the essential points is the presence of noise in the data, which is a common problem in current biomedical techniques when carrying out experiments.
In the data processing part of this experiment, we thoroughly considered the effect of noisy data on the experimental results. The acquired fMRI data are screened, and the collected unqualified data will be excluded from the subsequent experiments. Since the scanned subjects have a slight head movement during the experimental data acquisition, it will cause the acquired images to be noisy. Head movement correction is used to make the processed images overlap with the target images as much as possible so that the same voxel corresponds to the same position in the brain at each moment. In this way, the interference of noise in the experiment is reduced. Several pre-processing processes, such as normalization, aim to minimize errors due to data acquisition and physiological properties. However, it is still not possible to completely exclude the effect of noise in the data.
Therefore, our deep learning framework also incorporates a mechanism to resist noise to some extent: using adversarial learning to get the generated augmented data allows the model to learn more about the distribution of the original data to extract more robust features, and to this view is weakening the negative impact of noisy samples. From the results of the above experimental section, we can observe that the strategy of generating adversarial improves the performance of our framework. However, in general, we still do not get particularly ideal detection results and classification accuracy due to the negative impact of noisy samples.
In summary, the current method still has some limitations, and the noise in many places using the current method can only reduce the impact as much as possible. But the noise problem still cannot be completely eliminated. How to better eliminate the impact of noisy data is the direction of our future research direction.
In this research, we develop a new framework termed time-sequential graph adversarial learning (TGAL) for incorporating community detection into an adversarial regularizer-directed graph representation learning process. Moreover, TGAL framework use a temporal graph attention encoder to incorporate topological input properties, temporal contextual representation, and latent factors. It also employs adversarial training with a neuroscientific prior to reconstructing the embedding space. In dynamic brain network datasets, our strategy surpassed several unsupervised deep embedding and community identification methods. Using the resulting graph representations for classification yielded better results than the comparison method, demonstrating that learning graph representations may result in more accurate graph embeddings in the latent space. Detailed model discussions were held to explore, among other concerns, the suggested TGAL and the superiority of the encoder and adversarial regularizer. Last but not least, our framework, which achieved improved performance, can be applied as a technology for medical-relevant research, and the detected dynamic brain communities altered by drugs can be used as an underlying biological marker to facilitate biomedical research on the pharmacodynamic properties and pharmacokinetics of neurological diseases and make efforts towards drug therapy for brain diseases.
This work was supported by the National Natural Science Foundations of China under Grants 62172403 and 61872351, the Distinguished Young Scholars Fund of Guangdong under Grant 2021B1515020019, the Excellent Young Scholars of Shenzhen under Grant RCYX20200714114641211 and Shenzhen Key Basic Research Projects under Grant JCYJ20200109115641762.
The authors declare there is no conflict of interest.
[1] |
V. R. Steele, J. M. Maurer, M. R. Arbabshirani, E. D. Claus, B. C. Fink, V. Rao, et al., Machine learning of functional magnetic resonance imaging network connectivity predicts substance abuse treatment completion, Biol. Psychiatry Cognit. Neurosci. Neuroimaging, 3 (2018), 141–149. https://doi.org/10.1016/j.bpsc.2017.07.003 doi: 10.1016/j.bpsc.2017.07.003
![]() |
[2] |
R. F. etzel, J. D. Medaglia, D. S. Bassett, Diversity of meso-scale architecture in human and non-human connectomes, Nat. Comm., 9 (2018), 346. https://doi.org/10.1038/s41467-017-02681-z doi: 10.1038/s41467-017-02681-z
![]() |
[3] |
M. Rubinov, O. Sporns, Weight-conserving characterization of complex functional brain networks, Neuroimage, 56 (2011), 2068–2079. https://doi.org/10.1016/j.neuroimage.2011.03.069 doi: 10.1016/j.neuroimage.2011.03.069
![]() |
[4] | R. F. Betzel, D. S. Bassett, Multi-scale brain networks, Neuroimage, 160 (2017), 73–83. https://doi.org/10.1016/j.neuroimage.2016.11.006 |
[5] |
M. Rubinov, O. Sporns, Complex network measures of brain connectivity: Uses and interpretations, Neuroimage, 52 (2010), 1059–1069. https://doi.org/10.1016/j.neuroimage.2009.10.003 doi: 10.1016/j.neuroimage.2009.10.003
![]() |
[6] | J. Pan, B. Lei, Y. Shen, Y. Liu, Z. Feng, S. Wang, Characterization multimodal connectivity of brain network by hypergraph GAN for Alzheimer's disease analysis, in Chinese Conference on Pattern Recognition and Computer Vision (PRCV), (2021), 467–478. https://doi.org/10.1007/978-3-030-88010-1_39 |
[7] | S. Hu, J. Yuan, S. Wang, Cross-modality synthesis from MRI to PET using adversarial U-net with different normalization, in 2019 International Conference on Medical Imaging Physics and Engineering (ICMIPE), (2019), 1–5. https://doi.org/10.1109/ICMIPE47306.2019.9098219 |
[8] | S. Hu, Y. Shen, S. Wang, B. Lei, Brain MR to PET synthesis via bidirectional generative adversarial network, in Medical Image Computing and Computer Assisted Intervention–-MICCAI 2020: 23rd International Conference, 10 (2020), 698–707. https://doi.org/10.1007/978-3-030-59713-9_67 |
[9] |
O. Sporns, Contributions and challenges for network models in cognitive neuroscience, Nat. Neurosci., 17 (2014), 652–660. https://doi.org/10.1038/nn.3690 doi: 10.1038/nn.3690
![]() |
[10] |
M. E. Newman, Communities, modules and large-scale structure in networks, Nat. Phys., 8 (2012), 25–31. https://doi.org/10.1038/nphys2162 doi: 10.1038/nphys2162
![]() |
[11] | O. Sporns, R. F. Betzel, Modular brain networks, Ann. Rev. Psychol., 67 (2016), 613–640. https://doi.org/10.1146/annurev-psych-122414-033634 |
[12] |
M. A. Bertolero, B. T. T. Yeo, M. D'Esposito, The modular and integrative functional architecture of the human brain, Proc. Nat. Acad. Sci. U. S. A., 112 (2015), E6798–E6807. https://doi.org/10.1073/pnas.1510619112 doi: 10.1073/pnas.1510619112
![]() |
[13] |
M. P. van den Heuvel, O. Sporns, Network hubs in the human brain, Trends Cogn. Sci., 17 (2013), 683–696. https://doi.org/10.1016/j.tics.2013.09.012 doi: 10.1016/j.tics.2013.09.012
![]() |
[14] |
S. Fortunato, Community detection in graphs, Phys. Rep., 486 (2010), 75–174. https://doi.org/10.1016/j.physrep.2009.11.002 doi: 10.1016/j.physrep.2009.11.002
![]() |
[15] | D. S. Bassett, O. Sporns, Network neuroscience, Nat. Neurosci., 20 (2017), 353–364. https://doi.org/10.1038/nn.4502 |
[16] |
M. Girvan, M. E. J. Newman, Community structure in social and biological networks, Proc. Nat. Acad. Sci. U. S. A., 99 (2002), 7821–7826. https://doi.org/10.1073/pnas.122653799 doi: 10.1073/pnas.122653799
![]() |
[17] | J. Pan, B. Lei, S. Wang, B. Wang, Y. Liu, Y. Shen, DecGAN: Decoupling generative adversarial network detecting abnormal neural circuits for Alzheimer's disease, preprint, arXiv: 2110.05712. |
[18] |
A. Avena-Koenigsberger, B. Misic, O. Sporns, Communication dynamics in complex brain networks, Nat. Rev. Neurosci., 19 (2018), 17–33. https://doi.org/10.1038/nrn.2017.149 doi: 10.1038/nrn.2017.149
![]() |
[19] | S. Hu, W. Yu, Z. Chen, S. Wang, Medical image reconstruction using generative adversarial network for Alzheimer disease assessment with class-imbalance problem, in 2020 IEEE 6th International Conference on Computer and Communications (ICCC), (2020), 1323–1327. https://doi.org/10.1109/ICCC51575.2020.9344912 |
[20] |
S. Wang, X. Li, J. Cui, H. X. Li, K. D. Luk, Y. Hu, Prediction of myelopathic level in cervical spondylotic myelopathy using diffusion tensor imaging, J. Magn. Reson. Imaging, 41 (2015), 1682–1688. https://doi.org/10.1002/jmri.24709 doi: 10.1002/jmri.24709
![]() |
[21] |
B. Lei, E. Liang, M. Yang, P. Yang, F. Zhou, E. L. Tan, et al., Predicting clinical scores for Alzheimer's disease based on joint and deep learning, Exp. Syst. Appl., 187 (2022), 115966. https://doi.org/10.1016/j.eswa.2021.115966 doi: 10.1016/j.eswa.2021.115966
![]() |
[22] | A. K. F. Lui, Y. H. Chan, M. F. Leung, Modelling of destinations for data-driven pedestrian trajectory prediction in public buildings, in 2021 IEEE International Conference on Big Data (Big Data), (2021), 1709–1717. https://doi.org/10.1109/BigData52589.2021.9671813 |
[23] |
S. Hu, B. Lei, S. Wang, Y. Wang, Z. Feng, Y. Shen, Bidirectional mapping generative adversarial networks for brain MR to PET synthesis, IEEE Trans. Med. Imaging, 41 (2021), 145–157. https://doi.org/10.1109/TMI.2021.3107013 doi: 10.1109/TMI.2021.3107013
![]() |
[24] | L. Yang, X. Cao, D. He, C. Wang, X. Wang, W. Zhang, Modularity based community detection with deep learning, in International Joint Conferences on Artificial Intelligence Organization, (2016), 2252–2258. https://doi.org/10.5555/3060832.3060936 |
[25] |
S. Wang, X. Wang, Y. Shen, B. He, X. Zhao, W. Cheung, et al., An ensemble-based densely-connected deep learning system for assessment of skeletal maturity, IEEE Trans. Syst. Man Cybern. Syst., 52 (2020), 426–437. https://doi.org/10.1109/TSMC.2020.2997852 doi: 10.1109/TSMC.2020.2997852
![]() |
[26] | G. Sperlí, A deep learning based community detection approach, in Proceedings of the 34th ACM/SIGAPP Symposium on Applied Computing, (2019), 1107–1110. https://doi.org/10.1145/3297280.3297574 |
[27] |
S. Wang, X. Wang, Y. Hu, Y. Shen, Z. Yang, M. Gan, et al., Diabetic retinopathy diagnosis using multichannel generative adversarial network with semisupervision, IEEE Trans. Autom. Sci. Eng., 18 (2021), 574–585. https://doi.org/10.1109/TASE.2020.2981637 doi: 10.1109/TASE.2020.2981637
![]() |
[28] |
W. Yu, B. Lei, M. K. Ng, A. C. Cheung, Y. Shen, S. Wang, Tensorizing GAN with high-order pooling for Alzheimer's disease assessment, IEEE Trans. Neural Networks Learn. Syst., 2021 (2021), 1–15. https://doi.org/10.1109/TNNLS.2021.3063516 doi: 10.1109/TNNLS.2021.3063516
![]() |
[29] |
B. Lei, Z. Xia, J. Feng, X. Jiang, Z. Ge, Y. Xu, et al., Skin lesion segmentation via generative adversarial networks with dual discriminators, Med. Image Anal., 64 (2020), 101716. https://doi.org/10.1016/j.media.2020.101716 doi: 10.1016/j.media.2020.101716
![]() |
[30] | Z. Chen, L. Li, J. Bruna, Supervised community detection with line graph neural networks, preprint, arXiv: 1705.08415. https://doi.org/10.48550/arXiv.1705.08415 |
[31] | G. Ian, P. A. Jean, M. Mehdi, Generative adversarial nets, Adv. Neural Inf. Process. Syst., 27 (2014), 1–9. |
[32] |
L. Mo, S. Wang, A variational approach to nonlinear two-point boundary value problems, Nonlinear Anal. Theory Methods Appl., 71 (2009), e834–e838. https://doi.org/10.1016/j.na.2008.12.006 doi: 10.1016/j.na.2008.12.006
![]() |
[33] | B. Jake, Notes on convolutional neural networks, Neural Nets, 2006 (2006), 1–8. |
[34] | T. N. Kipf, M. Welling, Semi-supervised classification with graph convolutional networks, preprint, arXiv: 1609.02907. |
[35] | W. Yu, B. Lei, S. Wang, Y. Liu, Z. Feng, Y. Hu, et al., Morphological feature visualization of Alzheimer's disease via multidirectional perception GAN, IEEE Trans. Neural Networks Learn. Syst., (2022), 1–15. https://doi.org/10.1109/TNNLS.2021.3118369 |
[36] | S. You, Y. Liu, B. Lei, S. Wang, Fine perceptive GANs for brain MR image super-resolution in wavelet domain, IEEE Trans. Neural Networks Learn. Syst., (2022), 1–13. https://doi.org/10.1109/TNNLS.2022.3153088 |
[37] | E. Abbe, Community detection and stochastic block models: Recent developments, J. Mach. Learn. Res., 18 (2018), 6446–6531. |
[38] | F. Liu, S. Xue, J. Wu, C. Zhou, W. Hu, C. Paris, et al., Deep learning for community detection: Progress, challenges and opportunities, in International Joint Conferences on Artificial Intelligence Organization, (2020), 4981–4987. https://doi.org/10.24963/ijcai.2020/693 |
[39] | P. Veličković, G. Cucurull, A. Casanova, A. Romero, P. Lio, Y. Bengio, Graph attention networks, preprint, arXiv: 1710.10903. |
[40] | A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, et al., Attention is all You need, in Proceedings of the 31st International Conference on Neural Information Processing Systems, (2017), 6000–6010. https://doi.org/10.5555/3295222.3295349 |
[41] | Z. Tao, H. Liu, J. Li, Z. Wang, Y. Fu, Adversarial graph embedding for ensemble clustering, in International Joint Conferences on Artificial Intelligence Organization, (2019), 3562–3568. https://doi.org/10.24963/ijcai.2019/494 |
[42] | I. Lobov, S. Ivanov, Unsupervised community detection with modularity-based attention model, preprint, arXiv: 1905.10350. |
[43] | W. Kool, H. van Hoof, M. Welling, Attention, learn to solve routing problems, in International Conference on Learning Representations, 2019. https://doi.org/10.48550/arXiv.1803.08475 |
[44] | T. Kipf, M. Welling, Variational graph auto-encoders, preprint, arXiv: 1611.07308. |
[45] | S. Pan, R. Hu, G. Long, J. Jiang, L. Yao, C. Zhang, Adversarially regularized graph autoencoder for graph embedding, in Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, (2018), 2609–2615. https://doi.org/10.24963/ijcai.2018/362 |
[46] | P. Veličković, W. Fedus, W. L. Hamilton, P. Lio, Y. Bengio, R. D. Hjelm, Deep graph infomax, in 7th International Conference on Learning Representations, ICLR 2019, 2018. https://doi.org/10.17863/CAM.40744 |
1. | Qiankun Zuo, Libin Lu, Lin Wang, Jiahui Zuo, Tao Ouyang, Constructing brain functional network by Adversarial Temporal-Spatial Aligned Transformer for early AD analysis, 2022, 16, 1662-453X, 10.3389/fnins.2022.1087176 | |
2. | Changwei Gong, Changhong Jing, Xin-an Liu, Victoria X. Wang, Cheuk Ying Tang, Paul J. Kenny, Ye Li, Zuxin Chen, Shuqiang Wang, Generative artificial intelligence-enabled dynamic detection of rat nicotine-related circuits, 2024, 36, 0941-0643, 4693, 10.1007/s00521-023-09307-0 | |
3. | Changwei Gong, Changhong Jing, Xuhang Chen, Chi Man Pun, Guoli Huang, Ashirbani Saha, Martin Nieuwoudt, Han-Xiong Li, Yong Hu, Shuqiang Wang, Generative AI for brain image computing and brain network computing: a review, 2023, 17, 1662-453X, 10.3389/fnins.2023.1203104 |
notation | description |
E | encoder |
G | generator |
D | discriminator |
X | attributes of brain region(node) |
A | adjacency(connectivity) matrix of brain networks |
H | hidden representations of region(node) attributes |
Z | embeddings of region(node) attributes |
ϕ | parameters of discriminator |
ψ | parameters of generator |
E | expectation |
i,j | No.i and No.j brain region |
encoder | AL | loss | Metrics(%) | ||||||
GCN | TG | K-means | modularity | C | Q | NMI | F1 | ||
√ | × | × | √ | × | 68.4 | 16.7 | 55.5 | 28.5 | |
√ | × | √ | √ | × | 72.5 | 24.3 | 49.6 | 44.4 | |
√ | × | × | × | √ | 28.4 | 64.5 | 50.8 | 60.2 | |
√ | × | √ | × | √ | 22.8 | 58.8 | 47.5 | 59.8 | |
× | √ | √ | × | √ | 42.9 | 69.5 | 67.5 | 66.7 |
notation | description |
E | encoder |
G | generator |
D | discriminator |
X | attributes of brain region(node) |
A | adjacency(connectivity) matrix of brain networks |
H | hidden representations of region(node) attributes |
Z | embeddings of region(node) attributes |
ϕ | parameters of discriminator |
ψ | parameters of generator |
E | expectation |
i,j | No.i and No.j brain region |
encoder | AL | loss | Metrics(%) | ||||||
GCN | TG | K-means | modularity | C | Q | NMI | F1 | ||
√ | × | × | √ | × | 68.4 | 16.7 | 55.5 | 28.5 | |
√ | × | √ | √ | × | 72.5 | 24.3 | 49.6 | 44.4 | |
√ | × | × | × | √ | 28.4 | 64.5 | 50.8 | 60.2 | |
√ | × | √ | × | √ | 22.8 | 58.8 | 47.5 | 59.8 | |
× | √ | √ | × | √ | 42.9 | 69.5 | 67.5 | 66.7 |