Given input and corresponding desired output , the overall goal of the IB principle [1] is to learn a latent representation that is maximally predictive to and contains as little information of as possible. Formally, the objective of IB can be formulated as:
where denotes mutual information and is a Lagrange multiplier that controls the trade-off between the sufficiency (the performance to down-stream task, as measured by $\max I(Y;T)$) and the minimality (the complexity of the representation, as quantified by $\min I(X;T)$).
To implement IB with deep neural networks, the maximization of equals to the minimization of cross-entropy loss; whereas the minimization of differs in each method by mutual information variational or non-parametric upper bounds. BrainIB estimates by the matrix-based Renyi’s α-order entropy [2,3] without any approximations or distributional assumptions.
Figure 1 The overview of the pipeline. The resting-state fMRI raw data are preprocessed and then parcellated into regions of interest (ROIs) according to the automated anatomical labelling (AAL) atlas. The functional connectivity (FC) matrices are calculated using Pearson correlation between ROIs. From the FC we construct the brain functional graph G = {A,X}, where A is the graph adjacency matrix characterizing the graph structure and X is node feature matrix. Specifically, A is a binarized FC matrix, where only the top 20-percentile absolute values of the correlations of the matrix are transformed into ones, while the rest are transformed into zeros. For node feature
Figure 2 Architecture of our proposed BrainIB. BrainIB consists of three modules: subgraph generator, graph encoder, and mutual information estimation module. The subgraph generator is used to sample subgraph G from the original graph Gsub. The graph encoder is used to learn graph embeddings.The mutual information estimation module evaluates the mutual information between G or Gsub.
where $ \mathcal{G} $ is input graph.
$ \mathcal{G}_{\text{sub}}=\mathcal{G}\odot M, $
where
where
where
We use matrix-based Renyi’s α-order mutual information to estimate , which significantly stabilizes the training.
Figure 3 Training dynamics of
BrainIB achieves better accuracy for the leave-one-site-out cross validation on REST-meta-MDD and ABIDE.
Table 1 Leave-one-site-out cross validation on REST-meta-MDD and ABIDE. The highest performance is highlighted with bold face.
Rest-meta-MDD |
ABIDE |
||||||
---|---|---|---|---|---|---|---|
Site |
DIR-GNN |
ProtGNN |
BrainIB |
Site |
DIR-GNN |
ProtGNN |
BrainIB |
site1 |
56.8% |
56.8% |
63.3% |
CMU |
83.3% |
75.0% |
83.3% |
site2 |
70.0% |
70.0% |
70.0% |
CALTECH |
68.4% |
68.4% |
71.1% |
site3 |
78.0% |
68.3% |
85.4% |
KKI |
65.5% |
65.5% |
72.7% |
site4 |
75.0% |
63.9% |
77.8% |
LEUVEN |
68.8% |
73.4% |
73.4% |
site5 |
64.4% |
63.2% |
67.8% |
MAX_MUN |
64.9% |
68.4% |
66.7% |
site6 |
68.8% |
64.6% |
68.8% |
NYU |
63.6% |
57.1% |
70.1% |
site7 |
70.4% |
67.6% |
73.2% |
OHSU |
71.4% |
71.4% |
67.9% |
site8 |
67.6% |
75.7% |
75.7% |
OLIN |
77.8% |
75.0% |
75.0% |
site9 |
80.6% |
72.2% |
75.0% |
PITT |
70.2% |
64.9% |
66.7% |
site10 |
71.0% |
66.7% |
72.0% |
SBL |
76.7% |
73.3% |
83.3% |
site11 |
70.0% |
64.2% |
82.1% |
SDSU |
75.0% |
72.2% |
75.0% |
site12 |
64.6% |
63.4% |
67.1% |
STANFORD |
80.0% |
72.5% |
75.0% |
site13 |
67.3% |
65.3% |
69.4% |
TRINITY |
67.3% |
73.5% |
65.3% |
site14 |
56.6% |
56.8% |
63.2% |
UCLA |
68.7% |
64.6% |
74.7% |
site15 |
61.1% |
62.5% |
70.1% |
UM |
66.2% |
62.8% |
64.8% |
site16 |
68.4% |
73.7% |
71.1% |
USM |
67.3% |
61.4% |
73.3% |
site17 |
68.9% |
71.1% |
68.9% |
YALE |
73.2% |
67.9% |
82.1% |
Mean |
68.2% |
66.2% |
71.8% |
Mean |
71.1% |
68.7% |
73.0% |
Figure 4 Comparison of explanation graph connections in brain networks of healthy controls and patients on MDD datasets. The colors of brain neural systems are described as: visual network, somatomotor network, dorsal attention network, ventral attention network, limbic network, frontoparietal network, default mode network, cerebellum and subcortial network respectively. Patients with MDD exhibits tight interactions between default mode network and limbic network, while these connections in healthy controls are much sparser.
We provide two papers to illustrate the BrainIB:
-
Title: BrainIB: Interpretable Brain Network-based Psychiatric Diagnosis with Graph Information Bottleneck
Published: IEEE Transactions on Neural Networks and Learning Systems (accepted)
Arxiv: https://arxiv.org/abs/2205.03612
Details:
where
-
Title: Towards a more stable and general subgraph information bottleneck
Published: ICASSP 2023
Details:
where $[ \cdot;\cdot;\cdot] $ is the vector concatenation operation,
We provide two demos: 1) BrainIB_V1 (IEEE TNNLS [6]) on ABIDE dataset (Figure 1-3 in the manuscript); and 2) BrainIB_V2 (IEEE ICASSP [7]) on ABIDE dataset.
Data is available at google drive (https://drive.google.com/drive/folders/1EkvBOoXF0MB2Kva9l4GQbuWX25Yp81a8?usp=sharing).
[1] N. Tishby, F. C. Pereira, and W. Bialek, “The information bottleneck method,” in Proc. 37th Annual Allerton Conference on Communications, Control and Computing,
1999, pp. 368–377.
[2] Giraldo, Luis Gonzalo Sanchez, Murali Rao, and Jose C. Principe. "Measures of entropy from data using infinitely divisible kernels." IEEE Transactions on Information Theory 61.1 (2014): 535-548.
[3] Yu, Shujian, et al. "Multivariate Extension of Matrix-Based Rényi's
[4] Gallo, Selene, et al. "Functional connectivity signatures of major depressive disorder: machine learning analysis of two multicenter neuroimaging studies." Molecular Psychiatry (2023): 1-10.
[5] J. Yu, T. Xu, and Y. Rong, “Graph information bottleneck for subgraph recognition,”
in International Conference on Learning Representations, 2020
[6] Zheng, Kaizhong, et al. "BrainIB: Interpretable brain network-based psychiatric diagnosis with graph information bottleneck.", under major revision of IEEE Transactions on Neural Networks and Learning Systems.
[7] Liu, Hongzhi, et al. “Towards a more stable and general subgraph information bottleneck”, accepted by IEEE ICASSP-23 (oral presentation)
If you have any questions, please feel free to contact us by [email protected] (Shujian Yu) or [email protected] (Kaizhong Zheng).