Source code for “Exploring Dynamic Interpretable Brain Networks via Hierarchical Graph Transformer”, published in Pattern Recognition.
Authors: Hao Hu†, Rundong Xue†, Shaoyi Du, Xiangmin Han, Jingxi Feng, Zeyu Zhang, Wei Zeng, Yue Gao, Juan Wang
- Dynamic Brain Transformer: learns time-varying functional connectivity for adaptive graph construction.
- Hierarchical Representation Learning: models intra-subnetwork homogeneity and inter-subnetwork heterogeneity.
- Cross-scale Modeling: bridges ROI-level dynamics and subnetwork-level coordination.
- Validated on 4 datasets for neurological disorder diagnosis.
Default config: setting/abide.yaml
Common options:
data.time_seires: path to*.npytrain.epochs/lr/weight_decay- Hierarchical constraints:
train.group_loss(Lintra+Linter)train.hierarchical_loss(LKL)train.hier_alpha/train.hier_beta/train.hier_gamma
Download the ABIDE dataset from here.
cd DIBrain
python main.py --config_filename setting/abide.yaml@article{hu2026exploring,
title={Exploring Dynamic Interpretable Brain Networks via Hierarchical Graph Transformer},
author={Hu, Hao and Xue, Rundong and Du, Shaoyi and Han, Xiangmin and Feng, Jingxi and Zhang, Zeyu and Zeng, Wei and Gao, Yue and Wang, Juan},
journal={Pattern Recognition},
pages={113371},
year={2026},
publisher={Elsevier}
}The source code is free for research and educational use only. Any commercial use should get formal permission first.