Skip to content

FOD-Net: A Deep Learning Method for Fiber Orientation Distribution Angular Super Resolution

License

Notifications You must be signed in to change notification settings

ruizengalways/FOD-Net

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

17 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

License: CC BY-NC-SA 4.0


FOD-Net: A Deep Learning Method for Fiber Orientation Distribution Angular Super Resolution

Introduction

The fiber orientation distribution, which is able to perform high-quality tractography permitting the study of brain connectivity, is optimally estimated from high angular resolution diffusion imaging (HARDI) data. However, acquiring HARDI data is very time consuming, thus limiting broad application in both clinical and clinical research environments. In this regard, single-shell low angular resolution diffusion imaging (LARDI) data are more commonly used in clinical settings, and often combined with lower b-values (e.g. $b = 1000 s/mm^2$). However, LARDI acquisitions cannot reliably resolve complex white matter fiber bundles configurations due to insufficient angular resolution, producing both spurious and missing connections. This can degrade the connectome reconstruction, potentially impacting the validity of conclusions drawn from such data. To address this issue, we present fiber orientation distribution (FOD) network (FOD-Net), a deep-learning-based FOD angular super resolution method that directly enhances FOD data from single-shell LARDI computed from typical clinical-quality data, to obtain the super-resolved FOD data equivalent to those derived from high-quality multi-shell HARDI acquisitions, thereby enabling reliable tractography and structural connectome reconstruction using widely available clinical protocols.

The code was written and is under maintenance by Rui Zeng.

This updated PyTorch implementation produces results comparable to or better than our original pytorch software. If you would like to reproduce the same results as in the papers, check out the initial version of the paper in the BioRxiv.

If you use this code for your research, please cite:

FOD-Net: A Deep Learning Method for Fiber Orientation Distribution Angular Super Resolution.
Rui Zeng, Jinglei Lv, He Wang, Luping Zhou, Michael Barnett, Fernando Calamante*, Chenyu Wang*. In Medical Image Analysis. (* equal contributions) [Bibtex].

Abstract and Highlights

Our highlights can be concluded as follows:

  • A deep-learning-based framework for FOD angular super-resolution in dMRI data is proposed (Please see the architecture in Figure 1).
  • FOD-Net improves the FOD angular resolution of FOD images acquired with single-shell LARDI dMRI data to approximate the quality of multi-shell high-angular-resolution dMRI FODs (Please see the demonstration in Figure 2).
  • The benefits of the super-resolved FODs in tractography and connectomics are demonstrated compared to single-shell LARDI dMRI data (please see Figure 3 and Figure 4 for tractography and brain connectivity improvements).


Figure 1 FOD-Net achitecture


Figure 2 FOD super resolution demonstration


Figure 3 Tractography improvement


Figure 4 Connectomic improvement

Prerequisites

Getting Started

Installation

  • Clone this repo:
git clone https://github.com/ruizengalways/FOD-Net.git
cd FOD-Net
  • Install PyTorch and 1.4.0 and other dependencies (e.g., nibabel and dipy).
  • Please use pip install to install the missing python packages if ImportError: No module named error message occurs

Human Connectome Project dataset preparation

The description about this dataset can be found in this link

To train FOD-Net, we first need to generate single-shell LARDI dMRI data from multi-shell dMRI data first. As the size of HCP dataset is too large, I have prepared the toy dataset, which only contains one subject, for you to play around. Please download this compressed dataset (link) and uncompress it into /dataset.

The directory tree of /dataset should look like the following after uncompressing the file:

.
└── 111009
    ├── brain_mask.nii.gz
    ├── fsl_5ttgen.nii.gz
    ├── HARDI_data
    │   ├── bvals
    │   ├── bvecs
    │   └── data.nii.gz
    └── LARDI_data
        ├── data_b1000_g32.bvals
        ├── data_b1000_g32_bvecs
        └── data_b1000_g32.nii.gz

HARDI_data refers to the original dMRI data copied from the Human Connectome Project dataset. bvecs and bvals are its corresponding diffusion gradient table and b value respectively (288 directions in total). LARDI_data is the angular downsampling version of HARDI_data, in which diffusion gradients and b values are downsampled to a number equaling to 32. The method for downsampling diffusion MRI is introduced in the original FOD-Net paper. brain_mask.nii.gz is the brain mask of the subject. fsl_5ttgen.nii.gz is the five-tissue mask generated using 5ttgen (link) in MRTrix.

  1. Perform ./util/HCP_dataset_preparation.py to generate normalised multi-shell multi-tissue constrained spherical deconvolution (MSMT CSD) FOD images for HARDI data as well obtaining normalised single-shell three-tissue constrained spherical deconvolution (SS3T CSD) FOD images for LARDI data. All FOD images are normalised using mtnormalise (link) to put them into the same scalar level. Once the process is done, you should get the directory tree as the following:
.
└── 111009
    ├── brain_mask.nii.gz
    ├── fsl_5ttgen.nii.gz
    ├── HARDI_data
    │   ├── bvals
    │   ├── bvecs
    │   └── data.nii.gz
    ├── LARDI_data
    │   ├── data_b1000_g32.bvals
    │   ├── data_b1000_g32_bvecs
    │   └── data_b1000_g32.nii.gz
    ├── msmt_csd
    │   ├── CSF_FODs.nii.gz
    │   ├── CSF_FODs_normalised.nii.gz
    │   ├── GM_FODs.nii.gz
    │   ├── GM_FODs_normalised.nii.gz
    │   ├── RF_CSF.txt
    │   ├── RF_GM.txt
    │   ├── RF_WM.txt
    │   ├── WM_FODs.nii.gz
    │   └── WM_FODs_normalised.nii.gz
    └── ss3t_csd
        ├── CSF_FODs.nii.gz
        ├── CSF_FODs_normalised.nii.gz
        ├── GM_FODs.nii.gz
        ├── GM_FODs_normalised.nii.gz
        ├── RF_CSF.txt
        ├── RF_GM.txt
        ├── RF_WM.txt
        ├── WM_FODs.nii.gz
        └── WM_FODs_normalised.nii.gz

The ./${subject_id}/ss3t_csd/WM_FODs_normalised.nii.gz and ./${subject_id}/msmt_csd/WM_FODs_normalised.nii.gz will be used as the LARDI (low resolution input) and HARDI data (ground truth target) pair for the training process.

FODNet train

To train a model, please run

python train.py --dataroot ./dataset/ --dataset_mode hcp --cfg ./config/fodnet_updated_config.yml

You can adjust the FOD-Net parameters either using arguments or cfg files located at ./config/*.yml

FODNet inference

As this is a toy dataset, ./${subject_id}/ss3t_csd/WM_FODs_normalised.nii.gz is also served as a input for testing the performance of our model. The output path is set to ./dataset/${subject_id}/fodnet/SR_WM_FODs_normalised.nii.gz for the sake of convenience.

If you do not have time to train the model, you could download the pretrained model from this link, and then put it into ./

Please run the following

python test.py --fod_path ./dataset/${subject_id}/ss3t_csd/WM_FODs_normalised.nii.gz --weights_path ./fodnet.pth --output_path ./dataset/${subject_id}/fodnet/SR_WM_FODs_normalised.nii.gz --brain_mask_path ./dataset/${subject_id}/brain_mask.nii.gz --gpu_ids 0

For each subject, it takes around 3 hours to finish inference. Do some sports and then come back to enjoy the super-resolved FOD image!

Tractography and connectomic

We use tckgen link and tck2connectome link. Please read the documents to learn.

Citation

If you use this code for your research or find our work is inspiring, please cite our papers.

@article{ZENG2022102431,
title = {FOD-Net: A deep learning method for fiber orientation distribution angular super resolution},
journal = {Medical Image Analysis},
volume = {79},
pages = {102431},
year = {2022},
issn = {1361-8415},
doi = {https://doi.org/10.1016/j.media.2022.102431},
url = {https://www.sciencedirect.com/science/article/pii/S1361841522000822},
author = {Rui Zeng and Jinglei Lv and He Wang and Luping Zhou and Michael Barnett and Fernando Calamante and Chenyu Wang},
keywords = {Fiber orientation distribution, Connectomics, Angular super resolution, Diffusion magnetic resonance imaging},
abstract = {Mapping the human connectome using fiber-tracking permits the study of brain connectivity and yields new insights into neuroscience. However, reliable connectome reconstruction using diffusion magnetic resonance imaging (dMRI) data acquired by widely available clinical protocols remains challenging, thus limiting the connectome/tractography clinical applications. Here we develop fiber orientation distribution (FOD) network (FOD-Net), a deep-learning-based framework for FOD angular super-resolution. Our method enhances the angular resolution of FOD images computed from common clinical-quality dMRI data, to obtain FODs with quality comparable to those produced from advanced research scanners. Super-resolved FOD images enable superior tractography and structural connectome reconstruction from clinical protocols. The method was trained and tested with high-quality data from the Human Connectome Project (HCP) and further validated with a local clinical 3.0T scanner as well as with another public available multicenter-multiscanner dataset. Using this method, we improve the angular resolution of FOD images acquired with typical single-shell low-angular-resolution dMRI data (e.g., 32 directions, b=1000 s/mm2) to approximate the quality of FODs derived from time-consuming, multi-shell high-angular-resolution dMRI research protocols. We also demonstrate tractography improvement, removing spurious connections and bridging missing connections. We further demonstrate that connectomes reconstructed by super-resolved FODs achieve comparable results to those obtained with more advanced dMRI acquisition protocols, on both HCP and clinical 3.0T data. Advances in deep-learning approaches used in FOD-Net facilitate the generation of high quality tractography/connectome analysis from existing clinical MRI environments. Our code is freely available at https://github.com/ruizengalways/FOD-Net.}
}

License: CC BY-NC-SA 4.0

About

FOD-Net: A Deep Learning Method for Fiber Orientation Distribution Angular Super Resolution

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages