Abstract

Preclinical MR fingerprinting (MRF) suffers from long acquisition time for organ-level coverage due to demanding image resolution and limited undersampling capacity. This study aims to develop a deep learning-assisted fast MRF framework for sub-millimeter T1 and T2 mapping of entire macaque brain on a preclinical 9.4 T MR system. Three dimensional MRF images were reconstructed by singular value decomposition (SVD) compressed reconstruction. T1 and T2 mapping for each axial slice exploited a self-attention assisted residual U-Net to suppress aliasing-induced quantification errors, and the transmit-field (B1 + ) measurements for robustness against B1 + inhomogeneity. Supervised network training used MRF images simulated via virtual parametric maps and a desired undersampling scheme. This strategy bypassed the difficulties of acquiring fully sampled preclinical MRF data to guide network training. The proposed fast MRF framework was tested on experimental data acquired from ex vivo and in vivo macaque brains. The trained network showed reasonable adaptability to experimental MRF images, enabling robust delineation of various T1 and T2 distributions in the brain tissues. Further, the proposed MRF framework outperformed several existing fast MRF methods in handling the aliasing artifacts and capturing detailed cerebral structures in the mapping results. Parametric mapping of entire macaque brain at nominal resolution of 0.35 0.35 1 mm3 can be realized via a 20-min 3D MRF scan, which was sixfold faster than the baseline protocol. Introducing deep learning to MRF framework paves the way for efficient organ-level high-resolution quantitative MRI in preclinical applications.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.