Sparse-view computed tomography (CT) is one of the primal means to reduce radiation risk. However, the reconstruction of sparse-view CT with classic analytical method is usually contaminated by severe artifacts. By designing the regularization terms carefully, iterative reconstruction (IR) algorithms can produce promising results. Aided by the powerful deep learning techniques, learned regularization terms with convolution neural network (CNN) have attracted much attention and can further improve performance. In this paper, to further enhance the performance of existing learnable regularization-based networks, we propose a learnable local-nonlocal regularization-based model called RegFormer for sparse-view CT reconstruction. Specifically, we unroll the iterative scheme into a neural network and replace the gradient of handcrafted regularization terms with learnable kernels. The convolution layers are used to learn gradient of local regularization, resulting in excellent denoising performance. In addition, the transformer-based encoders and decoders incorporate the learned nonlocal prior into the model, preserving the structures and details. To enhance the ability to extract deep features, we propose an iteration transmission (IT) module that further improve the efficiency of each iteration. The experimental results show that our proposed RegFormer outperforms several state-of-the-art methods in artifact reduction and detail preservation.
Read full abstract