Abstract

Image quality metrics (IQMs) such as root mean square error (RMSE) and structural similarity index (SSIM) are commonly used in the evaluation and optimization of accelerated magnetic resonance imaging (MRI) acquisition and reconstruction strategies. However, it is unknown how well these indices relate to a radiologist's perception of diagnostic image quality. In this study, we compare the image quality scores of five radiologists with the RMSE, SSIM, and other potentially useful IQMs: peak signal to noise ratio (PSNR) multi-scale SSIM (MSSSIM), information-weighted SSIM (IWSSIM), gradient magnitude similarity deviation (GMSD), feature similarity index (FSIM), high dynamic range visible difference predictor (HDRVDP), noise quality metric (NQM), and visual information fidelity (VIF). The comparison uses a database of MR images of the brain and abdomen that have been retrospectively degraded by noise, blurring, undersampling, motion, and wavelet compression for a total of 414 degraded images. A total of 1017 subjective scores were assigned by five radiologists. IQM performance was measured via the Spearman rank order correlation coefficient (SROCC) and statistically significant differences in the residuals of the IQM scores and radiologists' scores were tested. When considering SROCC calculated from combining scores from all radiologists across all image types, RMSE and SSIM had lower SROCC than six of the other IQMs included in the study (VIF, FSIM, NQM, GMSD, IWSSIM, and HDRVDP). In no case did SSIM have a higher SROCC or significantly smaller residuals than RMSE. These results should be considered when choosing an IQM in future imaging studies.

Highlights

  • T HE quality of a magnetic resonance (MR) image can be difficult to assess in a robust, objective and quantitative manner

  • Ten full-reference objective image quality metrics (IQMs) were included in this study: root mean square error (RMSE), peak signal to noise ratio (PSNR), structural similarity index (SSIM), multiscale SSIM (MSSSIM [25]), information-weighted SSIM (IWSSIM [26]), gradient magnitude similarity deviation (GMSD [27]), feature similarity index (FSIM [28]), high dynamic range visible difference predictor (HDRVDP [29]), noise quality metric (NQM [30]), and visual information fidelity (VIF [31])

  • IWSSIM performed poorly for images degraded by undersampling artifacts, showing statistically larger residuals compared to all other IQMs for these images

Read more

Summary

INTRODUCTION

T HE quality of a magnetic resonance (MR) image can be difficult to assess in a robust, objective and quantitative manner. This leads to challenges in the comparison of different image acquisition and reconstruction techniques and the validation of new ones. Clinical MR images are typically viewed by an expert radiologist, so the radiologist’s opinion of the diagnostic quality of the image can be considered an appropriate measure of MR image quality Applying this standard on a large scale is often infeasible due to large image library sizes, limited time availability of expert radiologists, and the inherent variability in a subjective scoring technique. The relationship between radiologists’ opinion of medical image quality and IQM scores is not well explored

Objective
Generation of Image Library
Objective IQMs
Radiologist Image Quality Assessment
Data Analysis
IQM Calculation Times
RESULTS
DISCUSSION
CONCLUSION
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call