Abstract

In this paper, we consider a regularized least squares problem subject to convex constraints. Our algorithm is based on the superiorization technique, equipped with a new step size rule which uses subgradient projections. The superiorization method is a two-step method where one step reduces the value of the penalty term and the other step reduces the residual of the underlying linear system (using an algorithmic operator T). For the new step size rule, we present a convergence analysis for the case when T belongs to a large subclass of strictly quasi-nonexpansive operators. To examine our algorithm numerically, we consider box constraints and use the total variation (TV) functional as a regularization term. The specific test cases are chosen from computed tomography using both noisy and noiseless data. We compare our algorithm with previously used parameters in superiorization. The T operator is based on sequential block iteration (for which our convergence analysis is valid), but we also use the conjugate gradient method (without theoretical support). Finally, we compare with the well-known “fast iterative shrinkage-thresholding algorithm” (FISTA). The numerical results demonstrate that our new step size rule improves previous step size rules for the superiorization methodology and is competitive with, and in several instances behaves better than, the other methods.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.