Abstract

Discrete thresholded neural networks of the Hopfield type have been employed to recover the regularized least-squares (LS) solu- tion from noisy image data. The goal is to attain this solution efficiently by conserving computational and storage requirements as the dimensions of the problem grow large. This paper discusses configurations of these networks that recover the LS solution by partitioning the networks and adopting a switching operation between active and inactive partitions to optimize the objective function. Sequential and parallel update proce- dures on active partitions offer a means to avoid limit cycling and impos- ing zero-self-feedback constraints. Examples of image noise reduction and the identification of the LS solution from corrupted image data are presented. © 2003 Society of Photo-Optical Instrumentation Engineers.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.