Abstract

A nonsmooth distributed optimization problem subject to affine equality and convex inequality is considered in this paper. All the local objective functions in the distributed optimization problem possess a common decision variable. And taking privacy into consideration, each agent doesn’t share its local information with other agents, including the information about the local objective function and constraint set. To cope with this distributed optimization, a neurodynamic approach based on the penalty-like methods is proposed. It is proved that the presented neurodynamic approach is convergent to an optimal solution to the considered distributed optimization problem. The proposed neurodynamic approach in this paper has lower model complexity and computational load via reducing auxiliary variables. In the end, two illustrative examples are given to show the effectiveness and practical application of the proposed neural network.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.