Abstract
A novel method based on minimizing the Euclidean distance is proposed for generating a well-distributed Pareto set in multi-objective optimization for a class of ergodic controllable Markov chains. The proposed approach is based on the concept of strong Pareto policy. We consider the case where the search space is a non-strictly convex set. For solving the problem we introduce the Tikhonov’s regularization method and implement the Lagrange principle. We formulate the original problem introducing linear constraints over the nonlinear problem employing the c-variable method and constraining the cost-functions allowing points in the Pareto front to have a small distance from one another. As a result, the proposed method generates an even representation of the entire Pareto surface. Then, we propose an algorithm to compute the Pareto front and provide all the details needed to implement the method in an efficient and numerically stable way. As well, we prove the main Theorems for describing the dependence of the saddle point for the regularizing parameter and analyzes its asymptotic behavior. Moreover, we analyze the step size parameter of the Lagrange principle and also its asymptotic behavior. The suggested approach is validated theoretically and verified by a numerical example related to security patrolling that present a technique for visualizing the Pareto front.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.