Abstract

Real-world optimization problems may involve a number of computationally expensive functions with a large number of input variables. Metamodel-based optimization methods can reduce the computational costs of evaluating expensive functions, but this does not reduce the dimension of the search domain nor mitigate the curse of dimensionality effects. The dimension of the search domain can be reduced by functional anova decomposition involving Sobol' sensitivity indices. This approach allows one to rank decision variables according to their impact on the objective function values. On the basis of the sparsity of effects principle, typically only a small number of decision variables significantly affects an objective function. Therefore, neglecting the variables with the smallest impact should lead to an acceptably accurate and simpler metamodel for the original problem. This appealing strategy has been applied only to single-objective optimization problems so far. Given a high-dimensional optimization problem with multiple objectives, a method called anova-mop is proposed for defining a number of independent low-dimensional subproblems with a smaller number of objectives. The method allows one to define approximated solutions for the original problem by suitably combining the solutions of the subproblems. The quality of the approximated solutions and both practical and theoretical aspects related to decision making are discussed.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.