Abstract When performing time-intensive optimization tasks, such as those in Topology or Shape Optimization, researchers have turned to machine-learned (ML) Inverse Design methods i.e., predicting the optimized geometry from input conditions to replace or warm start traditional optimizers. Such methods are often optimized to reduce the Mean Squared Error (MSE) or Binary Cross-Entropy between the output and a training dataset of optimized designs. While convenient, we show that this choice may be myopic. Specifically, we compare two methods of optimizing the hyper-parameters of easily reproducible Machine Learning models including random forest (RF), k-nearest neighbors (KNN), and Deconvolutional Neural Network (DeCNN) model for predicting the three optimal topology problems. We show that under both direct Inverse Design and when warm starting further Topology Optimization (TO), using MSE metrics to tune hyper-parameters produces less performant models than directly evaluating the objective function, though both produce designs that are almost one order of magnitude better than using the common uniform initialization. We also illustrate how warm starting impacts both the convergence time, the type of solutions obtained during optimization, and the final designs. Overall, our initial results portend that researchers may need to revisit common choices for evaluating ID methods that subtly trade-off factors in how an ID method will actually be used. We hope our open-source dataset and evaluation environment will spur additional research in those directions.
Read full abstract