Abstract
When machine learning (ML) is used to solve practical problems in structural engineering (SE), the associated data may not reflect the overall distribution of the population. With this in mind, a common practice in prior ML-SE studies is to optimize the model such that the testing error is minimized. However, this strategy can lead to a situation where the testing error is significantly greater than the error obtained during training, which is a clear sign of overfitting. Therefore, simply selecting a model with the highest testing performance does not necessarily guarantee generalizability. To address this challenge, this study aims to provide some practical guidelines that consider both the model complexity and generalizability. The proposed solution is fairly straightforward to implement. The generalized bias-variance decomposition is first introduced, followed by a discussion of different error variation patterns (bell-shaped, double-descent, and monotonically decreasing). A mathematical formulation that establishes a relationship between the model complexity and generalizability via an objective function is then presented to facilitate the model parameter selection. Subsequently, four numerical experiments are conducted to apply the formulation within the structural engineering domain. Finally, the proposed model selection approach is compared with existing methods to demonstrate its advantages and limitations.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.