Abstract

In the research area of structural reliability analysis (SRA), the dominant failure modes (DFMs) of a structural system make significant contributions to life-span failure prediction and safety assessment. However, the high computational cost caused by the combinatorial explosion is the main problem in DFMs searching that hinders its application and further development. Recently, many successful applications have proved that the self-play deep reinforcement learning (DRL) has a strong ability to obtain action policy in the face of combinatorial explosion problems. Inspired by this, a self-play strategy is designed to optimize the DRL-based DFMs searching process and reduce the computational effort. A scoring function is designed and used as the refereeing standard of the self-play games and helps improve the efficiency of Monte Carlo tree search (MCTS) in an asynchronous training process. In comparison with the β-unzipping method and exploration-based DFMs searching method, the proposed method significantly improved training efficiency with an accuracy of over 95% and a lower requirement of the number of finite element analysis (FEA), both of which contribute to the policy learning of failure component selection. In summary, the method shows potential applications for actual structures and makes valuable contributions to the problem with high computing costs.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.