Abstract

Machine learning, by virtue of its powerful predictive capabilities in discerning material properties, has emerged as an indispensable tool in the pursuit of new material discovery and design. Its specific impact on advanced catalyst design is unlocking breakthroughs across various scientific disciplines. However, the success of machine learning (ML) in predicting material properties is strongly influenced by the size of the dataset. Small datasets can lead to overfitting, resulting in suboptimal predictive performance. In this study, we compare various pre-training models that leverage expert knowledge with those based on large, mixed datasets for specific downstream property prediction tasks. We argue that transfer learning, particularly self-supervised learning based on large datasets, offers more generalizable, scalable, and accurate predictions than supervised pre-training models. Our analysis demonstrates the advantages of self-supervised pre-training, particularly in cases where expert knowledge is limited or insufficient. We evaluate the performance of these models on the challenging task of predicting the adsorption energy associated with CO2 reduction by M−X−C single-atom catalysts. Using the optimal model, the self-supervised pre-training model, we conducted a screening of single-atom catalysts for CO2 reduction to CO, leading to the discovery of two new, more efficient catalysts, namely Zn-NP3 and Sn-P4. Our findings highlight the potential of self-supervised pre-training models in advancing material property prediction and guiding the discovery of novel materials.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call