Abstract
Abstract As the field of quantum physics evolves, researchers naturally form subgroups focusing on specialized problems. 
While this encourages in-depth exploration, it can limit the exchange of ideas across structurally similar problems in different subfields. 
To encourage cross-talk among these different specialized areas, data-driven approaches using machine learning have recently shown promise to uncover meaningful connections between research concepts, promoting cross-disciplinary innovation.
Current state-of-the-art approaches represent concepts using knowledge graphs and frame the task as a link prediction problem, where connections between concepts are explicitly modeled.
In this work, we introduce a novel approach based on dynamic word embeddings for concept combination prediction. 
Unlike knowledge graphs, our method captures implicit relationships between concepts, can be learned in a fully unsupervised manner, and encodes a broader spectrum of information. 
We demonstrate that this representation enables accurate predictions about the co-occurrence of concepts within research abstracts over time.
To validate the effectiveness of our approach, we provide a comprehensive benchmark against existing methods and offer insights into the interpretability of these embeddings, particularly in the context of quantum physics research. 
Our findings suggest that this representation offers a more flexible and informative way of modeling conceptual relationships in scientific literature.
Published Version
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have