The vigorous development of deep learning (DL) has been propelled by big data and high-performance computing. For brain–computer interfaces (BCIs) to benefit from DL in a reliable and scalable manner, the scale and quality of data are crucial. Special emphasis is placed on the zero-shot learning (ZSL) paradigm, which is essential for enhancing the flexibility and scalability of BCI systems. ZSL enables models to generalise from limited examples to new, unseen tasks, addressing data scarcity challenges and accelerating the development of robust, adaptable BCIs. Despite a growing number of BCI surveys in recent years, there is a notable gap in clearly presenting public data resources. This paper explores the fundamental data capital necessary for large-scale deep learning BCI (DBCI) models. Our key contributions include (1) a systematic review and comprehensive understanding of the current industrial landscape of DBCI datasets; (2) an in-depth analysis of research gaps and trends in DBCI devices, data and applications, offering insights into the progress and prospects for high-quality data foundation and developing large-scale DBCI models; (3) a focus on the paradigm shift brought by ZSL, which is pivotal for the technical potential and readiness of BCIs in the era of multimodal large AI models.
Read full abstract