Articles published on Representation Learning
Authors
Select Authors
Journals
Select Journals
Duration
Select Duration
13133 Search results
Sort by Recency
- New
- Research Article
- 10.1016/j.neunet.2025.108438
- Apr 1, 2026
- Neural networks : the official journal of the International Neural Network Society
- Zhen Hao Wong + 3 more
Robust heterogeneous network representation learning by multifaceted curriculum training.
- New
- Research Article
1
- 10.1016/j.engappai.2026.114108
- Apr 1, 2026
- Engineering Applications of Artificial Intelligence
- Xiaonan Ni + 6 more
Structural-aware key node identification in hypergraphs via representation learning and fine-tuning
- New
- Research Article
- 10.1016/j.jprocont.2026.103677
- Apr 1, 2026
- Journal of Process Control
- Syed Meesam Raza Naqvi + 4 more
Anomaly detection using case-based reasoning and representation learning: Application to industrial maintenance
- New
- Research Article
- 10.1016/j.dsp.2026.105937
- Apr 1, 2026
- Digital Signal Processing
- Xiaoran Li + 1 more
Harnessing structure-aware graph representation and adaptive anchor graph learning for multi-view clustering
- New
- Research Article
- 10.1109/tpami.2025.3645279
- Apr 1, 2026
- IEEE transactions on pattern analysis and machine intelligence
- Xiaoxia Zhang + 3 more
Graph Neural Networks (GNNs) have made significant strides in the analysis and modeling of complex network data, particularly excelling in graph and node classification tasks. However, the "closed box" nature of GNNs impedes user understanding and trust, thereby restricting their broader application. This challenge has spurred a growing focus on demystifying GNNs to make their decision-making processes more transparent. Traditional methods for explaining GNNs often rely on selecting subgraphs and employing combinatorial optimization to generate understandable outputs. However, these methods are closely linked to the inherent complexity of GNNs, leading to higher explanation costs. To address this issue, we introduce a lower-complexity proxy model to explain GNNs. Our approach leverages knowledge distillation with inter-layer alignment, specifically targeting the challenge of over-smoothing and its detrimental impact on model explanation. Initially, we distill critical insights from complex GNN models into a more manageable proxy model. We then apply an inter-layer alignment-based distillation technique to ensure alignment between the proxy and the original model, facilitating the extraction of node or edge-level explanations within the proxy framework. We theoretically prove that the explanations derived from the proxy model are faithful to both the proxy and the original model. Additionally, we show that the upper bound of unfaithfulness between the proxy and the original model remains consistent when the distillation error is infinitesimal. This inter-layer alignment knowledge distillation technique enables the proxy model to retain the knowledge learning and topological representation capabilities of the original model to the greatest extent. Experimental evaluations on numerous real-world datasets confirm the effectiveness of our method, demonstrating robust performance.
- New
- Research Article
- 10.1016/j.knosys.2026.115634
- Apr 1, 2026
- Knowledge-Based Systems
- Chunyan Li + 3 more
Disentangled molecular representation learning with context-aware codebook for OOD generalization
- New
- Research Article
- 10.1016/j.ins.2025.122777
- Apr 1, 2026
- Information Sciences
- Liang Zhang + 3 more
Hierarchical prototype-guided representation learning for robust graph classification
- New
- Research Article
5
- 10.1109/tpami.2025.3643453
- Apr 1, 2026
- IEEE transactions on pattern analysis and machine intelligence
- Hanbo Bi + 13 more
The rapid advancement of foundation models has revolutionized visual representation learning in a self-supervised manner. However, their application in remote sensing (RS) remains constrained by a fundamental gap: existing models predominantly handle single or limited modalities, overlooking the inherently multi-modal nature of RS observations. Optical, synthetic aperture radar (SAR), and multi-spectral data offer complementary insights that significantly reduce the inherent ambiguity and uncertainty in single-source analysis. To bridge this gap, we introduce RingMoE, a unified multi-modal RS foundation model with 14.7 billion parameters, pre-trained on 400 million multi-modal RS images from nine satellites. RingMoE incorporates three key innovations: 1) A hierarchical Mixture-of-Experts (MoE) architecture comprising modal-specialized, collaborative, and shared experts, effectively modeling intra-modal knowledge while capturing cross-modal dependencies to mitigate conflicts between modal representations; 2) Physics-informed self-supervised learning, explicitly embedding sensor-specific radiometric characteristics into the pre-training objectives; 3) Dynamic expert pruning, enabling adaptive model compression from 14.7B to 1B parameters while maintaining performance, facilitating efficient deployment in Earth observation applications. Evaluated across 23 benchmarks spanning six key RS tasks (i.e., classification, detection, segmentation, tracking, change detection, and depth estimation), RingMoE outperforms existing foundation models and sets new SOTAs, demonstrating remarkable adaptability from single-modal to multi-modal scenarios. Beyond theoretical progress, it has been deployed and trialed in multiple sectors, including emergency response, land management, marine sciences, and urban planning.
- New
- Research Article
- 10.1016/j.asoc.2026.114776
- Apr 1, 2026
- Applied Soft Computing
- Yongbo Ni + 1 more
Multi-view feature representation learning: Integrating medical team information to enhance online physician recommendation
- New
- Research Article
- 10.1016/j.ipm.2025.104557
- Apr 1, 2026
- Information Processing & Management
- Zhentao Liang + 4 more
Citation importance-aware document representation learning for large-scale science mapping
- New
- Research Article
- 10.1016/j.eswa.2025.130853
- Apr 1, 2026
- Expert Systems with Applications
- Yong Lu + 2 more
DFSA-Net: an effective single image dehazing framework via frequency-domain processing and multi-scale representation learning
- New
- Research Article
- 10.1016/j.neunet.2025.108358
- Apr 1, 2026
- Neural networks : the official journal of the International Neural Network Society
- Zhirong Zhang + 4 more
MC-HSTA: A multi-source cross-domain hybrid spatio-temporal attention network for traffic flow prediction.
- New
- Research Article
- 10.1016/j.patcog.2025.112607
- Apr 1, 2026
- Pattern Recognition
- Hengheng Xiong + 7 more
Decoupling representation learning and classifier for long-tailed adversarial training
- New
- Research Article
- 10.1016/j.artmed.2026.103371
- Apr 1, 2026
- Artificial intelligence in medicine
- Woohyeok Choi + 5 more
EEG-based epileptic seizure prediction with patient-tailored spectral-spatial-temporal feature learning.
- New
- Research Article
- 10.1016/j.neunet.2025.108410
- Apr 1, 2026
- Neural networks : the official journal of the International Neural Network Society
- Jinzhong Ning + 7 more
CNER-Omni: A unified dynamic modality learning framework for Chinese named entity recognition across text and speech.
- New
- Research Article
- 10.1016/j.patcog.2025.112518
- Apr 1, 2026
- Pattern Recognition
- Mingyu Zhao + 7 more
GraphProbe: Knowledge Probing for Graph Representation Learning
- New
- Research Article
- 10.1016/j.neunet.2025.108420
- Apr 1, 2026
- Neural networks : the official journal of the International Neural Network Society
- Jiaqi Chu + 4 more
Granger-TSllm: Granger causality enhanced LLMs with residual-quantized tokenizer for multivariate time series forecasting.
- New
- Research Article
- 10.1016/j.knosys.2026.115638
- Apr 1, 2026
- Knowledge-Based Systems
- Dae Hyeon Kim + 1 more
Adaptive semi-supervised graph contrastive representation learning for robust emotion recognition using electroencephalogram
- New
- Research Article
- 10.1016/j.patcog.2025.112602
- Apr 1, 2026
- Pattern Recognition
- Chengchao Shen + 2 more
Multiple object stitching for unsupervised representation learning
- New
- Research Article
- 10.1016/j.knosys.2026.115541
- Apr 1, 2026
- Knowledge-Based Systems
- Yanglan Gan + 4 more
HASNN: Hierarchical attention spiking neural network for dynamic graph representation learning