Step-Wise Distribution-Aligned Style Prompt Tuning for Source-Free Cross-Domain Few-Shot Learning.

  • Abstract
  • Literature Map
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

Existing cross-domain few-shot learning (CDFSL) methods, which develop training strategies in the source domain to enhance model transferability, face challenges when applied to large-scale pre-trained models (LMs), as their source domains and training strategies are not accessible. Besides, fine-tuning LMs specifically for CDFSL requires substantial computational resources, which limits their practicality. Therefore, this paper investigates the source-free CDFSL (SF-CDFSL) problem to solve the few-shot learning (FSL) task in target domain using only a pre-trained model and a few target samples, without requiring source data or training strategies. However, the inaccessibility of source data prevents explicitly reducing the domain gaps between the source and target. To tackle this challenge, this paper proposes a novel approach, Step-wise Distribution-aligned Style Prompt Tuning (StepSPT), to implicitly narrow the domain gaps from the perspective of prediction distribution optimization. StepSPT initially proposes a style prompt that adjusts the target samples to mirror the expected distribution. Furthermore, StepSPT tunes the style prompt and classifier by exploring a dual-phase optimization process (external and internal processes). In the external process, a step-wise distribution alignment strategy is introduced to tune the proposed style prompt by factorizing the prediction distribution optimization problem into the multi-step distribution alignment problem. In the internal process, the classifier is updated via standard cross-entropy loss. Evaluation on 5 datasets illustrates the superiority of StepSPT over existing prompt tuning-based methods and state-of-the-art methods (SOTAs). Furthermore, ablation studies and performance analyzes highlight the efficacy of StepSPT.

Similar Papers
  • Research Article
  • Cite Count Icon 2
  • 10.1016/j.knosys.2024.112024
Learning general features to bridge the cross-domain gaps in few-shot learning
  • Jun 8, 2024
  • Knowledge-Based Systems
  • Xiang Li + 8 more

Learning general features to bridge the cross-domain gaps in few-shot learning

  • Research Article
  • Cite Count Icon 1
  • 10.3389/fpls.2024.1434222
From laboratory to field: cross-domain few-shot learning for crop disease identification in the field.
  • Dec 18, 2024
  • Frontiers in plant science
  • Sen Yang + 5 more

Few-shot learning (FSL) methods have made remarkable progress in the field of plant disease recognition, especially in scenarios with limited available samples. However, current FSL approaches are usually limited to a restrictive setting where base classes and novel classes come from the same domain such as PlantVillage. Consequently, when the model is generalized to new domains (field disease datasets), its performance drops sharply. In this work, we revisit the cross-domain performance of existing FSL methods from both data and model perspectives, aiming to better achieve cross-domain generalization of disease by exploring inter-domain correlations. Specifically, we propose a broader cross-domain few-shot learning(CD-FSL) framework for crop disease identification that allows the classifier to generalize to previously unseen categories and domains. Within this framework, three representative CD-FSL models were developed by integrating the Brownian distance covariance (BCD) module and improving the general feature extractor, namely metric-based CD-FSL(CDFSL-BDC), optimization-based CD-FSL(CDFSL-MAML), and non-meta-learning-based CD-FSL (CDFSL-NML). To capture the impact of domain shift on model performance, six public datasets with inconsistent feature distributions between domains were selected as source domains. We provide a unified testbed to conduct extensive meta-training and meta-testing experiments on the proposed benchmarks to evaluate the generalization performance of CD-FSL in the disease domain. The results showed that the accuracy of the three CD-FSL models improved significantly as the inter-domain similarity increased. Compared with other state-of-the-art CD-FSL models, the CDFSL-BDC models had the best average performance under different domain gaps. Shifting from the pest domain to the crop disease domain, the CDFSL-BDC model achieved an accuracy of 63.95% and 80.13% in the 1-shot/5-shot setting, respectively. Furthermore, extensive evaluation on a multi-domain datasets demonstrated that multi-domain learning exhibits stronger domain transferability compared to single-domain learning when there is a large domain gap between the source and target domain. These comparative results suggest that optimizing the CD-FSL method from a data perspective is highly effective for solving disease identification tasks in field environments. This study holds promise for expanding the application of deep learning techniques in disease detection and provides a technical reference for cross-domain disease detection.

  • Research Article
  • Cite Count Icon 8
  • 10.1016/j.patcog.2021.108293
Who is closer: A computational method for domain gap evaluation
  • Sep 1, 2021
  • Pattern Recognition
  • Xiaobin Liu + 1 more

Who is closer: A computational method for domain gap evaluation

  • Research Article
  • Cite Count Icon 2
  • 10.1016/j.neunet.2024.106536
Spectral Decomposition and Transformation for Cross-domain Few-shot Learning
  • Jul 14, 2024
  • Neural Networks
  • Yicong Liu + 3 more

Spectral Decomposition and Transformation for Cross-domain Few-shot Learning

  • Research Article
  • Cite Count Icon 4
  • 10.1016/j.knosys.2024.111968
A fine-grained self-adapting prompt learning approach for few-shot learning with pre-trained language models
  • Jun 1, 2024
  • Knowledge-Based Systems
  • Xiaojun Chen + 5 more

A fine-grained self-adapting prompt learning approach for few-shot learning with pre-trained language models

  • Research Article
  • Cite Count Icon 1
  • 10.1609/aaai.v37i9.26306
High-Level Semantic Feature Matters Few-Shot Unsupervised Domain Adaptation
  • Jun 26, 2023
  • Proceedings of the AAAI Conference on Artificial Intelligence
  • Lei Yu + 4 more

In few-shot unsupervised domain adaptation (FS-UDA), most existing methods followed the few-shot learning (FSL) methods to leverage the low-level local features (learned from conventional convolutional models, e.g., ResNet) for classification. However, the goal of FS-UDA and FSL are relevant yet distinct, since FS-UDA aims to classify the samples in target domain rather than source domain. We found that the local features are insufficient to FS-UDA, which could introduce noise or bias against classification, and not be used to effectively align the domains. To address the above issues, we aim to refine the local features to be more discriminative and relevant to classification. Thus, we propose a novel task-specific semantic feature learning method (TSECS) for FS-UDA. TSECS learns high-level semantic features for image-to-class similarity measurement. Based on the high-level features, we design a cross-domain self-training strategy to leverage the few labeled samples in source domain to build the classifier in target domain. In addition, we minimize the KL divergence of the high-level feature distributions between source and target domains to shorten the distance of the samples between the two domains. Extensive experiments on DomainNet show that the proposed method significantly outperforms SOTA methods in FS-UDA by a large margin (i.e., ~10%).

  • Research Article
  • 10.1007/s40747-025-01871-3
Adaptive integrated weight unsupervised multi-source domain adaptation without source data
  • Apr 22, 2025
  • Complex & Intelligent Systems
  • Zhirui Wang + 2 more

Unsupervised multi-source domain adaptation methods transfer knowledge learned from multiple labeled source domains to an unlabeled target domain. Existing methods assume that all source domain data can be accessed directly. However, such an assumption is unrealistic and causes data privacy concerns, especially when the source domain labels include personal information. In such a setting, it is prohibited to minimize domain gaps by pairwise calculation of the data from the source and target domains. Therefore, this work addresses the source-free unsupervised multi-source domain adaptation problem, where only the source models are available during the adaptation. We propose trust center sample-based source-free domain adaptation (TSDA) method to solve this problem. The key idea is to leverage the pre-trained models from the source domain and progressively train the target model in a self-learning manner. Because target samples with low entropy measured from the pre-trained source model achieve high accuracy, the trust center samples are selected first using the entropy function. Then pseudo labels are assigned for target samples based on a self-supervised pseudo-labeling strategy. For multiple source domains, corresponding target models are learned based on the assigned pseudo labels. Finally, multiple target models are integrated to predict the label for unlabeled target data. Extensive experiment results on some benchmark datasets and generated adversarial samples demonstrate that our approach outperforms existing UMDA methods, even though some methods can always access source data.

  • Research Article
  • Cite Count Icon 270
  • 10.1016/j.compag.2020.105542
Few-Shot Learning approach for plant disease classification using images taken in the field
  • Jun 20, 2020
  • Computers and Electronics in Agriculture
  • David Argüeso + 6 more

Few-Shot Learning approach for plant disease classification using images taken in the field

  • Research Article
  • 10.1016/j.knosys.2024.112548
Gradient-guided channel masking for cross-domain few-shot learning
  • Oct 9, 2024
  • Knowledge-Based Systems
  • Siqi Hui + 4 more

Gradient-guided channel masking for cross-domain few-shot learning

  • Supplementary Content
  • Cite Count Icon 44
  • 10.3390/s22155507
Deep Unsupervised Domain Adaptation with Time Series Sensor Data: A Survey
  • Jul 23, 2022
  • Sensors (Basel, Switzerland)
  • Yongjie Shi + 2 more

Sensors are devices that output signals for sensing physical phenomena and are widely used in all aspects of our social production activities. The continuous recording of physical parameters allows effective analysis of the operational status of the monitored system and prediction of unknown risks. Thanks to the development of deep learning, the ability to analyze temporal signals collected by sensors has been greatly improved. However, models trained in the source domain do not perform well in the target domain due to the presence of domain gaps. In recent years, many researchers have used deep unsupervised domain adaptation techniques to address the domain gap between signals collected by sensors in different scenarios, i.e., using labeled data in the source domain and unlabeled data in the target domain to improve the performance of models in the target domain. This survey first summarizes the background of recent research on unsupervised domain adaptation with time series sensor data, the types of sensors used, the domain gap between the source and target domains, and commonly used datasets. Then, the paper classifies and compares different unsupervised domain adaptation methods according to the way of adaptation and summarizes different adaptation settings based on the number of source and target domains. Finally, this survey discusses the challenges of the current research and provides an outlook on future work. This survey systematically reviews and summarizes recent research on unsupervised domain adaptation for time series sensor data to provide the reader with a systematic understanding of the field.

  • Research Article
  • Cite Count Icon 22
  • 10.1109/tcsvt.2020.3035890
Fast Adapting Without Forgetting for Face Recognition
  • Nov 4, 2020
  • IEEE Transactions on Circuits and Systems for Video Technology
  • Hao Liu + 4 more

Although face recognition has made dramatic improvements in recent years, there are still many challenges in real-world applications such as face recognition for the elderly and children, for the surveillance scenes and for Near infrared vs. Visible light (NIR-VIS) heterogeneous scene, etc. Due to the existence of these challenges, there are usually domain gaps between training (source domain) and test (target domain). A common way to improve the performance on the target domain is fine-tuning the base model trained on source domain using target data. However, it will severely degrade performance on the source domain. Another way which jointly trains models using both source and target data, suffers from the heavy computations and large data storage, especially when we continue to encounter new domains. In response to these problems, we introduce a new challenging task: Single Exemplar Domain Incremental Learning (SE-DIL), which utilizes the target domain data and just one exemplar per identity from source domain data to quickly improve the performance on the target domain while keeping the performance on the source domain. To deal with SE-DIL, we propose our Fast Adapting without Forgetting (FAwF) method with three components: margin-based exemplar selection, prototype-based class extension and hard&soft knowledge distillation. Through FAwF, we can well maintain the source domain performance with only one sample per source domain class, greatly reducing the fine-tuning time-cost and data storage. Besides, we collected a large-scale children face dataset KidsFace with 12 K identities for studying the SE-DIL in face recognition. Extensive analysis and experiments on our KidsFace-Test protocol and other challenging face test sets show that our method performs better than the state-of-the-art methods on both target and source domain.

  • Research Article
  • Cite Count Icon 8
  • 10.1109/tip.2024.3374222
Enhancing Information Maximization With Distance-Aware Contrastive Learning for Source-Free Cross-Domain Few-Shot Learning.
  • Jan 1, 2024
  • IEEE Transactions on Image Processing
  • Huali Xu + 6 more

Existing Cross-Domain Few-Shot Learning (CDFSL) methods require access to source domain data to train a model in the pre-training phase. However, due to increasing concerns about data privacy and the desire to reduce data transmission and training costs, it is necessary to develop a CDFSL solution without accessing source data. For this reason, this paper explores a Source-Free CDFSL (SF-CDFSL) problem, in which CDFSL is addressed through the use of existing pretrained models instead of training a model with source data, avoiding accessing source data. However, due to the lack of source data, we face two key challenges: effectively tackling CDFSL with limited labeled target samples, and the impossibility of addressing domain disparities by aligning source and target domain distributions. This paper proposes an Enhanced Information Maximization with Distance-Aware Contrastive Learning (IM-DCL) method to address these challenges. Firstly, we introduce the transductive mechanism for learning the query set. Secondly, information maximization (IM) is explored to map target samples into both individual certainty and global diversity predictions, helping the source model better fit the target data distribution. However, IM fails to learn the decision boundary of the target task. This motivates us to introduce a novel approach called Distance-Aware Contrastive Learning (DCL), in which we consider the entire feature set as both positive and negative sets, akin to Schrödinger's concept of a dual state. Instead of a rigid separation between positive and negative sets, we employ a weighted distance calculation among features to establish a soft classification of the positive and negative sets for the entire feature set. We explore three types of negative weights to enhance the performance of CDFSL. Furthermore, we address issues related to IM by incorporating contrastive constraints between object features and their corresponding positive and negative sets. Evaluations of the 4 datasets in the BSCD-FSL benchmark indicate that the proposed IM-DCL, without accessing the source domain, demonstrates superiority over existing methods, especially in the distant domain task. Additionally, the ablation study and performance analysis confirmed the ability of IM-DCL to handle SF-CDFSL. The code will be made public at https://github.com/xuhuali-mxj/IM-DCL.

  • Conference Article
  • Cite Count Icon 5
  • 10.1145/3474085.3475565
Self-Supervised Pre-training on the Target Domain for Cross-Domain Person Re-identification
  • Oct 17, 2021
  • Junyin Zhang + 4 more

Most existing cluster-based cross-domain person re-identification (re-id) methods only pre-train the re-id model on the source domain. Unfortunately, the pre-trained model may not perform well on the target domain due to the large domain gap between source and target domains, which is harmful to the following optimization. In this paper, we propose a novel Self-supervised Pre-training method on the Target Domain (SPTD), which pre-trains the model on both the source and target domains in a self-supervised manner. Specifically, SPTD uses different kinds of data augmentation manners to simulate different intra-class changes and constraints the consistency between the augmented data distribution and the original data distribution. As a result, the pre-trained model involves some specific discriminative knowledge on the target domain and is beneficial to the following optimization. It is easy to combine the proposed SPTD with other cluster-based cross-domain re-id methods just by replacing the original pre-trained model with our pre-trained model. Comprehensive experiments on three widely used datasets, i.e. Market1501, DukeMTMC-ReID and MSMT17, demonstrate the effectiveness of SPTD. Especially, the final results surpass previous state-of-the-art methods by a large margin.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 6
  • 10.3390/jmse12020264
Cross-Domain Contrastive Learning-Based Few-Shot Underwater Acoustic Target Recognition
  • Feb 1, 2024
  • Journal of Marine Science and Engineering
  • Xiaodong Cui + 5 more

Underwater Acoustic Target Recognition (UATR) plays a crucial role in underwater detection devices. However, due to the difficulty and high cost of collecting data in the underwater environment, UATR still faces the problem of small datasets. Few-shot learning (FSL) addresses this challenge through techniques such as Siamese networks and prototypical networks. However, it also suffers from the issue of overfitting, which leads to catastrophic forgetting and performance degradation. Current underwater FSL methods primarily focus on mining similar information within sample pairs, ignoring the unique features of ship radiation noise. This study proposes a novel cross-domain contrastive learning-based few-shot (CDCF) method for UATR to alleviate overfitting issues. This approach leverages self-supervised training on both source and target domains to facilitate rapid adaptation to the target domain. Additionally, a base contrastive module is introduced. Positive and negative sample pairs are generated through data augmentation, and the similarity in the corresponding frequency bands of feature embedding is utilized to learn fine-grained features of ship radiation noise, thereby expanding the scope of knowledge in the source domain. We evaluate the performance of CDCF in diverse scenarios on ShipsEar and DeepShip datasets. The experimental results indicate that in cross-domain environments, the model achieves accuracy rates of 56.71%, 73.02%, and 76.93% for 1-shot, 3-shot, and 5-shot scenarios, respectively, outperforming other FSL methods. Moreover, the model demonstrates outstanding performance in noisy environments.

  • Research Article
  • Cite Count Icon 35
  • 10.1109/lgrs.2019.2956490
Unsupervised Adversarial Domain Adaptation Network for Semantic Segmentation
  • Dec 26, 2019
  • IEEE Geoscience and Remote Sensing Letters
  • Wei Liu + 1 more

With the rapid development of deep learning technology, semantic segmentation methods have been widely used in remote sensing data. A pretrained semantic segmentation model usually cannot perform well when the testing images (target domain) have an obvious difference from the training data set (source domain), while a large enough labeled data set is almost impossible to be acquired for each scenario. Unsupervised domain adaptation (DA) techniques aim to transfer knowledge learned from the source domain to a totally unlabeled target domain. By reducing the domain shift, DA methods have shown the ability to improve the classification accuracy for the target domain. Hence, in this letter, we propose an unsupervised adversarial DA network that converts deep features into 2-D feature curves and reduces the discrepancy between curves from the source domain and curves from the target domain based on a conditional generative adversarial networks (cGANs) model. Our proposed DA network is able to improve the semantic labeling accuracy when we apply a pretrained semantic segmentation model to the target domain. To test the effectiveness of the proposed method, experiments are conducted on the International Society for Photogrammetry and Remote Sensing (ISPRS) 2-D Semantic Labeling data set. Results show that our proposed network is able to stably improve overall accuracy not only when the source and target domains are from the same city but with different building styles but also when the source and target domains are from different cities and acquired by different sensors. By comparing with a few state-of-the-art DA methods, we demonstrate that our proposed method achieves the best cross-domain semantic segmentation performance.

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.