Few-Shot Classification Of Brain Cancer Images Using Meta-Learning Algorithms

  • Abstract
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

The primary objective of deep learning is to have good performance on a large dataset. However, when the model lacks sufficient data, it becomes a challenge to achieve high accuracy in predicting these unfamiliar classes. In fact, the real-world dataset often introduces new classes, and some types of data are difficult to collect or simulate, such as medical images. A subset of machine learning is meta learning, or "learning-to-learn", which can tackle these problems. In this paper, a few-shot classification model is proposed to classify three types of brain cancer: Glioma brain cancer, Meningioma brain cancer, and brain Tumor cancer. To achieve this, we employ an episodic meta-training paradigm that integrates the model-agnostic meta-learning (MAML) framework with a prototypical network (ProtoNet) to train the model. In detail, ProtoNet focuses on learning a metric space by computing distances to class prototypes of each class, while MAML concentrates on finding the optimal initialization parameters for the model to enable the model to learn quickly on a few labeled samples. In addition, we compute and report the average accuracy for the baseline and our methods to assess the quality of the prediction confidence. Simulation results indicate that our proposed approach substantially surpasses the performance of the baseline ResNet18 model, achieving an average accuracy improvement from 46.33% to 92.08% across different few-shot settings. These findings highlight the potential of combining metric-based and optimization-based meta-learning techniques to improve diagnostic support in healthcare applications.

Similar Papers
  • Conference Article
  • Cite Count Icon 20
  • 10.1109/icassp43922.2022.9746899
Improved Meta Learning for Low Resource Speech Recognition
  • May 23, 2022
  • Satwinder Singh + 2 more

We propose a new meta learning based framework for low resource speech recognition that improves the previous model agnostic meta learning (MAML) approach. The MAML is a simple yet powerful meta learning approach. However, the MAML presents some core deficiencies such as training instabilities and slower convergence speed. To address these issues, we adopt multi-step loss (MSL). The MSL aims to calculate losses at every step of the inner loop of MAML and then combines them with a weighted importance vector. The importance vector ensures that the loss at the last step has more importance than the previous steps. Our empirical evaluation shows that MSL significantly improves the stability of the training procedure and it thus also improves the accuracy of the overall system. Our proposed system outperforms MAML based low resource ASR system on various languages in terms of character error rates and stable training behavior.

  • Research Article
  • Cite Count Icon 8
  • 10.1007/s13042-021-01316-6
Model-agnostic multi-stage loss optimization meta learning
  • Apr 26, 2021
  • International Journal of Machine Learning and Cybernetics
  • Xiao Yao + 5 more

Model Agnostic Meta Learning (MAML) has become the most representative meta learning algorithm to solve few-shot learning problems. This paper mainly discusses MAML framework, focusing on the key problem of solving few-shot learning through meta learning. However, MAML is sensitive to the base model for the inner loop, and training instability occur during the training process, resulting in an increase of the training difficulty of the model in the process of training and verification process, causing degradation of model performance. In order to solve these problems, we propose a multi-stage loss optimization meta-learning algorithm. By discussing a learning mechanism for inner and outer loops, it improves the training stability and accelerates the convergence for the model. The generalization ability of MAML has been enhanced.

  • Conference Article
  • Cite Count Icon 10
  • 10.1109/icdm50108.2020.00091
A Primal-Dual Subgradient Approach for Fair Meta Learning
  • Nov 1, 2020
  • Chen Zhao + 3 more

The problem of learning to generalize on unseen classes during the training step, also known as few-shot classification, has attracted considerable attention. Initialization based methods, such as the gradient-based model agnostic meta-learning (MAML) [1], tackle the few-shot learning problem by “learning to fine-tune”. The goal of these approaches is to learn proper model initialization, so that the classifiers for new classes can be learned from a few labeled examples with a small number of gradient update steps. Few shot meta-learning is well-known with its fast-adapted capability and accuracy generalization onto unseen tasks [2]. Learning fairly with unbiased outcomes is another significant hallmark of human intelligence, which is rarely touched in few-shot meta-learning. In this work, we propose a Primal-Dual Fair Meta-learning framework, namely PDFM, which learns to train fair machine learning models using only a few examples based on data from related tasks. The key idea is to learn a good initialization of a fair model's primal and dual parameters so that it can adapt to a new fair learning task via a few gradient update steps. Instead of manually tuning the dual parameters as hyperparameters via a grid search, PDFM optimizes the initialization of the primal and dual parameters jointly for fair meta-learning via a subgradient primal-dual approach. We further instantiate an example of bias controlling using decision boundary covariance (DBC) [3] as the fairness constraint for each task, and demonstrate the versatility of our proposed approach by applying it to classification on a variety of three realworld datasets. Our experiments show substantial improvements over the best prior work for this setting. Our code and datasets are available at https://github.com/charliezhaoyinpeng/PDFM.git.

  • Conference Article
  • Cite Count Icon 1
  • 10.1109/icassp49357.2023.10094936
Meta Learning with Adaptive Loss Weight for Low-Resource Speech Recognition
  • Jun 4, 2023
  • Qiulin Wang + 3 more

Model Agnostic Meta-Learning (MAML) is an effective meta-learning algorithm for low-resource automatic speech recognition (ASR). It uses gradient descent to learn the initialization parameters of the model through various languages, making the model quickly adapt to unseen low-resource languages. But MAML is unstable due to its unique bilevel loss backward structure, which significantly affects the stability and generalization of the model. Since various languages have different contributions to the target language, the loss weights corresponding to the effects of diverse languages require costly manual adjustment in the training stage. Proper selection of these weights will influence the performance of the entire model. In this paper, we propose to apply a loss weight adaption method to MAML using Convolutional Neural Network (CNN) with Homoscedastic Uncertainty. The results of experiments showed that the proposed method outperformed previous gradient-based meta-learning methods and other loss weights adaption methods, and it further improved the stability and effectiveness of MAML.

  • Research Article
  • Cite Count Icon 21
  • 10.1088/2634-4386/ac8828
Meta-learning spiking neural networks with surrogate gradient descent
  • Sep 30, 2022
  • Neuromorphic Computing and Engineering
  • Kenneth M Stewart + 1 more

Adaptive ‘life-long’ learning at the edge and during online task performance is an aspirational goal of artificial intelligence research. Neuromorphic hardware implementing spiking neural networks (SNNs) are particularly attractive in this regard, as their real-time, event-based, local computing paradigm makes them suitable for edge implementations and fast learning. However, the long and iterative learning that characterizes state-of-the-art SNN training is incompatible with the physical nature and real-time operation of neuromorphic hardware. Bi-level learning, such as meta-learning is increasingly used in deep learning to overcome these limitations. In this work, we demonstrate gradient-based meta-learning in SNNs using the surrogate gradient method that approximates the spiking threshold function for gradient estimations. Because surrogate gradients can be made twice differentiable, well-established, and effective second-order gradient meta-learning methods such as model agnostic meta learning (MAML) can be used. We show that SNNs meta-trained using MAML perform comparably to conventional artificial neural networks meta-trained with MAML on event-based meta-datasets. Furthermore, we demonstrate the specific advantages that accrue from meta-learning: fast learning without the requirement of high precision weights or gradients, training-to-learn with quantization and mitigating the effects of approximate synaptic plasticity rules. Our results emphasize how meta-learning techniques can become instrumental for deploying neuromorphic learning technologies on real-world problems.

  • Conference Article
  • Cite Count Icon 3
  • 10.1109/slt48900.2021.9383489
Meta Learning to Classify Intent and Slot Labels with Noisy Few Shot Examples
  • Jan 19, 2021
  • Shang-Wen Li + 4 more

Recently deep learning has dominated many machine learning areas, including spoken language understanding (SLU). However, deep learning models are notorious for being data-hungry, and the heavily optimized models are usually sensitive to the quality of the training examples provided and the consistency between training and inference conditions. To improve the performance of SLU models on tasks with noisy and low training resources, we propose a new SLU benchmarking task: few-shot robust SLU, where SLU comprises two core problems, intent classification (IC) and slot labeling (SL). We establish the task by defining few-shot splits on three public IC/SL datasets, ATIS, SNIPS, and TOP, and adding two types of natural noises (adaptation example missing/replacing and modality mismatch) to the splits. We further propose a novel noise-robust few-shot SLU model based on prototypical networks. We show the model consistently outperforms the conventional fine-tuning baseline and another popular meta-learning method, Model-Agnostic Meta-Learning (MAML), in terms of achieving better IC accuracy and SL F1, and yielding smaller performance variation when noises are present.

  • Video Transcripts
  • 10.48448/0yw3-gw31
Task Similarity Aware Meta Learning: Theory-inspired Improvement on MAML
  • Jul 17, 2021
  • Pan Zhou + 5 more

Few-shot learning ability is heavily desired for machine intelligence. By meta-learning a model initialization from training tasks with fast adaptation ability to new tasks, model-agnostic meta-learning (MAML) has achieved remarkable success in a number of few-shot learning applications. However, theoretical understandings on the learning ability of MAML remain absent yet, hindering developing new and more advanced meta learning methods in a principle way. In this work, we solve this problem by theoretically justifying the fast adaptation capability of MAML when applied to new tasks. Specifically, we prove that the learnt meta-initialization can quickly adapt to new tasks with only a few steps of gradient descent. This result, for the first time, explicitly reveals the benefits of the unique designs in MAML. Then we propose a theory-inspired task similarity aware MAML which clusters tasks into multiple groups according to the estimated optimal model parameters and learns group-specific initializations. The proposed method improves upon MAML by speeding up the adaptation and giving stronger few-shot learning ability. Experimental results on the few-shot classification tasks testify its advantages. TensorFlow code will be released to reproduce our results.

  • Research Article
  • Cite Count Icon 5
  • 10.1016/j.engappai.2023.105978
Parametric unsteady flow modeling by using meta learning
  • Feb 21, 2023
  • Engineering Applications of Artificial Intelligence
  • Xinshuai Zhang + 5 more

Parametric unsteady flow modeling by using meta learning

  • Research Article
  • Cite Count Icon 95
  • 10.1109/tim.2021.3082264
Intelligent Fault Diagnosis of Rotary Machines: Conditional Auxiliary Classifier GAN Coupled With Meta Learning Using Limited Data
  • Jan 1, 2021
  • IEEE Transactions on Instrumentation and Measurement
  • Sonal Dixit + 2 more

The industrial advancement has promoted the development of deep learning (DL)-based intelligent fault diagnosis methods for condition-based maintenance (CBM). Though these methods rely on large dataset for training, the collection of large number of fault samples is not practically feasible. For this purpose, generative adversarial networks (GANs) are capable to generate high-quality synthetic samples. However, the problem still persists with the training of GAN using limited fault samples that are present in practical conditions. This article proposes a novel conditional auxiliary classifier GAN framework coupled with model agnostic meta learning (MAML) to resolve this problem. The objective is to initialize and update the network parameters using MAML instead of regular stochastic gradient learning. This modification enables GAN to learn the task of synthetic sample generation using the limited training dataset. The effectiveness of the proposed framework has been compared with several famous state-of-the-art intelligent fault diagnosis methods existing in the literature. The comparative performance has been validated on benchmarked datasets, i.e., air compressor and bearing datasets collected from a single-stage reciprocating air compressor. The proposed framework is able to achieve the classification accuracy of 99.26% and 98.55% for bearing and air compressor datasets, respectively, with only ten samples per class. Moreover, a real-time case study is performed to validate the proposed method in real time.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 35
  • 10.1109/access.2020.3045225
Meta Learn on Constrained Transfer Learning for Low Resource Cross Subject EEG Classification
  • Jan 1, 2020
  • IEEE Access
  • Tiehang Duan + 6 more

Electroencephalogram (EEG) signal has large variance and its pattern differs significantly across subjects. Cross subject EEG classification is a challenging task due to such pattern variation and the limited target data available, as collecting and annotating EEG data for a new user is costly and involve efforts from human experts. We model the task as a transfer learning problem and propose to tackle it with meta learning on constrained transfer learning (MLCL). MLCL is an end to end trainable learning paradigm that trains on large standard datasets of known subjects and then quickly adapt to a new subject with minimal target data. The transfer process is accelerated by applying model-agnostic meta-learning (MAML) algorithm, performed under a novel constrained setting which keeps enough flexibility to adapt to new subject while significantly reducing number of parameters to transfer. This enables the adaptation done with a small amount of target data. The method can be applied to all deep learning oriented models. We performed extensive experiments across three public datasets. The proposed model outperforms current state of the arts in terms of both accuracy and AUC-ROC score for low target resource configurations. We further conducted interpretation analysis on the model, which reveals detailed information at the resolution of individual channels for the transfer process.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 3
  • 10.3390/ai2020012
Meta Learning for Few-Shot One-Class Classification
  • Apr 22, 2021
  • AI
  • Gabriel Dahia + 1 more

We propose a method that can perform one-class classification given only a small number of examples from the target class and none from the others. We formulate the learning of meaningful features for one-class classification as a meta-learning problem in which the meta-training stage repeatedly simulates one-class classification, using the classification loss of the chosen algorithm to learn a feature representation. To learn these representations, we require only multiclass data from similar tasks. We show how the Support Vector Data Description method can be used with our method, and also propose a simpler variant based on Prototypical Networks that obtains comparable performance, indicating that learning feature representations directly from data may be more important than which one-class algorithm we choose. We validate our approach by adapting few-shot classification datasets to the few-shot one-class classification scenario, obtaining similar results to the state-of-the-art of traditional one-class classification, and that improves upon that of one-class classification baselines employed in the few-shot setting.

  • Video Transcripts
  • 10.48448/1h13-vc36
Minimax and Neyman–Pearson Meta-Learning for Outlier Languages
  • Aug 3, 2021
  • Rahul Aralikatte + 4 more

Model-agnostic meta-learning (MAML) has been recently put forth as a strategy to learn resource-poor languages in a sample-efficient fashion. Nevertheless, the properties of these languages are often not well represented by those available during training. Hence, we argue that the i.i.d. assumption ingrained in MAML makes it ill-suited for cross-lingual NLP. In fact, under a decision-theoretic framework, MAML can be interpreted as minimising the expected risk across training languages (with a uniform prior), which is known as Bayes criterion. To increase its robustness to outlier languages, we create two variants of MAML based on alternative criteria: Minimax MAML reduces the maximum risk across languages, while Neyman–Pearson MAML constrains the risk in each language to a maximum threshold. Both criteria constitute fully differentiable two-player games. In light of this, we propose a new adaptive optimiser solving for a local approximation to their Nash equilibrium. We evaluate both model variants on two popular NLP tasks, part-of-speech tagging and question answering. We report gains for their average and minimum performance across low-resource languages in zero- and few-shot settings, compared to joint multi-source transfer and vanilla MAML. The code for our experiments is available at https://github.com/rahular/robust-maml

  • PDF Download Icon
  • Conference Article
  • Cite Count Icon 1
  • 10.18653/v1/2021.findings-acl.106
Minimax and Neyman–Pearson Meta-Learning for Outlier Languages
  • Jan 1, 2021
  • Edoardo Maria Ponti + 4 more

Model-agnostic meta-learning (MAML) has been recently put forth as a strategy to learn resource-poor languages in a sample-efficient fashion. Nevertheless, the properties of these languages are often not well represented by those available during training. Hence, we argue that the i.i.d. assumption ingrained in MAML makes it ill-suited for cross-lingual NLP. In fact, under a decision-theoretic framework, MAML can be interpreted as minimising the expected risk across training languages (with a uniform prior), which is known as Bayes criterion. To increase its robustness to outlier languages, we create two variants of MAML based on alternative criteria: Minimax MAML reduces the maximum risk across languages, while Neyman-Pearson MAML constrains the risk in each language to a maximum threshold. Both criteria constitute fully differentiable two-player games. In light of this, we propose a new adaptive optimiser solving for a local approximation to their Nash equilibrium. We evaluate both model variants on two popular NLP tasks, part-of-speech tagging and question answering. We report gains for their average and minimum performance across low-resource languages in zero- and few-shot settings, compared to joint multi-source transfer and vanilla MAML.

  • Conference Article
  • Cite Count Icon 1
  • 10.1109/wcnc51071.2022.9771840
Evaluation of Adaptation Methods for Deep Learning-based Wi-Fi Receivers
  • Apr 10, 2022
  • William Blount + 4 more

Machine-learning based transceivers have received increasing attention for next-generation wireless systems. We investigate the application of two meta-learning algorithms – Model Agnostic Meta Learning (MAML) and Reptile – to a deep learning-based Wi-Fi channel estimation and tracking system, called DeepWiPHY. The meta-learning algorithms were compared against conventional methods such as random initialization, cross-evaluation, and retraining on multiple channel models with varying severity of multipath fading. Comparisons were made fairly with respect to the complexity of the adaptation of the model necessary for a new environment. The results indicate that perhaps surprisingly, conventional training methods are adequate and in fact can outperform meta-learning methods over a wide variety of channels. The key is to train the receiver using the worst-case (most severe) multipath channel model, which then allows strong performance across a wide class of channels without requiring the additional burden of meta-learning.

  • Research Article
  • 10.7717/peerj-cs.3161
Quantification of left ventricular mass in multiple views of echocardiograms using model-agnostic meta learning in a few-shot setting
  • Sep 16, 2025
  • PeerJ Computer Science
  • Yeong Hyeon Kim + 7 more

BackgroundReliable measurement of left ventricular mass (LVM) in echocardiography is essential for early detection of left ventricular dysfunction, coronary artery disease, and arrhythmia risk, yet growing patient volumes have created critical shortage of experts in echocardiography. Recent deep learning approaches reduce inter‐operator variability but require large, fully labeled datasets for each standard view—an impractical demand in many clinical settings.MethodsTo overcome these limitations, we propose a heatmap-based point-estimation segmentation model trained via model-agnostic meta-learning (MAML) for few-shot LVM quantification across multiple echocardiographic views. Our framework adapts rapidly to new views by learning a shared representation and view-specific head performing K inner-loop updates, and then meta-updating in the outer loop. We used the EchoNet-LVH dataset for the PLAX view, the TMED-2 dataset for the PSAX view and the CAMUS dataset for both the apical 2-chamber and apical 4-chamber views under 1-, 5-, and 10-shot scenarios.ResultsAs a result, the proposed MAML methods demonstrated comparable performance using mean distance error, mean angle error, successful distance error and spatial angular similarity in a few-shot setting compared to models trained with larger labeled datasets for each view of the echocardiogram.

More from: EAI Endorsed Transactions on Industrial Networks and Intelligent Systems
  • Research Article
  • 10.4108/eetinis.131.9499
COSMN: Clustering-Based Optimization for 360-Degree Live Streaming over Mobile Networks
  • Nov 11, 2025
  • EAI Endorsed Transactions on Industrial Networks and Intelligent Systems
  • Hung Nguyen Viet + 3 more

  • Research Article
  • 10.4108/eetinis.124.10405
Few-Shot Classification Of Brain Cancer Images Using Meta-Learning Algorithms
  • Nov 6, 2025
  • EAI Endorsed Transactions on Industrial Networks and Intelligent Systems
  • Tuyet-Nhi Thi Nguyen + 4 more

  • Research Article
  • 10.4108/eetinis.124.10134
Optimizing UAV Trajectories in Optical IRS-Aided Hybrid FSO/RF Aerial Access Networks Using DRL Technique
  • Nov 6, 2025
  • EAI Endorsed Transactions on Industrial Networks and Intelligent Systems
  • Cuong Nguyen + 4 more

  • Research Article
  • 10.4108/eetinis.v12i4.10461
Deep Reinforcement Learning Approaches Against Jammers with Unequal Sweeping Probability Attacks
  • Nov 4, 2025
  • EAI Endorsed Transactions on Industrial Networks and Intelligent Systems
  • Lan Nguyen + 3 more

  • Research Article
  • 10.4108/eetinis.v12i3.8720
FedNDA: Enhancing Federated Learning with Noisy Client Detection and Robust Aggregation
  • Jul 3, 2025
  • EAI Endorsed Transactions on Industrial Networks and Intelligent Systems
  • Tuan Dung Kieu + 6 more

  • Research Article
  • 10.4108/eetinis.v12i3.8948
Return Loss Optimization in Rectangular Microstrip Patch Antennas Using Response Surface Methodology (RSM) for 5G Applications
  • Jun 12, 2025
  • EAI Endorsed Transactions on Industrial Networks and Intelligent Systems
  • Thi Bich Ngoc Tran + 1 more

  • Open Access Icon
  • Research Article
  • 10.4108/eetinis.v12i3.8587
QoE-Energy Consumption Optimization for End-User Devices in Adaptive Bitrate Video Streaming Using the Lagrange Multiplier Method
  • Apr 14, 2025
  • EAI Endorsed Transactions on Industrial Networks and Intelligent Systems
  • Tien Vu Huu + 1 more

  • Research Article
  • 10.4108/eetinis.v12i3.8604
Security-Reliability Analysis of NOMA-Assisted Hybrid Satellite-Terrestrial Relay Multi-Cast Transmission Networks Using Fountain Codes and Partial Relay Selection with Presence of Multiple Eavesdroppers
  • Apr 11, 2025
  • EAI Endorsed Transactions on Industrial Networks and Intelligent Systems
  • Toan Van Nguyen + 4 more

  • Research Article
  • 10.4108/eetinis.v12i2.8600
Integrated Cloud-Twin Synchronization for Supply Chain 5.0
  • Mar 12, 2025
  • EAI Endorsed Transactions on Industrial Networks and Intelligent Systems
  • Divya Sasi Latha + 1 more

  • Research Article
  • 10.4108/eetinis.v12i2.7612
Single-level Discrete Two Dimensional Wavelet Transform Based Multiscale Deep Learning Framework for Two-Wheeler Helmet Detection
  • Mar 11, 2025
  • EAI Endorsed Transactions on Industrial Networks and Intelligent Systems
  • Amrutha Annadurai + 3 more

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.

Search IconWhat is the difference between bacteria and viruses?
Open In New Tab Icon
Search IconWhat is the function of the immune system?
Open In New Tab Icon
Search IconCan diabetes be passed down from one generation to the next?
Open In New Tab Icon