Adapting Generative Large Language Models for Information Extraction from Unstructured Electronic Health Records in Residential Aged Care: A Comparative Analysis of Training Approaches

  • Abstract
  • Highlights & Summary
  • Literature Map
  • References
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

Information extraction (IE) of unstructured electronic health records is challenging due to the semantic complexity of textual data. Generative large language models (LLMs) offer promising solutions to address this challenge. However, identifying the best training methods to adapt LLMs for IE in residential aged care settings remains underexplored. This research addresses this challenge by evaluating the effects of zero-shot and few-shot learning, both with and without parameter-efficient fine-tuning (PEFT) and retrieval-augmented generation (RAG) using Llama 3.1-8B. The study performed named entity recognition (NER) to nursing notes from Australian aged care facilities (RACFs), focusing on agitation in dementia and malnutrition risk factors. Performance evaluation includes accuracy, macro-averaged precision, recall, and F1 score. We used non-parametric statistical methods to compare if the differences were statistically significant. Results show that zero-shot and few-shot learning, whether combined with PEFT or RAG, achieve comparable performance across the clinical domains when the same prompting template is used. Few-shot learning significantly outperforms zero-shot learning when neither PEFT nor RAG is applied. Notably, PEFT significantly improves model performance in both zero-shot and few-shot learning; however, RAG significantly improves performance only in few-shot learning. After PEFT, the performance of zero-shot learning reaches a comparable level with few-shot learning. However, few-shot learning with RAG significantly outperforms zero-shot learning with RAG. We also found a similar level of performance between few-shot learning with RAG and zero-shot learning with PEFT. These findings provide valuable insights for researchers, practitioners, and stakeholders to optimize the use of generative LLMs in clinical IE.

Highlights

  • The rapid digitization of healthcare has led to the widespread adoption of electronic health records (EHRs), which store vast amounts of patient data

  • No statistically significant difference is found in accuracy, precision, recall, and F1 score between the zero-shot and fewshot learning with parameter-efficient fine-tuning (PEFT) in named entity recognition (NER) (Fig. 3d, p > 0.05), there is a trend that few-shot learning performs above zero-shot learning

  • No significant difference is found in accuracy, precision, recall, and F1 score for the other clinical domains between the two training models, pure few-shot learning and few-shot learning with PEFT for NER

Read more Highlights Expand/Collapse icon

Summary

IntroductionExpand/Collapse icon

The rapid digitization of healthcare has led to the widespread adoption of electronic health records (EHRs), which store vast amounts of patient data. Electronic health records include structured data, such as demographic information and laboratory results, and unstructured data, such as clinical notes. Structured data, characterized by its predefined format, can be queried and utilized. Unstructured data, often found in free-text formats, encompass a significant portion of comprehensive and rich patient data [1]. These narratives, written by healthcare professionals, provide detailed insights into patient conditions, diagnoses, treatments, and outcomes. The large volume and complex nature of unstructured EHRs have posed significant challenges in effectively utilizing this valuable information [2].

MethodsExpand/Collapse icon
ResultsExpand/Collapse icon
DiscussionExpand/Collapse icon
ConclusionExpand/Collapse icon
ReferencesShowing 10 of 46 papers
  • Open Access Icon
  • Cite Count Icon 8
  • 10.1016/j.artmed.2024.102822
Automatic quantitative stroke severity assessment based on Chinese clinical named entity recognition with domain-adaptive pre-trained large language model
  • Feb 27, 2024
  • Artificial Intelligence In Medicine
  • Zhanzhong Gu + 12 more

  • Cite Count Icon 30
  • 10.1111/j.1365-2702.2008.02670.x
Aged‐care nurses’ knowledge of nursing documentation: an Australian perspective
  • Jun 5, 2009
  • Journal of Clinical Nursing
  • Robyn Daskein + 2 more

  • Cite Count Icon 2
  • 10.1016/j.jbi.2024.104596
Clinical natural language processing for secondary uses
  • Jan 24, 2024
  • Journal of biomedical informatics
  • Yanjun Gao + 3 more

  • Cite Count Icon 17
  • 10.18653/v1/2023.bionlp-1.37
Zero-Shot Information Extraction for Clinical Meta-Analysis using Large Language Models
  • Jan 1, 2023
  • David Kartchner + 4 more

  • Open Access Icon
  • PDF Download Icon
  • Cite Count Icon 107
  • 10.18653/v1/2022.naacl-main.191
Learning To Retrieve Prompts for In-Context Learning
  • Jan 1, 2022
  • Ohad Rubin + 2 more

  • Cite Count Icon 2
  • 10.3233/shti231055
Extracting Symptoms of Agitation in Dementia from Free-Text Nursing Notes Using Advanced Natural Language Processing.
  • Jan 25, 2024
  • Studies in health technology and informatics
  • Dinithi Vithanage + 5 more

  • Cite Count Icon 3
  • 10.3233/shti230937
A Five-Step Workflow to Manually Annotate Unstructured Data into Training Dataset for Natural Language Processing.
  • Jan 25, 2024
  • Studies in health technology and informatics
  • Yunshu Zhu + 4 more

  • Open Access Icon
  • Cite Count Icon 5
  • 10.1145/3664190.3672514
Retrieval Augmented Zero-Shot Text Classification
  • Aug 2, 2024
  • Tassallah Abdullahi + 2 more

  • Open Access Icon
  • Cite Count Icon 19
  • 10.14428/esann/2024.es2024-222
Fine-Tuning Llama 2 Large Language Models for Detecting Online Sexual Predatory Chats and Abusive Texts
  • Jan 1, 2024
  • Thanh Thi Nguyen + 2 more

  • Open Access Icon
  • Cite Count Icon 2
  • 10.3390/psychiatryint5040046
COVID-19 and Its Influence on Prevalence of Dementia and Agitation in Australian Residential Aged Care: A Comparative Study
  • Sep 30, 2024
  • Psychiatry International
  • Yunshu Zhu + 8 more

Similar Papers
  • Research Article
  • 10.1007/s00117-025-01416-2
Optimized interaction with Large Language Models : A practical guide to Prompt Engineering and Retrieval-Augmented Generation
  • Feb 21, 2025
  • Radiologie (Heidelberg, Germany)
  • Anna Fink + 4 more

Given the increasing number of radiological examinations, large language models (LLMs) offer promising support in radiology. Optimized interaction is essential to ensure reliable results. This article provides an overview of interaction techniques such as prompt engineering, zero-shot learning, and retrieval-augmented generation (RAG) and gives practical tips for their application in radiology. Demonstration of interaction techniques based on practical examples with concrete recommendations for their application in routine radiological practice. Advanced interaction techniques allow task-specific adaptation of LLMs without the need for retraining. The creation of precise prompts and the use of zero-shot and few-shot learning can significantly improve response quality. RAG enables the integration of current and domain-specific information into LLM tools, increasing the accuracy and relevance of the generated content. The use of prompt engineering, zero-shot and few-shot learning, and RAG can optimize interaction with LLMs in radiology. Through these targeted strategies, radiologists can efficiently integrate general chatbots into routine practice to improve patient care.

  • Research Article
  • 10.1007/s42452-025-07225-5
A review on NLP zero-shot and few-shot learning: methods and applications
  • Aug 21, 2025
  • Discover Applied Sciences
  • G Ramesh + 6 more

Zero-shot and few-shot learning techniques in natural language processing (NLP), this comprehensive review traces their evolution from traditional methods to cutting-edge approaches like transfer learning and pre-trained language models, semantic embedding, attribute-based approaches, generative models for data augmentation in zero-shot learning, and meta-learning, model-agnostic meta-learning, relationship networks, model-agnostic meta-learning (MAML), prototypical networks in few-shot learning. Real-world applications underscore the adaptability and efficacy of these techniques across various NLP tasks in both industry and academia. Acknowledging challenges inherent in zero-shot and few-shot learning, this review identifies limitations and suggests avenues for improvement. It emphasizes theoretical foundations alongside practical considerations such as accuracy and generalization across diverse NLP tasks. By consolidating key insights, this review provides researchers and practitioners with valuable guidance on the current state and future potential of zero-shot and few-shot learning techniques in addressing real-world NLP challenges. Looking ahead, this review aims to stimulate further research, fostering a deeper understanding of the complexities and applicability of zero-shot and few-shot learning techniques in NLP. By offering a roadmap for future exploration, it seeks to contribute to the ongoing advancement and practical implementation of NLP technologies across various domains.

  • Research Article
  • Cite Count Icon 18
  • 10.1055/a-2264-5631
Improving the use of LLMs in radiology through prompt engineering: from precision prompts to zero-shot learning.
  • Feb 26, 2024
  • RoFo : Fortschritte auf dem Gebiete der Rontgenstrahlen und der Nuklearmedizin
  • Maximilian Frederik Russe + 3 more

Large language models (LLMs) such as ChatGPT have shown significant potential in radiology. Their effectiveness often depends on prompt engineering, which optimizes the interaction with the chatbot for accurate results. Here, we highlight the critical role of prompt engineering in tailoring the LLMs' responses to specific medical tasks. Using a clinical case, we elucidate different prompting strategies to adapt the LLM ChatGPT using GPT4 to new tasks without additional training of the base model. These approaches range from precision prompts to advanced in-context methods such as few-shot and zero-shot learning. Additionally, the significance of embeddings, which serve as a data representation technique, is discussed. Prompt engineering substantially improved and focused the chatbot's output. Moreover, embedding of specialized knowledge allows for more transparent insight into the model's decision-making and thus enhances trust. Despite certain challenges, prompt engineering plays a pivotal role in harnessing the potential of LLMs for specialized tasks in the medical domain, particularly radiology. As LLMs continue to evolve, techniques like few-shot learning, zero-shot learning, and embedding-based retrieval mechanisms will become indispensable in delivering tailored outputs. · Large language models might impact radiological practice and decision-masking.. · However, implementation and performance are dependent on the assigned task.. · Optimization of prompting strategies can substantially improve model performance.. · Strategies for prompt engineering range from precision prompts to zero-shot learning.. · Russe MF, Reisert M, Bamberg F et al. Improving the use of LLMs in radiology through prompt engineering: from precision prompts to zero-shot learning . Fortschr Röntgenstr 2024; 196: 1166 - 1170.

  • Conference Article
  • Cite Count Icon 3
  • 10.2118/217671-ms
Enhancing Information Retrieval in the Drilling Domain: Zero-Shot Learning with Large Language Models for Question-Answering
  • Feb 27, 2024
  • S Alyaev + 3 more

Finding information across multiple databases, formats, and documents remains a manual job in the drilling industry. Large Language Models (LLMs) have proven effective in data-aggregation tasks, including answering questions. However, using LLMs for domain-specific factual responses poses a nontrivial challenge. The expert labor cost for training domain-specific LLMs prohibits niche industries from developing custom question-answering bots. This paper tests several commercial LLMs for information retrieval tasks for drilling data using zero-shot in-context learning. In addition, we studied the model’s calibration using a few-shot multiple-choice drilling questionnaire. To create an LLM benchmark for drilling, we collated the text data from publicly available databases: the Norwegian Petroleum Directorate (NPD), company annual reports, and petroleum glossary. We used a zero-shot learning technique that relies on an LLM’s ability to generate responses for tasks outside its training. We implemented a controlled zero-shot learning "in-context" procedure that sends a user’s query augmented with text data to the LLM as inputs. This implementation encourages the LLM to take the answer from the data while leveraging its pre-trained contextual-learning capability. We evaluated several state-of-the-art generic LLMs available through an API, including G4, G3.5-TI, J2-ultra model, and L2 series. The paper documents the pre-trained LLMs’ ability to provide correct answers and identify petroleum industry jargon from the collated dataset. Our zero-shot in-context learning implementation helps vanilla LLMs provide relevant factual responses for the drilling domain. While each LLM’s performance varies, we have identified models suitable for a drilling chatbot application. In particular, G4 outperformed on all the tasks. This finding suggests that training expensive domain-specific LLMs is not necessary for question-answering tasks in the context of drilling data. We demonstrate the utility of zero-shot in-context learning using pre-trained LLMs for question-answering tasks relevant to the drilling industry. Additionally, we prepared and publicly released the collated datasets from the NPD database and companies’ annual reports to enable results reproducibility and to foster acceleration of language model adoption and development for the subsurface and drilling industries. The petroleum industry may find our solution beneficial for enhancing personnel training and career development. It also offers a method for conducting data analytics and overcoming challenges in retrieving historical well data.

  • Research Article
  • 10.55544/ijrah.5.1.24
Zero-Shot Learning and Few-Shot Learning with Generative AI: Bridging the Data Gap for Real-World Applications
  • Jan 30, 2025
  • Integrated Journal for Research in Arts and Humanities
  • Vinay Kumar Gali + 1 more

Modern artificial intelligence systems frequently rely on vast amounts of labeled data to achieve robust performance, yet many real-world scenarios suffer from limited data availability. This paper investigates the potential of integrating zero-shot and few-shot learning paradigms with generative AI models to bridge the persistent data gap. Zero-shot learning empowers models to recognize and classify instances from unseen categories by leveraging semantic descriptors, while few-shot learning focuses on adapting models to new classes using only a handful of examples. Generative AI techniques, such as advanced generative adversarial networks and transformer-based models, can synthesize realistic data samples that mimic complex distributions found in natural environments. By combining these approaches, our methodology offers a dual advantage: it not only enhances model generalization across diverse tasks but also mitigates the challenges posed by data scarcity. We demonstrate the effectiveness of this hybrid framework through experiments in domains including computer vision, natural language processing, and anomaly detection, where traditional data collection is prohibitive. Our analysis reveals that the strategic use of generated data significantly boosts learning outcomes, even when initial training samples are sparse. Furthermore, the adaptability of the proposed system makes it suitable for dynamic, real-world applications where new categories continuously emerge. Overall, this study provides a comprehensive overview of leveraging generative AI to enhance zero-shot and few-shot learning, paving the way for more resilient and scalable solutions in environments constrained by limited data resources. These innovations promise to reshape the future of machine learning by opening new pathways for robust AI development.

  • Research Article
  • Cite Count Icon 177
  • 10.1109/tip.2018.2861573
A Unified approach for Conventional Zero-shot, Generalized Zero-shot and Few-shot Learning.
  • Oct 26, 2017
  • IEEE Transactions on Image Processing
  • Shafin Rahman + 2 more

Prevalent techniques in zero-shot learning do not generalize well to other related problem scenarios. Here, we present a unified approach for conventional zero-shot, generalized zero-shot and few-shot learning problems. Our approach is based on a novel Class Adapting Principal Directions (CAPD) concept that allows multiple embeddings of image features into a semantic space. Given an image, our method produces one principal direction for each seen class. Then, it learns how to combine these directions to obtain the principal direction for each unseen class such that the CAPD of the test image is aligned with the semantic embedding of the true class, and opposite to the other classes. This allows efficient and class-adaptive information transfer from seen to unseen classes. In addition, we propose an automatic process for selection of the most useful seen classes for each unseen class to achieve robustness in zero-shot learning. Our method can update the unseen CAPD taking the advantages of few unseen images to work in a few-shot learning scenario. Furthermore, our method can generalize the seen CAPDs by estimating seen-unseen diversity that significantly improves the performance of generalized zero-shot learning. Our extensive evaluations demonstrate that the proposed approach consistently achieves superior performance in zero-shot, generalized zero-shot and few/one-shot learning problems.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 5
  • 10.2196/56243
Extraction of Substance Use Information From Clinical Notes: Generative Pretrained Transformer–Based Investigation
  • Aug 19, 2024
  • JMIR Medical Informatics
  • Fatemeh Shah-Mohammadi + 1 more

BackgroundUnderstanding the multifaceted nature of health outcomes requires a comprehensive examination of the social, economic, and environmental determinants that shape individual well-being. Among these determinants, behavioral factors play a crucial role, particularly the consumption patterns of psychoactive substances, which have important implications on public health. The Global Burden of Disease Study shows a growing impact in disability-adjusted life years due to substance use. The successful identification of patients’ substance use information equips clinical care teams to address substance-related issues more effectively, enabling targeted support and ultimately improving patient outcomes.ObjectiveTraditional natural language processing methods face limitations in accurately parsing diverse clinical language associated with substance use. Large language models offer promise in overcoming these challenges by adapting to diverse language patterns. This study investigates the application of the generative pretrained transformer (GPT) model in specific GPT-3.5 for extracting tobacco, alcohol, and substance use information from patient discharge summaries in zero-shot and few-shot learning settings. This study contributes to the evolving landscape of health care informatics by showcasing the potential of advanced language models in extracting nuanced information critical for enhancing patient care.MethodsThe main data source for analysis in this paper is Medical Information Mart for Intensive Care III data set. Among all notes in this data set, we focused on discharge summaries. Prompt engineering was undertaken, involving an iterative exploration of diverse prompts. Leveraging carefully curated examples and refined prompts, we investigate the model’s proficiency through zero-shot as well as few-shot prompting strategies.ResultsResults show GPT’s varying effectiveness in identifying mentions of tobacco, alcohol, and substance use across learning scenarios. Zero-shot learning showed high accuracy in identifying substance use, whereas few-shot learning reduced accuracy but improved in identifying substance use status, enhancing recall and F1-score at the expense of lower precision.ConclusionsExcellence of zero-shot learning in precisely extracting text span mentioning substance use demonstrates its effectiveness in situations in which comprehensive recall is important. Conversely, few-shot learning offers advantages when accurately determining the status of substance use is the primary focus, even if it involves a trade-off in precision. The results contribute to enhancement of early detection and intervention strategies, tailor treatment plans with greater precision, and ultimately, contribute to a holistic understanding of patient health profiles. By integrating these artificial intelligence–driven methods into electronic health record systems, clinicians can gain immediate, comprehensive insights into substance use that results in shaping interventions that are not only timely but also more personalized and effective.

  • Book Chapter
  • Cite Count Icon 4
  • 10.4018/979-8-3693-1822-5.ch007
Challenges and Limitations of Few-Shot and Zero-Shot Learning
  • Apr 5, 2024
  • V Dankan Gowda + 4 more

Essential to the development of AI and machine learning, this chapter explores the complex areas of few-shot and zero-shot learning. There have been great advancements towards more efficient and adaptive AI systems with few-shot learning and zero-shot learning, respectively, which can learn from minimal data and infer from particular data instances without previous exposure. Nevertheless, there are several limits and difficulties associated with these procedures. This chapter delves deeply into the theoretical foundations of both techniques, explaining how they work and what problems they solve in different ways. It examines the semantic gap, domain adaptation problems, and model bias, as well as the computational restrictions, overfitting, and model generalizability that are intrinsic to few-shot learning and zero-shot learning, respectively. We may better understand the ideas' potential use in different real-world contexts by comparing and contrasting them.

  • Research Article
  • Cite Count Icon 2
  • 10.3390/ai6040072
Multimodal Data Fusion for Tabular and Textual Data: Zero-Shot, Few-Shot, and Fine-Tuning of Generative Pre-Trained Transformer Models
  • Apr 7, 2025
  • AI
  • Shadi Jaradat + 5 more

In traffic safety analysis, previous research has often focused on tabular data or textual crash narratives in isolation, neglecting the potential benefits of a hybrid multimodal approach. This study introduces the Multimodal Data Fusion (MDF) framework, which fuses tabular data with textual narratives by leveraging advanced Large Language Models (LLMs), such as GPT-2, GPT-3.5, and GPT-4.5, using zero-shot (ZS), few-shot (FS), and fine-tuning (FT) learning strategies. We employed few-shot learning with GPT-4.5 to generate new labels for traffic crash analysis, such as driver fault, driver actions, and crash factors, alongside the existing label for severity. Our methodology was tested on crash data from the Missouri State Highway Patrol, demonstrating significant improvements in model performance. GPT-2 (fine-tuned) was used as the baseline model, against which more advanced models were evaluated. GPT-4.5 few-shot learning achieved 98.9% accuracy for crash severity prediction and 98.1% accuracy for driver fault classification. In crash factor extraction, GPT-4.5 few-shot achieved the highest Jaccard score (82.9%), surpassing GPT-3.5 and fine-tuned GPT-2 models. Similarly, in driver actions extraction, GPT-4.5 few-shot attained a Jaccard score of 73.1%, while fine-tuned GPT-2 closely followed with 72.2%, demonstrating that task-specific fine-tuning can achieve performance close to state-of-the-art models when adapted to domain-specific data. These findings highlight the superior performance of GPT-4.5 few-shot learning, particularly in classification and information extraction tasks, while also underscoring the effectiveness of fine-tuning on domain-specific datasets to bridge performance gaps with more advanced models. The MDF framework’s success demonstrates its potential for broader applications beyond traffic crash analysis, particularly in domains where labeled data are scarce and predictive modeling is essential.

  • Research Article
  • Cite Count Icon 8
  • 10.1016/j.cogsys.2023.101188
Improving deep learning with prior knowledge and cognitive models: A survey on enhancing explainability, adversarial robustness and zero-shot learning
  • Nov 30, 2023
  • Cognitive Systems Research
  • Fuseini Mumuni + 1 more

Improving deep learning with prior knowledge and cognitive models: A survey on enhancing explainability, adversarial robustness and zero-shot learning

  • Research Article
  • 10.1016/j.compbiomed.2025.111013
A comprehensive evaluation of large language models for information extraction from unstructured electronic health records in residential aged care.
  • Oct 1, 2025
  • Computers in biology and medicine
  • Dinithi Vithanage + 5 more

A comprehensive evaluation of large language models for information extraction from unstructured electronic health records in residential aged care.

  • Research Article
  • 10.1609/aaai.v39i11.33286
STD-PLM: Understanding Both Spatial and Temporal Properties of Spatial-Temporal Data with PLM
  • Apr 11, 2025
  • Proceedings of the AAAI Conference on Artificial Intelligence
  • Yiheng Huang + 7 more

Spatial-temporal forecasting and imputation are important for real-world intelligent systems. Most existing methods are tailored for individual forecasting or imputation tasks but are not designed for both. Additionally, they are less effective for zero-shot and few-shot learning. While pre-trained language model (PLM) have exhibited strong pattern recognition and reasoning abilities across various tasks, including few-shot and zero-shot learning, their applications in spatial-temporal data understanding has been constrained by insufficient modeling of complex correlations such as the temporal correlations, spatial connectivity, non-pairwise and high-order spatial-temporal correlations within data. In this paper, we propose STD-PLM for understanding both spatial and temporal properties of Spatial-Temporal Data with PLM, which is capable of implementing both spatial-temporal forecasting and imputation tasks. STD-PLM understands spatial-temporal correlations via explicitly designed spatial and temporal tokenizers. Topology-aware node embeddings are designed for PLM to comprehend and exploit the topology structure of data in inductive manner. Furthermore, to mitigate the efficiency issues introduced by the PLM, we design a sandglass attention module(SGA) combined with a specific constrained loss function, which significantly improves the model's efficiency while ensuring performance. Extensive experiments demonstrate that STD-PLM exhibits competitive performance and generalization capabilities across the forecasting and imputation tasks on various datasets. Moreover, STD-PLM achieves promising results on both few-shot and zero-shot tasks.

  • Research Article
  • 10.2118/0125-0092-jpt
Zero-Shot Learning With Large Language Models Enhances Drilling-Information Retrieval
  • Jan 1, 2025
  • Journal of Petroleum Technology
  • Chris Carpenter

_ This article, written by JPT Technology Editor Chris Carpenter, contains highlights of paper SPE 217671, “Enhancing Information Retrieval in the Drilling Domain: Zero-Shot Learning With Large Language Models for Question Answering,” by Felix J. Pacis, SPE, University of Stavanger, and Sergey Alyaev and Gilles Pelfrene, SPE, NORCE, et al. The paper has not been peer reviewed. _ Finding information across multiple databases, formats, and documents remains a manual job in the drilling industry. Large language models (LLMs) have proven effective in data-aggregation tasks, including answering questions. However, using LLMs for domain-specific factual responses poses a nontrivial challenge. The expert-labor cost for training domain-specific LLMs prohibits niche industries from developing custom question-answering bots. The complete paper tests several commercial LLMs for information-retrieval tasks for drilling data using zero-shot in-context learning. In addition, the model’s calibration is tested with a few-shot multiple-choice drilling questionnaire. Introduction While LLMs have proven effective in various tasks ranging from sentiment analysis to text completion, using LLMs for question-answering tasks presents a challenge in providing factual responses. Pretrained LLMs only serve as a parameterized implicit knowledge base and cannot access recent data; thus, information is bounded by the time of training. Retrieval augmented generation (RAG) can address some of these issues by extending the utility of LLMs to specific data sources. Fig. 1 shows a simplified RAG-based LLM question/answer application. RAG involves two primary components: document retrieval (green boxes), which retrieves the most relevant context based on the query, and LLM response generation (blue boxes). During the response generation, LLM operates based on the prompt, query, and retrieved context without any change in the model parameters, a process the authors term as “in-context learning.” Methodology Two experiments have been conducted: The first one is a few-shot multiple-choice experiment evaluated using the SLB drilling glossary; the second is a zero-shot in-context experiment evaluated on drilling reports and company reports. Multiple-Choice Experiment. SLB Drilling Glossary. For the multiple-choice experiment, a publicly available drilling glossary served as a basis for evaluation. A total of 409 term/definition pairs were considered. Five term/definition pairs were chosen, serving as few-shot default values, while the remaining 404 pairs served as the multiple-choice questions. Four choices were given for each term/definition question pair, where one was the correct answer. The three incorrect choices were picked randomly from all possible terms minus the true answer. Zero-Shot In-Context Experiment. Norwegian Petroleum Directorate (NPD) Database. The authors explored the wellbore history of all individual exploration wells drilled in the Norwegian shelf in the NPD database. In this experiment, 12 exploration wells were randomly chosen for evaluation. In addition to these drilling reports, information about the stratigraphy of three additional wells was added. Annual Reports. Annual reports of two major operators in Norway for 2020 and 2021 also were considered. These consisted of short summaries that presented the main operational and economic results achieved by the company throughout the year. These reports were added to the evaluation to balance the higher technical content of the wellbore-history reports.

  • Research Article
  • Cite Count Icon 34
  • 10.1109/jproc.2023.3279374
Zero-Shot and Few-Shot Learning With Knowledge Graphs: A Comprehensive Survey
  • Jun 1, 2023
  • Proceedings of the IEEE
  • Jiaoyan Chen + 7 more

Machine learning (ML), especially deep neural networks, has achieved great success, but many of them often rely on a number of labeled samples for supervision. As sufficient labeled training data are not always ready due to, e.g., continuously emerging prediction targets and costly sample annotation in real-world applications, ML with sample shortage is now being widely investigated. Among all these studies, many prefer to utilize auxiliary information including those in the form of knowledge graph (KG) to reduce the reliance on labeled samples. In this survey, we have comprehensively reviewed over <inline-formula xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink"> <tex-math notation="LaTeX">$\text{90}$</tex-math> </inline-formula> articles about KG-aware research for two major sample shortage settings—zero-shot learning (ZSL) where some classes to be predicted have no labeled samples and few-shot learning (FSL) where some classes to be predicted have only a small number of labeled samples that are available. We first introduce KGs used in ZSL and FSL as well as their construction methods and then systematically categorize and summarize KG-aware ZSL and FSL methods, dividing them into different paradigms, such as the mapping-based, the data augmentation, the propagation-based, and the optimization-based. We next present different applications, including not only KG augmented prediction tasks such as image classification, question answering, text classification, and knowledge extraction but also KG completion tasks and some typical evaluation resources for each task. We eventually discuss some challenges and open problems from different perspectives.

  • Research Article
  • 10.62408/ai-ling.v1i1.13
Exploiting ChatGPT to simplify Italian bureaucratic and professional texts
  • Aug 7, 2024
  • AI-Linguistica. Linguistic Studies on AI-Generated Texts and Discourses
  • Walter Paci + 4 more

This paper investigates the use of ChatGPT, a large language model, for simplifying long sentences and nominal clusters in professional texts belonging to administrative and legal domains. We apply three prompt engineering techniques — zero-shot learning, few-shot learning, and Chain-of-Thought reasoning — to generate alternative sentences from a corpus of Italian texts. We evaluate the generated sentences using a survey with expert and non-expert readers of bureaucratic and legal Italian, focusing on ease of understanding, coherence, and preferences in rephrasing. Our results show that ChatGPT can effectively address the linguistic challenges outlined by UNI 11482:2013 Standard, and that complex prompting techniques yield better outcomes than simpler ones. We also discuss the implications of our findings for the optimization of text understanding and simplification using large language models.

More from: Journal of Healthcare Informatics Research
  • New
  • Research Article
  • 10.1007/s41666-025-00216-6
Persuasive Design in a Digital Mindfulness Intervention: A Randomized Trial of a Skill-Based Achievement System and Automated Peer Encouragement
  • Oct 29, 2025
  • Journal of Healthcare Informatics Research
  • Abdul Rahman Idrees + 8 more

  • Research Article
  • 10.1007/s41666-025-00215-7
Impact Detection in Fall Events: Leveraging Spatio-temporal Graph Convolutional Networks and Recurrent Neural Networks Using 3D Skeleton Data
  • Sep 30, 2025
  • Journal of Healthcare Informatics Research
  • Tresor Y Koffi + 3 more

  • Research Article
  • 10.1007/s41666-025-00212-w
Multimodal Data Fusion for Whole-Slide Histopathology Image Classification
  • Sep 3, 2025
  • Journal of Healthcare Informatics Research
  • Yiran Song + 5 more

  • Research Article
  • 10.1007/s41666-025-00211-x
Socio-Demographic Modifiers Shape Large Language Models’ Ethical Decisions
  • Aug 12, 2025
  • Journal of Healthcare Informatics Research
  • Vera Sorin + 9 more

  • Research Article
  • 10.1007/s41666-025-00210-y
Instruction-Tuned Large Language Models for Clinical Data Extraction: Creating an Aortic Measurement Database from CT Radiology Reports
  • Aug 3, 2025
  • Journal of Healthcare Informatics Research
  • Ely Erez + 7 more

  • Research Article
  • 10.1007/s41666-025-00209-5
A Hybrid Bidirectional Deep Learning Model Using HRV for Prediction of ICU Mortality Risk in TBI Patients
  • Jul 30, 2025
  • Journal of Healthcare Informatics Research
  • Hasitha Kuruwita A + 7 more

  • Research Article
  • 10.1007/s41666-025-00208-6
Multi-Cohort Federated Learning Shows Synergy in Mortality Prediction for MRI-Based and Metabolomics-Based Age Scores
  • Jul 30, 2025
  • Journal of Healthcare Informatics Research
  • Pedro Mateus + 21 more

  • Research Article
  • 10.1007/s41666-025-00207-7
Smarter Together: Combining Large Language Models and Small Models for Physiological Signals Visual Inspection
  • Jul 28, 2025
  • Journal of Healthcare Informatics Research
  • Huayu Li + 10 more

  • Research Article
  • 10.1007/s41666-025-00203-x
Clinical Assessment of Fine-Tuned Open-Source LLMs in Cardiology: From Progress Notes to Discharge Summary
  • Jul 25, 2025
  • Journal of Healthcare Informatics Research
  • Hyoje Jung + 13 more

  • Research Article
  • 10.1007/s41666-025-00206-8
FairAlloc: Learning Fair Organ Allocation Policy for Liver Transplant
  • Jun 25, 2025
  • Journal of Healthcare Informatics Research
  • Sirui Ding + 6 more

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.

Search IconWhat is the difference between bacteria and viruses?
Open In New Tab Icon
Search IconWhat is the function of the immune system?
Open In New Tab Icon
Search IconCan diabetes be passed down from one generation to the next?
Open In New Tab Icon