LEVA: Using Large Language Models to Enhance Visual Analytics.

  • Abstract
  • References
  • Citations
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

Visual analytics supports data analysis tasks within complex domain problems. However, due to the richness of data types, visual designs, and interaction designs, users need to recall and process a significant amount of information when they visually analyze data. These challenges emphasize the need for more intelligent visual analytics methods. Large language models have demonstrated the ability to interpret various forms of textual data, offering the potential to facilitate intelligent support for visual analytics. We propose LEVA, a framework that uses large language models to enhance users' VA workflows at multiple stages: onboarding, exploration, and summarization. To support onboarding, we use large language models to interpret visualization designs and view relationships based on system specifications. For exploration, we use large language models to recommend insights based on the analysis of system status and data to facilitate mixed-initiative exploration. For summarization, we present a selective reporting strategy to retrace analysis history through a stream visualization and generate insight reports with the help of large language models. We demonstrate how LEVA can be integrated into existing visual analytics systems. Two usage scenarios and a user study suggest that LEVA effectively aids users in conducting visual analytics.

ReferencesShowing 10 of 47 papers
  • Open Access Icon
  • Cite Count Icon 28
  • 10.1109/tvcg.2021.3114211
ChartStory: Automated Partitioning, Layout, and Captioning of Charts into Comic-Style Narratives.
  • Feb 1, 2023
  • IEEE Transactions on Visualization and Computer Graphics
  • Jian Zhao + 8 more

  • Open Access Icon
  • Cite Count Icon 770
  • 10.1007/978-3-540-70956-5_7
Visual Analytics: Definition, Process, and Challenges
  • Apr 11, 2008
  • Daniel Keim + 5 more

  • Cite Count Icon 65
  • 10.1111/cgf.13730
A Review of Guidance Approaches in Visual Data Analysis: A Multifocal Perspective
  • Jun 1, 2019
  • Computer Graphics Forum
  • Davide Ceneda + 2 more

  • Open Access Icon
  • Cite Count Icon 123
  • 10.1109/tvcg.2020.3030378
NL4DV: A Toolkit for Generating Analytic Specifications for Data Visualization from Natural Language Queries.
  • Oct 13, 2020
  • IEEE Transactions on Visualization and Computer Graphics
  • Arpit Narechania + 2 more

  • Open Access Icon
  • Cite Count Icon 36
  • 10.1145/3313831.3376533
Paths Explored, Paths Omitted, Paths Obscured: Decision Points & Selective Reporting in End-to-End Data Analysis
  • Apr 21, 2020
  • Yang Liu + 2 more

  • Cite Count Icon 82
  • 10.1109/tvcg.2021.3114848
Natural Language to Visualization by Neural Machine Translation.
  • Jan 1, 2022
  • IEEE Transactions on Visualization and Computer Graphics
  • Yuyu Luo + 5 more

  • Cite Count Icon 96
  • 10.18653/v1/2023.acl-long.870
Can Large Language Models Be an Alternative to Human Evaluations?
  • Jan 1, 2023
  • Cheng-Han Chiang + 1 more

  • Open Access Icon
  • Cite Count Icon 616
  • 10.1109/tvcg.2013.124
A Multi-Level Typology of Abstract Visualization Tasks
  • Dec 1, 2013
  • IEEE Transactions on Visualization and Computer Graphics
  • Matthew Brehmer + 1 more

  • Cite Count Icon 201
  • 10.1057/ivs.2008.31
Characterizing Users' Visual Analytic Activity for Insight Provenance
  • Jan 1, 2009
  • Information Visualization
  • David Gotz + 1 more

  • Open Access Icon
  • Cite Count Icon 44
  • 10.1109/dasfaa.2003.1192383
Prefetching for visual data exploration
  • Jan 1, 2003
  • P.R Doshi + 2 more

CitationsShowing 10 of 20 papers
  • Conference Article
  • 10.1145/3664476.3670943
Evaluating Cyber Security Dashboards for Smart Cities and Buildings: Enhancing User Modeling with LLMs
  • Jul 30, 2024
  • Hanning Zhao + 1 more

Evaluating Cyber Security Dashboards for Smart Cities and Buildings: Enhancing User Modeling with LLMs

  • Research Article
  • Cite Count Icon 3
  • 10.1109/tvcg.2024.3456378
How Aligned are Human Chart Takeaways and LLM Predictions? A Case Study on Bar Charts with Varying Layouts.
  • Jan 1, 2025
  • IEEE transactions on visualization and computer graphics
  • Huichen Will Wang + 4 more

Large Language Models (LLMs) have been adopted for a variety of visualizations tasks, but how far are we from perceptually aware LLMs that can predict human takeaways? Graphical perception literature has shown that human chart takeaways are sensitive to visualization design choices, such as spatial layouts. In this work, we examine the extent to which LLMs exhibit such sensitivity when generating takeaways, using bar charts with varying spatial layouts as a case study. We conducted three experiments and tested four common bar chart layouts: vertically juxtaposed, horizontally juxtaposed, overlaid, and stacked. In Experiment 1, we identified the optimal configurations to generate meaningful chart takeaways by testing four LLMs, two temperature settings, nine chart specifications, and two prompting strategies. We found that even state-of-the-art LLMs struggled to generate semantically diverse and factually accurate takeaways. In Experiment 2, we used the optimal configurations to generate 30 chart takeaways each for eight visualizations across four layouts and two datasets in both zero-shot and one-shot settings. Compared to human takeaways, we found that the takeaways LLMs generated often did not match the types of comparisons made by humans. In Experiment 3, we examined the effect of chart context and data on LLM takeaways. We found that LLMs, unlike humans, exhibited variation in takeaway comparison types for different bar charts using the same bar layout. Overall, our case study evaluates the ability of LLMs to emulate human interpretations of data and points to challenges and opportunities in using LLMs to predict human chart takeaways.

  • Open Access Icon
  • Research Article
  • Cite Count Icon 3
  • 10.1109/tvcg.2024.3496112
LightVA: Lightweight Visual Analytics With LLM Agent-Based Task Planning and Execution.
  • Sep 1, 2025
  • IEEE transactions on visualization and computer graphics
  • Yuheng Zhao + 7 more

Visual analytics (VA) requires analysts to iteratively propose analysis tasks based on observations and execute tasks by creating visualizations and interactive exploration to gain insights. This process demands skills in programming, data processing, and visualization tools, highlighting the need for a more intelligent, streamlined VA approach. Large language models (LLMs) have recently been developed as agents to handle various tasks with dynamic planning and tool-using capabilities, offering the potential to enhance the efficiency and versatility of VA. We propose LightVA, a lightweight VA framework that supports task decomposition, data analysis, and interactive exploration through human-agent collaboration. Our method is designed to help users progressively translate high-level analytical goals into low-level tasks, producing visualizations and deriving insights. Specifically, we introduce an LLM agent-based task planning and execution strategy, employing a recursive process involving a planner, executor, and controller. The planner is responsible for recommending and decomposing tasks, the executor handles task execution, including data analysis, visualization generation and multi-view composition, and the controller coordinates the interaction between the planner and executor. Building on the framework, we develop a system with a hybrid user interface that includes a task flow diagram for monitoring and managing the task planning process, a visualization panel for interactive data exploration, and a chat view for guiding the model through natural language instructions. We examine the effectiveness of our method through a usage scenario and an expert study.

  • Conference Article
  • 10.1145/3708359.3712087
A picture is worth a thousand words? Investigating the Impact of Image Aids in AR on Memory Recall for Everyday Tasks
  • Mar 24, 2025
  • Elizaveta Lukianova + 2 more

A picture is worth a thousand words? Investigating the Impact of Image Aids in AR on Memory Recall for Everyday Tasks

  • New
  • Conference Article
  • 10.1145/3746270.3760233
Zero-shot Emotion Annotation in Facial Images Using Large Multimodal Models: Benchmarking and Prospects for Multi-Class, Multi-Frame Approaches
  • Oct 26, 2025
  • He Zhang + 1 more

Zero-shot Emotion Annotation in Facial Images Using Large Multimodal Models: Benchmarking and Prospects for Multi-Class, Multi-Frame Approaches

  • Open Access Icon
  • Conference Article
  • 10.1145/3706598.3713913
Jupybara: Operationalizing a Design Space for Actionable Data Analysis and Storytelling with LLMs
  • Apr 25, 2025
  • Huichen Will Wang + 2 more

Jupybara: Operationalizing a Design Space for Actionable Data Analysis and Storytelling with LLMs

  • Open Access Icon
  • Research Article
  • Cite Count Icon 5
  • 10.1109/tvcg.2024.3456350
DracoGPT: Extracting Visualization Design Preferences from Large Language Models.
  • Jan 1, 2025
  • IEEE transactions on visualization and computer graphics
  • Huichen Will Wang + 3 more

Trained on vast corpora, Large Language Models (LLMs) have the potential to encode visualization design knowledge and best practices. However, if they fail to do so, they might provide unreliable visualization recommendations. What visualization design preferences, then, have LLMs learned? We contribute DracoGPT, a method for extracting, modeling, and assessing visualization design preferences from LLMs. To assess varied tasks, we develop two pipelines-DracoGPT-Rank and DracoGPT-Recommend-to model LLMs prompted to either rank or recommend visual encoding specifications. We use Draco as a shared knowledge base in which to represent LLM design preferences and compare them to best practices from empirical research. We demonstrate that DracoGPT can accurately model the preferences expressed by LLMs, enabling analysis in terms of Draco design constraints. Across a suite of backing LLMs, we find that DracoGPT-Rank and DracoGPT-Recommend moderately agree with each other, but both substantially diverge from guidelines drawn from human subjects experiments. Future work can build on our approach to expand Draco's knowledge base to model a richer set of preferences and to provide a robust and cost-effective stand-in for LLMs.

  • Research Article
  • Cite Count Icon 2
  • 10.1109/tvcg.2024.3412241
Linking Text and Visualizations via Contextual Knowledge Graph.
  • Sep 1, 2025
  • IEEE transactions on visualization and computer graphics
  • Xiwen Cai + 5 more

The integration of visualizations and text is commonly found in data news, analytical reports, and interactive documents. For example, financial articles are presented along with interactive charts to show the changes in stock prices on Yahoo Finance. Visualizations enhance the perception of facts in the text while the text reveals insights of visual representation. However, effectively combining text and visualizations is challenging and tedious, which usually involves advanced programming skills. This paper proposes a semi-automatic pipeline that builds links between text and visualization. To resolve the relationship between text and visualizations, we present a method which structures a visualization and the underlying data as a contextual knowledge graph, based on which key phrases in the text are extracted, grouped, and mapped with visual elements. To support flexible customization of text-visualization links, our pipeline incorporates user knowledge to revise the links in a mixed-initiative manner. To demonstrate the usefulness and the versatility of our method, we replicate prior studies or cases in crafting interactive word-sized visualizations, annotating visualizations, and creating text-chart interactions based on a prototype system. We carry out two preliminary model tests and a user study and the results and user feedbacks suggest our method is effective.

  • Open Access Icon
  • Conference Article
  • 10.1109/beliv64461.2024.00012
The Visualization JUDGE: Can Multimodal Foundation Models Guide Visualization Design Through Visual Perception?
  • Oct 14, 2024
  • Matthew Berger + 1 more

Foundation models for vision and language are the basis of AI applications across numerous sectors of society. The success of these models stems from their ability to mimic human capabilities, namely visual perception in vision models, and analytical reasoning in large language models. As visual perception and analysis are fundamental to data visualization, in this position paper we ask: how can we harness foundation models to advance progress in visualization design? Specifically, how can multimodal foundation models (MFMs) guide visualization design through visual perception? We approach these questions by investigating the effectiveness of MFMs for perceiving visualization, and formalizing the overall visualization design and optimization space. Specifically, we think that MFMs can best be viewed as judges, equipped with the ability to criticize visualizations, and provide us with actions on how to improve a visualization. We provide a deeper characterization for text-to-image generative models, and multi-modal large language models, organized by what these models provide as output, and how to utilize the output for guiding design decisions. We hope that our perspective can inspire researchers in visualization on how to approach MFMs for visualization design.

  • Conference Article
  • 10.1109/pacificvis64226.2025.00012
SmartMLVs: LLM-enabled Multiple Linked Views Generation for Interactive Visualization
  • Apr 22, 2025
  • Tian Qiu + 6 more

SmartMLVs: LLM-enabled Multiple Linked Views Generation for Interactive Visualization

Similar Papers
  • Research Article
  • 10.1111/cgf.70112
InterChat: Enhancing Generative Visual Analytics using Multimodal Interactions
  • May 23, 2025
  • Computer Graphics Forum
  • Juntong Chen + 8 more

The rise of Large Language Models (LLMs) and generative visual analytics systems has transformed data‐driven insights, yet significant challenges persist in accurately interpreting users analytical and interaction intents. While language inputs offer flexibility, they often lack precision, making the expression of complex intents inefficient, error‐prone, and time‐intensive. To address these limitations, we investigate the design space of multimodal interactions for generative visual analytics through a literature review and pilot brainstorming sessions. Building on these insights, we introduce a highly extensible workflow that integrates multiple LLM agents for intent inference and visualization generation. We develop InterChat, a generative visual analytics system that combines direct manipulation of visual elements with natural language inputs. This integration enables precise intent communication and supports progressive, visually driven exploratory data analyses. By employing effective prompt engineering, and contextual interaction linking, alongside intuitive visualization and interaction designs, InterChat bridges the gap between user interactions and LLM‐driven visualizations, enhancing both interpretability and usability. Extensive evaluations, including two usage scenarios, a user study, and expert feedback, demonstrate the effectiveness of InterChat. Results show significant improvements in the accuracy and efficiency of handling complex visual analytics tasks, highlighting the potential of multimodal interactions to redefine user engagement and analytical depth in generative visual analytics.

  • Research Article
  • Cite Count Icon 3
  • 10.1109/tvcg.2024.3496112
LightVA: Lightweight Visual Analytics With LLM Agent-Based Task Planning and Execution.
  • Sep 1, 2025
  • IEEE transactions on visualization and computer graphics
  • Yuheng Zhao + 7 more

Visual analytics (VA) requires analysts to iteratively propose analysis tasks based on observations and execute tasks by creating visualizations and interactive exploration to gain insights. This process demands skills in programming, data processing, and visualization tools, highlighting the need for a more intelligent, streamlined VA approach. Large language models (LLMs) have recently been developed as agents to handle various tasks with dynamic planning and tool-using capabilities, offering the potential to enhance the efficiency and versatility of VA. We propose LightVA, a lightweight VA framework that supports task decomposition, data analysis, and interactive exploration through human-agent collaboration. Our method is designed to help users progressively translate high-level analytical goals into low-level tasks, producing visualizations and deriving insights. Specifically, we introduce an LLM agent-based task planning and execution strategy, employing a recursive process involving a planner, executor, and controller. The planner is responsible for recommending and decomposing tasks, the executor handles task execution, including data analysis, visualization generation and multi-view composition, and the controller coordinates the interaction between the planner and executor. Building on the framework, we develop a system with a hybrid user interface that includes a task flow diagram for monitoring and managing the task planning process, a visualization panel for interactive data exploration, and a chat view for guiding the model through natural language instructions. We examine the effectiveness of our method through a usage scenario and an expert study.

  • Research Article
  • 10.1109/mcg.2024.3517293
AI-in-The-Loop: The Future of Biomedical Visual Analytics Applications in the Era of AI.
  • Mar 1, 2025
  • IEEE computer graphics and applications
  • Katja Bühler + 3 more

AI is the workhorse of modern data analytics and omnipresent across many sectors. Large language models and multimodal foundation models are today capable of generating code, charts, visualizations, etc. How will these massive developments of AI in data analytics shape future data visualizations and visual analytics workflows? What is the potential of AI to reshape methodology and design of future visual analytics applications? What will be our role as visualization researchers in the future? What are opportunities, open challenges, and threats in the context of an increasingly powerful AI? This Visualization Viewpoints discusses these questions in the special context of biomedical data analytics as an example of a domain in which critical decisions are taken based on complex and sensitive data, with high requirements on transparency, efficiency, and reliability. We map recent trends and developments in AI on the elements of interactive visualization and visual analytics workflows and highlight the potential of AI to transform biomedical visualization as a research field. Given that agency and responsibility have to remain with human experts, we argue that it is helpful to keep the focus on human-centered workflows, and to use visual analytics as a tool for integrating "AI-in-the-loop." This is in contrast to the more traditional term "human-in-the-loop." which focuses on incorporating human expertise into AI-based systems.

  • Research Article
  • 10.1007/s10726-025-09935-y
Idea Evaluation for Solutions to Specialized Problems: Leveraging the Potential of Crowds and Large Language Models
  • Jun 28, 2025
  • Group Decision and Negotiation
  • Henner Gimpel + 4 more

Complex problems such as climate change pose severe challenges to societies worldwide. To overcome these challenges, digital innovation contests have emerged as a promising tool for idea generation. However, assessing idea quality in innovation contests is becoming increasingly problematic in domains where specialized knowledge is needed. Traditionally, expert juries are responsible for idea evaluation in such contests. However, experts are a substantial bottleneck as they are often scarce and expensive. To assess whether expert juries could be replaced, we consider two approaches. We leverage crowdsourcing and a Large Language Model (LLM) to evaluate ideas, two approaches that are similar in terms of the aggregation of collective knowledge and could therefore be close to expert knowledge. We compare expert jury evaluations from innovation contests on climate change with crowdsourced and LLM’s evaluations and assess performance differences. Results indicate that crowds and LLMs have the ability to evaluate ideas in the complex problem domain while contest specialization—the degree to which a contest relates to a knowledge-intensive domain rather than a broad field of interest—is an inhibitor of crowd evaluation performance but does not influence the evaluation performance of LLMs. Our contribution lies with demonstrating that crowds and LLMs (as opposed to traditional expert juries) are suitable for idea evaluation and allows innovation contest operators to integrate the knowledge of crowds and LLMs to reduce the resource bottleneck of expert juries.

  • Conference Article
  • Cite Count Icon 10
  • 10.1109/hicss.2011.199
Expanding the Scope: Interaction Design Perspectives for Visual Analytics
  • Jan 1, 2011
  • T M Green + 2 more

In this paper, we explore the current state of interaction design in visual analytics. Current visual analytics design is heavily focused on interface issues like scalability and tool functionality; this focus is necessary, but it should not be exclusive. Further, most consideration of human cognition is done after tool development in the form of limited evaluation. We argue that, by definition, visual analytics is the science of analytical reasoning facilitated by visual interfaces, and as such, should consider complex human cognition in visual analytics design before and after tool development. We discuss two extant approaches to interaction design (Activity Theory and Participatory Design) and discuss how they might be applied, as well as the potential benefits to these approaches. We also introduce a design tool adapted for visual analytics, and provide an example of visual analytics interaction design in action. Future implications of this work are also discussed.

  • Research Article
  • Cite Count Icon 1
  • 10.1109/tvcg.2024.3456215
PhenoFlow: A Human-LLM Driven Visual Analytics System for Exploring Large and Complex Stroke Datasets.
  • Jan 1, 2025
  • IEEE transactions on visualization and computer graphics
  • Jaeyoung Kim + 6 more

Acute stroke demands prompt diagnosis and treatment to achieve optimal patient outcomes. However, the intricate and irregular nature of clinical data associated with acute stroke, particularly blood pressure (BP) measurements, presents substantial obstacles to effective visual analytics and decision-making. Through a year-long collaboration with experienced neurologists, we developed PhenoFlow, a visual analytics system that leverages the collaboration between human and Large Language Models (LLMs) to analyze the extensive and complex data of acute ischemic stroke patients. PhenoFlow pioneers an innovative workflow, where the LLM serves as a data wrangler while neurologists explore and supervise the output using visualizations and natural language interactions. This approach enables neurologists to focus more on decision-making with reduced cognitive load. To protect sensitive patient information, PhenoFlow only utilizes metadata to make inferences and synthesize executable codes, without accessing raw patient data. This ensures that the results are both reproducible and interpretable while maintaining patient privacy. The system incorporates a slice-and-wrap design that employs temporal folding to create an overlaid circular visualization. Combined with a linear bar graph, this design aids in exploring meaningful patterns within irregularly measured BP data. Through case studies, PhenoFlow has demonstrated its capability to support iterative analysis of extensive clinical datasets, reducing cognitive load and enabling neurologists to make well-informed decisions. Grounded in long-term collaboration with domain experts, our research demonstrates the potential of utilizing LLMs to tackle current challenges in data-driven clinical decision-making for acute ischemic stroke patients.

  • Research Article
  • 10.1093/ndt/gfae069.792
#2924 Comparison of large language models and traditional natural language processing techniques in predicting arteriovenous fistula failure
  • May 23, 2024
  • Nephrology Dialysis Transplantation
  • Suman Lama + 6 more

Background and Aims Large language models (LLMs) have gained significant attention in the field of natural language processing (NLP), marking a shift from traditional techniques like Term Frequency-Inverse Document Frequency (TF-IDF). We developed a traditional NLP model to predict arteriovenous fistula (AVF) failure within next 30 days using clinical notes. The goal of this analysis was to investigate whether LLMs would outperform traditional NLP techniques, specifically in the context of predicting AVF failure within the next 30 days using clinical notes. Method We defined AVF failure as the change in status from active to permanently unusable status or temporarily unusable status. We used data from a large kidney care network from January 2021 to December 2021. Two models were created using LLMs and traditional TF-IDF technique. We used “distilbert-base-uncased”, a distilled version of BERT base model [1], and compared its performance with traditional TF-IDF-based NLP techniques. The dataset was randomly divided into 60% training, 20% validation and 20% test dataset. The test data, comprising of unseen patients’ data was used to evaluate the performance of the model. Both models were evaluated using metrics such as area under the receiver operating curve (AUROC), accuracy, sensitivity, and specificity. Results The incidence of 30 days AVF failure rate was 2.3% in the population. Both LLMs and traditional showed similar overall performance as summarized in Table 1. Notably, LLMs showed marginally better performance in certain evaluation metrics. Both models had same AUROC of 0.64 on test data. The accuracy and balanced accuracy for LLMs were 72.9% and 59.7%, respectively, compared to 70.9% and 59.6% for the traditional TF-IDF approach. In terms of specificity, LLMs scored 73.2%, slightly higher than the 71.2% observed for traditional NLP methods. However, LLMs had a lower sensitivity of 46.1% compared to 48% for traditional NLP. However, it is worth noting that training on LLMs took considerably longer than TF-IDF. Moreover, it also used higher computational resources such as utilization of graphics processing units (GPU) instances in cloud-based services, leading to higher cost. Conclusion In our study, we discovered that advanced LLMs perform comparably to traditional TF-IDF modeling techniques in predicting the failure of AVF. Both models demonstrated identical AUROC. While specificity was higher in LLMs compared to traditional NLP, sensitivity was higher in traditional NLP compared to LLMs. LLM was fine-tuned with a limited dataset, which could have influenced its performance to be similar to that of traditional NLP methods. This finding suggests that while LLMs may excel in certain scenarios, such as performing in-depth sentiment analysis of patient data for complex tasks, their effectiveness is highly dependent on the specific use case. It is crucial to weigh the benefits against the resources required for LLMs, as they can be significantly more resource-intensive and costly compared to traditional TF-IDF methods. This highlights the importance of a use-case-driven approach in selecting the appropriate NLP technique for healthcare applications.

  • Research Article
  • Cite Count Icon 2
  • 10.1371/journal.pone.0317084
Leveraging large language models for data analysis automation.
  • Feb 21, 2025
  • PloS one
  • Jacqueline A Jansen + 3 more

Data analysis is constrained by a shortage of skilled experts, particularly in biology, where detailed data analysis and subsequent interpretation is vital for understanding complex biological processes and developing new treatments and diagnostics. One possible solution to this shortage in experts would be making use of Large Language Models (LLMs) for generating data analysis pipelines. However, although LLMs have shown great potential when used for code generation tasks, questions regarding the accuracy of LLMs when prompted with domain expert questions such as omics related data analysis questions, remain unanswered. To address this, we developed mergen, an R package that leverages LLMs for data analysis code generation and execution. We evaluated the performance of this data analysis system using various data analysis tasks for genomics. Our primary goal is to enable researchers to conduct data analysis by simply describing their objectives and the desired analyses for specific datasets through clear text. Our approach improves code generation via specialized prompt engineering and error feedback mechanisms. In addition, our system can execute the data analysis workflows prescribed by the LLM providing the results of the data analysis workflow for human review. Our evaluation of this system reveals that while LLMs effectively generate code for some data analysis tasks, challenges remain in executable code generation, especially for complex data analysis tasks. The best performance was seen with the self-correction mechanism, in which self-correct was able to increase the percentage of executable code when compared to the simple strategy by 22.5% for tasks of complexity 2. For tasks for complexity 3, 4 and 5, this increase was 52.5%, 27.5% and 15% respectively. Using a chi-squared test, it was shown that significant differences could be found using the different prompting strategies. Our study contributes to a better understanding of LLM capabilities and limitations, providing software infrastructure and practical insights for their effective integration into data analysis workflows.

  • Research Article
  • Cite Count Icon 5
  • 10.1109/tvcg.2024.3456350
DracoGPT: Extracting Visualization Design Preferences from Large Language Models.
  • Jan 1, 2025
  • IEEE transactions on visualization and computer graphics
  • Huichen Will Wang + 3 more

Trained on vast corpora, Large Language Models (LLMs) have the potential to encode visualization design knowledge and best practices. However, if they fail to do so, they might provide unreliable visualization recommendations. What visualization design preferences, then, have LLMs learned? We contribute DracoGPT, a method for extracting, modeling, and assessing visualization design preferences from LLMs. To assess varied tasks, we develop two pipelines-DracoGPT-Rank and DracoGPT-Recommend-to model LLMs prompted to either rank or recommend visual encoding specifications. We use Draco as a shared knowledge base in which to represent LLM design preferences and compare them to best practices from empirical research. We demonstrate that DracoGPT can accurately model the preferences expressed by LLMs, enabling analysis in terms of Draco design constraints. Across a suite of backing LLMs, we find that DracoGPT-Rank and DracoGPT-Recommend moderately agree with each other, but both substantially diverge from guidelines drawn from human subjects experiments. Future work can build on our approach to expand Draco's knowledge base to model a richer set of preferences and to provide a robust and cost-effective stand-in for LLMs.

  • Research Article
  • Cite Count Icon 1
  • 10.1145/2644448.2644455
A comparative approach to enhance information interaction design of visual analytics systems
  • May 1, 2014
  • Communication Design Quarterly
  • Zhenyu Cheryl Qian + 2 more

This paper introduces a novel comparative strategy to access, synthesize, and redesign a mobile visual analytics (VA) system. Designing, evaluating, and improving VA tools are challenging because of the exploratory and unpredicted nature of their users' analysis activities in a real context. Often the system development approach is running rounds of iteration based on one or a few design ideas and related references. Inspired by ideation and design selection from design-thinking literature, we start to redesign systems from comparison and filtering based on a broad range of design ideas. This approach focuses on the information interaction design of systems; integrates design principles from information design, sensorial design, and interaction design as guidelines; compares VA systems at the component level; and seeks unique and adaptive design solutions. The Visual Analytics Benchmark Repository provides a rich collection of the Visual Analytics Science and Technology (VAST) challenges submission reports and videos. For each challenge design problem, there are multiple creative and mature design solutions. Based on this resource, we conducted a series of empirical user studies to understand the user experience by comparing different design solutions, enhanced one visual analytics system design MobileAnalymator by synthesizing new features and removing redundant functions, and accessed the redesign outcomes with the same comparative approach.

  • Research Article
  • 10.1200/jco.2025.43.16_suppl.11160
Evaluation of large language model (LLM)-based clinical abstraction of electronic health records (EHRs) for non-small cell lung cancer (NSCLC) patients.
  • Jun 1, 2025
  • Journal of Clinical Oncology
  • Kabir Manghnani + 10 more

11160 Background: Abstraction is a critical step for converting clinical data from unstructured EHRs into a structured format suitable for real-world data analyses. Typically this is a manual, labor-intensive activity requiring substantial training. While prior work has shown that abstraction by humans is reliable, advances in LLMs may improve the efficiency of abstraction. We aim to measure the performance of LLMs in abstracting a diverse set of oncology data elements. Methods: Two clinical abstractors independently abstracted unstructured records of 222 advanced or metastatic NSCLC patients (mean: 248 pages per case). A two-stage LLM system balancing cost and comprehensiveness was used to abstract clinical elements for demographics, diagnosis, third-party lab biomarker testing, and first line (1L) treatment. The initial stage extracted 16 documents semantically similar to the abstraction query and input them, along with abstraction rules, into an LLM (GPT-4o). The LLM was instructed to provide both the abstracted field and a completeness assessment of provided context. If the first phase resulted in a low completeness score, the entire patient record was then input into a long-context LLM (Gemini-Pro-1.5) to re-attempt abstraction. Gwet’s agreement coefficient (AC) was the primary measure of agreement between the LLM and each abstractor. Date agreement was calculated within ±30 days. Results: The LLM system yielded abstracted values for 90% of elements where both abstractors provided non-missing values. In these cases, the LLM also demonstrated high agreement with each abstractor (≥0.81 across all categories). Agreement was highest in demographic and diagnosis domains and lower for 1L treatment domain, which require deeper understanding of a patient's temporal journey. For elements where neither abstractor provided values, the LLM sometimes provided outputs (frequency: 4.9% for non-biomarker elements; 38.5% for biomarker elements). These discrepancies were primarily driven by nuances in abstraction rules; the LLM often included Tempus-tested biomarkers, while abstractors were more rigorous in abstracting only third-party biomarker results. Conclusions: LLMs show high completion rates and high agreement with human abstractors across a variety of critical abstraction fields. The use of LLMs may significantly reduce the burden of human abstraction and allow for large-scale curation of oncology records. Challenges in handling nuanced contexts underscore the need for careful refinement and evaluation prior to deployment. Domain LLM agreement with abstractors (AC, min-max) Demographic (birth date, sex, race, smoking status) 0.96-1 Diagnosis (stage, histology, year of diagnosis) 0.92-0.98 Third Party Biomarker (EGFR, ALK, ROS1, PDL1, BRAF, RET, NTRK) 0.87-1 1L Treatment (agents, initiation date) 0.81-0.86

  • Research Article
  • Cite Count Icon 40
  • 10.1145/3576935
Visualization and Visual Analytics Approaches for Image and Video Datasets: A Survey
  • Mar 9, 2023
  • ACM Transactions on Interactive Intelligent Systems
  • Shehzad Afzal + 6 more

Image and video data analysis has become an increasingly important research area with applications in different domains such as security surveillance, healthcare, augmented and virtual reality, video and image editing, activity analysis and recognition, synthetic content generation, distance education, telepresence, remote sensing, sports analytics, art, non-photorealistic rendering, search engines, and social media. Recent advances in Artificial Intelligence (AI) and particularly deep learning have sparked new research challenges and led to significant advancements, especially in image and video analysis. These advancements have also resulted in significant research and development in other areas such as visualization and visual analytics, and have created new opportunities for future lines of research. In this survey article, we present the current state of the art at the intersection of visualization and visual analytics, and image and video data analysis. We categorize the visualization articles included in our survey based on different taxonomies used in visualization and visual analytics research. We review these articles in terms of task requirements, tools, datasets, and application areas. We also discuss insights based on our survey results, trends and patterns, the current focus of visualization research, and opportunities for future research.

  • Conference Article
  • Cite Count Icon 8
  • 10.1109/immersive.2016.7932378
Towards mobile immersive analysis: A study of applications
  • Mar 20, 2016
  • Aidong Lu + 4 more

The latest new technologies of virtual and mixed realities are ready to be commercialized. Their innovative communication channels have created new opportunities and challenges for visualization and visual analytics. At this stage, it is important to understand the state-of-the-art of relevant technologies as well as the impacts to the research fields. This position paper presents the key features of mobile, immersive environments that are different from traditional visualization and visual analytics systems. Through a study of applications in the mixed reality, we present a rich example set of capabilities and applications that address discovery and visual communication in real-world uses. We hope that this work will provide references and promote the research on the topic of immersive visualization and visual analytics.

  • Research Article
  • 10.1200/jco.2025.43.16_suppl.e23161
Aiding data retrieval in clinical trials with large language models: The APOLLO 11 Consortium in advanced lung cancer patients.
  • Jun 1, 2025
  • Journal of Clinical Oncology
  • Federica Corso + 19 more

e23161 Background: Data retrieval is challenging in clinical research and traditional methods for data collection are often time-consuming and may be error-prone. Large Language Models (LLMs) have shown zero-shot capabilities in converting unstructured clinical text into structured data. These technologies could support the retrieval stage of clinical trials by leveraging the information reported in Electronic Health Records (EHRs) without relying any longer on manual curation. APOLLO 11 Consortium (NCT05550961) is a multicentric Italian trial which leverages a federated infrastructure for the analysis of advanced lung cancer patient data across Italy. Methods: We conducted a pilot study using Llama 3.1 8B on 358 Non-Small Cell Lung Cancer patients from the IRCCS Istituto Nazionale dei Tumori, leader of the APOLLO 11 Consortium. Anonymized EHRs have been analyzed within the LLM pipeline for feature extraction by Wiest et al. A combination of zero/few shot prompting techniques both in English and Italian languages was used. We selected smoking, histology, PD-L1 and staging as multiclass variables and bone/brain/liver metastases as binary variables. The ground truth collection involved a first Manual Data Entry (1-MDE) and a final full-revised MDE (2-MDE). The LLM accuracy was calculated only for the comparison LLM vs 2-MDE. In addition, we calculated the percentage of Missing Information (% MI) in 1-MDE, 2-MDE and LLM extraction. Results: Compared to 2-MDE, LLM achieved feature-specific accuracies of 0.78 for PD-L1, 0.85 for BONE METASTASIS, 0.83 for BRAIN METASTASIS, 0.89 for LIVER METASTASIS and 0.96 for TUMOUR STAGING. For smoking and staging, LLM extraction also reduced % MI relative to 1-MDE (Table 1). Only for PD-L1, we further analyzed the 12.8% of MI and found that 91.3% resulted from hallucinations (i.e., PD-L1 was misclassified as missing). Evaluations using English prompts confirmed the pipeline’s adaptability and high tasks accuracy. Conclusions: This study confirms the feasibility of LLMs for data retrieval in clinical trials demonstrating strong performance across diverse clinical features with minimal prompt optimization. LLMs could assist clinicians and data entry personnel in the 1-MDE process, streamlining initial data structuring and saving time. The 2-MDE step can remain as a quality check to address any discrepancies. Further improvements could focus on prompt optimization and integrating human feedback to reduce hallucination rates. Clinical trial information: NCT05550961 . %MI in 1-MDE, 2-MDE and LLM extraction. Accuracy refers only to LLM vs 2-MDE. Histology and metastasis sites were collected only in 2-MDE. NA = not available. Smoking PD-L1 Histology Bone Met Brain Met Liver Met T N M Stage % MI 1-MDE 6.4 8.9 NA NA NA NA 22.5 22.5 23.11 98.3 % MI 2-MDE 6.6 3 0 0 0 0 0 0 0 0 % MI LLM 2.7 12.8 10.3 0 0 0 0 0 0 6.9 % accuracy (LLM vs 2-MDE) 67 78 91 85 83 89 39 52 70 96

  • Conference Article
  • Cite Count Icon 2
  • 10.18260/p.23725
Computer-Vision-Aided Lip Movement Correction to Improve English Pronunciation
  • Jul 8, 2015
  • Shuang Wei + 3 more

Dr. Yingjie Chen is an assistant professor in the Department of Computer Graphics Technology of Purdue University. He received his Ph.D. degree in the areas of human-computer interaction, information visualization, and visual analytics from the School of Interaction Arts and Technology at Simon Fraser University (SFU) in Canada. He earned the Bachelor degree of Engineering from the Tsinghua University in China, and a Master of Science degree in Information Technology from SFU. His research covers interdisciplinary domains of information visualization, visual analytics, digital media, and human computer interaction. He seeks to design, model, and construct new forms of interaction in visualization and system design, by which the system can minimize its influence on design and analysis, and become a true free extension of human’s brain and hand.

More from: IEEE transactions on visualization and computer graphics
  • New
  • Research Article
  • 10.1109/tvcg.2025.3628181
Untangling Rhetoric, Pathos, and Aesthetics in Data Visualization.
  • Nov 7, 2025
  • IEEE transactions on visualization and computer graphics
  • Verena Prantl + 2 more

  • Research Article
  • 10.1109/tvcg.2025.3616763
Measurement of Visitor Behavioral Engagement in Heritage Informal Learning Environments Using Head-Mounted Displays.
  • Nov 1, 2025
  • IEEE transactions on visualization and computer graphics
  • Shuyu Luo + 4 more

  • Research Article
  • 10.1109/tvcg.2025.3616756
Selection at a Distance Through a Large Transparent Touch Screen.
  • Nov 1, 2025
  • IEEE transactions on visualization and computer graphics
  • Sebastian Rigling + 4 more

  • Research Article
  • 10.1109/tvcg.2025.3610275
IEEE ISMAR 2025 Introducing the Special Issue
  • Nov 1, 2025
  • IEEE Transactions on Visualization and Computer Graphics
  • Han-Wei Shen + 2 more

  • Research Article
  • 10.1109/tvcg.2025.3616842
Detecting Visual Information Manipulation Attacks in Augmented Reality: A Multimodal Semantic Reasoning Approach.
  • Nov 1, 2025
  • IEEE transactions on visualization and computer graphics
  • Yanming Xiu + 1 more

  • Research Article
  • 10.1109/tvcg.2025.3610302
IEEE ISMAR 2025 Science & Technology Program Committee Members for Journal Papers
  • Nov 1, 2025
  • IEEE Transactions on Visualization and Computer Graphics

  • Research Article
  • 10.1109/tvcg.2025.3616749
HAT Swapping: Virtual Agents as Stand-Ins for Absent Human Instructors in Virtual Training.
  • Nov 1, 2025
  • IEEE transactions on visualization and computer graphics
  • Jingjing Zhang + 8 more

  • Research Article
  • 10.1109/tvcg.2025.3616758
Viewpoint-Tolerant Depth Perception for Shared Extended Space Experience on Wall-Sized Display.
  • Nov 1, 2025
  • IEEE transactions on visualization and computer graphics
  • Dooyoung Kim + 3 more

  • Research Article
  • 10.1109/tvcg.2025.3616751
SGSG: Stroke-Guided Scene Graph Generation.
  • Nov 1, 2025
  • IEEE transactions on visualization and computer graphics
  • Qixiang Ma + 5 more

  • Research Article
  • 10.1109/tvcg.2025.3620888
IEEE Transactions on Visualization and Computer Graphics: 2025 IEEE International Symposium on Mixed and Augmented Reality
  • Nov 1, 2025
  • IEEE Transactions on Visualization and Computer Graphics

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.

Search IconWhat is the difference between bacteria and viruses?
Open In New Tab Icon
Search IconWhat is the function of the immune system?
Open In New Tab Icon
Search IconCan diabetes be passed down from one generation to the next?
Open In New Tab Icon