Network Approaches to Binary Assessment Data: Network Psychometrics Versus Latent Space Item Response Models.
This study compares two network-based approaches for analyzing binary psychological assessment data: network psychometrics and latent space item response modeling (LSIRM). Network psychometrics, a well-established method, infers relationships among items or symptoms based on pairwise conditional dependencies. In contrast, LSIRM is a more recent framework that represents item responses as a bipartite network of respondents and items embedded in a latent metric space, where the likelihood of a response decreases with increasing distance between the respondent and item. We evaluate the performance of both methods through simulation studies under varying data-generating conditions. In addition, we demonstrate their applications to real assessment data, showcasing the distinct insights each method offers to researchers and practitioners.
- Research Article
- 10.1037/met0000625
- Dec 21, 2023
- Psychological methods
Network psychometrics leverages pairwise Markov random fields to depict conditional dependencies among a set of psychological variables as undirected edge-weighted graphs. Researchers often intend to compare such psychometric networks across subpopulations, and recent methodological advances provide invariance tests of differences in subpopulation networks. What remains missing, though, is an analogue to an effect size measure that quantifies differences in psychometric networks. We address this gap by complementing recent advances for investigating whether psychometric networks differ with an intuitive similarity measure quantifying the extent to which networks differ. To this end, we build on graph-theoretic approaches and propose a similarity measure based on the Frobenius norm of differences in psychometric networks' weighted adjacency matrices. To assess this measure's utility for quantifying differences between psychometric networks, we study how it captures differences in subpopulation network models implied by both latent variable models and Gaussian graphical models. We show that a wide array of network differences translates intuitively into the proposed measure, while the same does not hold true for customary correlation-based comparisons. In a simulation study on finite-sample behavior, we show that the proposed measure yields trustworthy results when population networks differ and sample sizes are sufficiently large, but fails to identify exact similarity when population networks are the same. From these results, we derive a strong recommendation to only use the measure as a complement to a significant test for network similarity. We illustrate potential insights from quantifying psychometric network similarities through cross-country comparisons of human values networks. (PsycInfo Database Record (c) 2023 APA, all rights reserved).
- Research Article
7
- 10.1007/s11336-023-09920-x
- Sep 1, 2023
- Psychometrika
Traditional measurement models assume that all item responses correlate with each other only through their underlying latent variables. This conditional independence assumption has been extended in joint models of responses and response times (RTs), implying that an item has the same item characteristics fors all respondents regardless of levels of latent ability/trait and speed. However, previous studies have shown that this assumption is violated in various types of tests and questionnaires and there are substantial interactions between respondents and items that cannot be captured by person- and item-effect parameters in psychometric models with the conditional independence assumption. To study the existence and potential cognitive sources of conditional dependence and utilize it to extract diagnostic information for respondents and items, we propose a diffusion item response theory model integrated with the latent space of variations in information processing rate of within-individual measurement processes. Respondents and items are mapped onto the latent space, and their distances represent conditional dependence and unexplained interactions. We provide three empirical applications to illustrate (1) how to use an estimated latent space to inform conditional dependence and its relation to person and item measures, (2) how to derive diagnostic feedback personalized for respondents, and (3) how to validate estimated results with an external measure. We also provide a simulation study to support that the proposed approach can accurately recover its parameters and detect conditional dependence underlying data.
- Research Article
5
- 10.1371/journal.pone.0269376
- Jun 29, 2022
- PLoS ONE
We explore potential cross-informant discrepancies between child- and parent-report measures with an example of the Child Behavior Checklist (CBCL) and the Youth Self Report (YSR), parent- and self-report measures on children’s behavioral and emotional problems. We propose a new way of examining the parent- and child-report differences with an interaction map estimated using a Latent Space Item Response Model (LSIRM). The interaction map enables the investigation of the dependency between items, between respondents, and between items and respondents, which is not possible with the conventional approach. The LSIRM captures the differential positions of items and respondents in the latent spaces for CBCL and YSR and identifies the relationships between each respondent and item according to their dependent structures. The results suggest that the analysis of item response in the latent space using the LSIRM is beneficial in uncovering the differential structures embedded in the response data obtained from different perspectives in children and their parents. This study also argues that the differential hidden structures of children and parents’ responses should be taken together to evaluate children’s behavioral problems.
- Book Chapter
28
- 10.1016/s0169-7161(06)26015-2
- Jan 1, 2006
- Handbook of Statistics
15 Item Response Theory in a General Framework
- Supplementary Content
- 10.1136/gpsych-2025-102055
- Jul 13, 2025
- General Psychiatry
Loneliness is a complex and usually unpleasant emotional response to isolation, which has been considered the latest global health epidemic exacerbated by the coronavirus disease 2019 pandemic, affecting nearly two-thirds of older adults. Some profound health implications carried by loneliness include depression, cognitive impairment, hypertension and frailty. Across the world, there is no consensus definition of loneliness, and its measure is based on the phenomenological perspective of the individual. The 20-item University of California Los Angeles Loneliness Scale version 3 (UCLA-20) is the most common measure. This scale demonstrates acceptable psychometric properties but is too long and complex for a phone interview. This paper addresses the increasing need to shorten this scale by adopting classical item response theory and network psychometrics to advance scale development. Through an item reduction analysis, we trimmed the original scale into an effective short form, which is as valid as the original one. With respondents’ time at a premium in most research nowadays, this short-form scale is an efficient and practical alternative to the original UCLA-20.
- Research Article
- 10.3390/jintelligence12020023
- Feb 16, 2024
- Journal of Intelligence
There recently have been many studies examining conditional dependence between response accuracy and response times in cognitive tests. While most previous research has focused on revealing a general pattern of conditional dependence for all respondents and items, it is plausible that the pattern may vary across respondents and items. In this paper, we attend to its potential heterogeneity and examine the item and person specificities involved in the conditional dependence between item responses and response times. To this end, we use a latent space item response theory (LSIRT) approach with an interaction map that visualizes conditional dependence in response data in the form of item-respondent interactions. We incorporate response time information into the interaction map by applying LSIRT models to slow and fast item responses. Through empirical illustrations with three cognitive test datasets, we confirm the presence and patterns of conditional dependence between item responses and response times, a result consistent with previous studies. Our results further illustrate the heterogeneity in the conditional dependence across respondents, which provides insights into understanding individuals' underlying item-solving processes in cognitive tests. Some practical implications of the results and the use of interaction maps in cognitive tests are discussed.
- Research Article
- 10.1037/met0000791
- Aug 7, 2025
- Psychological methods
As with many other latent variable models, the confirmatory factor analysis model is built upon the conditional independence assumption, which states that latent variables and item parameters can fully explain covariations between item responses. However, growing evidence in psychological and educational measurement research challenges this assumption, raising concerns regarding conditional dependence (CD). As the main model parameters correspond to the main person and item effects, CD implies the presence of unexplained person-item interactions. To leverage this information from nonbinary item responses, we propose integrating a latent space model with confirmatory factor analysis. The resulting model assumes that persons and items have co-ordinates on a shared metric space called an interaction map, where distances between persons and items reflect CD and their interactions. With this approach, the model quantifies and visualizes person-item interactions, leading to further practical analyses of CD. Our simulation studies demonstrate that the proposed model can recover its parameters well and correctly detect underlying CD. We also provide empirical examples to demonstrate the utilities and advantages of the proposed model, such as (a) deriving personalized diagnoses and evaluations for respondents, (b) quantifying individual differences in perceived item properties, and (c) facilitating investigations of CD with external variables and method effects. (PsycInfo Database Record (c) 2025 APA, all rights reserved).
- Research Article
30
- 10.1007/s11336-021-09762-5
- Jun 1, 2021
- Psychometrika
Classic item response models assume that all items with the same difficulty have the same response probability among all respondents with the same ability. These assumptions, however, may very well be violated in practice, and it is not straightforward to assess whether these assumptions are violated, because neither the abilities of respondents nor the difficulties of items are observed. An example is an educational assessment where unobserved heterogeneity is present, arising from unobserved variables such as cultural background and upbringing of students, the quality of mentorship and other forms of emotional and professional support received by students, and other unobserved variables that may affect response probabilities. To address such violations of assumptions, we introduce a novel latent space model which assumes that both items and respondents are embedded in an unobserved metric space, with the probability of a correct response decreasing as a function of the distance between the respondent's and the item's position in the latent space. The resulting latent space approach provides an interaction map that represents interactions of respondents and items, and helps derive insightful diagnostic information on items as well as respondents. In practice, such interaction maps enable teachers to detect students from underrepresented groups who need more support than other students. We provide empirical evidence to demonstrate the usefulness of the proposed latent space approach, along with simulation results.
- Research Article
6
- 10.12738/estp.2017.2.0246
- Jan 1, 2017
- Educational Sciences: Theory & Practice
Tests used for such purposes as determining educational quality, defining educational needs, hiring an employee, student selection and placement and performing guidance and clinic services have an important place in education and psychology. Of course, they should have certain psychometric features related to test scores' validity and reliability. Various test theories have helped to create more valid and reliable measurements and, as a result, to make better decisions regarding individuals. In education and psychology, Classical Test Theory (CTT) and Item Response Theory (IRT) are both widely used. CTT assumes that an individual's observed score is the total of the true score and the error score, while IRT estimates an individual's ability or latent trait from responses to test items (Embretson & Reise, 2000).When IRT assumptions and model-data fit are ensured, item and ability parameters' invariance occurs; this is known as the most important advantage IRT has over CTT. Item and ability parameters' invariance means estimating ability parameters independently of item sample and estimating item parameters independently of ability sample. IRT's invariance feature makes it very practicable in many applications, for instance, test development, computerized adaptive testing, bias studies, test equating and item mapping (Hambleton & Swaminathan, 1985). IRT is classified under two main categories as parametric IRT (PIRT) and nonparametric IRT (NIRT) (Olivares, 2005; Sijtsma & Molenaar, 2002).To analyse ordered items, such as Likert-type attitude items, partial credit cognitive items or not ordered graded items such as multiple-choice test items, item response models are developed towards polytomous items in IRT (Ostini & Nering, 2006). In these models developed for polytomous items, a non-linear relationship between an individual's latent trait and the possibility of choosing a certain category of item answer is explained (Embretson & Reise, 2000). Graded Response Model (GRM), part of IRT models developed for polytomous items, is often preferred by researchers for applications since it is more useful in presentations, portfolios, essays and Likert-type items with ordered item categories (DeMars, 2010; Ostini & Nering, 2006). To scale tests that consist of polytomous items by making true estimates according to GRM, evaluating PIRT's assumptions and model-data fit is necessary. And to provide these assumptions and model-data fit, large samples are needed. At this point, NIRT models draw attention because they provide a practical advantage in determining psychometric properties of tests with fewer items and respondents (Stout, 2001).NIRT models are defined as statistical scaling methods that require fewer assumptions than PIRT models for measuring persons and items (Stochl, 2007). With their wide application area, NIRT models are used in ordinal scales, applied research areas, sociology, marketing research and health research on quality of life (Sijtsma, 2005). The literature reveals that two models, namely, the Mokken model and nonparametric regression estimation models, are employed. These two models are themselves divided into sub-models. The Mokken model consists of the sub-models Monotone Homogeneity Model (MHM) and the Double Monotonicity Model (DMM). Nonparametric regression estimation models consist of such sub-models as the Kernel Smoothing Approach Model (KSAM), the Isotonic Regression Estimation and the Smoothed Isotonic Regression Estimation models (Lee, 2007; Sijtsma & Molenaar, 2002). Along with theoretical studies being conducted, new sub-models are being added to nonparametric regression estimation models.As a NIRT model, MHM requires unidimensionality, local independence and monotonicity assumptions, and it defines the relationship that latent variables and items with homogeneous (unidimensional) and monotone item characteristic curve (ICC) have (Meijer & Baneke, 2004; Sijtsma & Molenaar, 2002). …
- Research Article
- 10.3390/jintelligence12040038
- Mar 28, 2024
- Journal of Intelligence
This article aims to provide an overview of the potential advantages and utilities of the recently proposed Latent Space Item Response Model (LSIRM) in the context of intelligence studies. The LSIRM integrates the traditional Rasch IRT model for psychometric data with the latent space model for network data. The model has person-wise latent abilities and item difficulty parameters, capturing the main person and item effects, akin to the Rasch model. However, it additionally assumes that persons and items can be mapped onto the same metric space called a latent space and distances between persons and items represent further decreases in response accuracy uncaptured by the main model parameters. In this way, the model can account for conditional dependence or interactions between persons and items unexplained by the Rasch model. With two empirical datasets, we illustrate that (1) the latent space can provide information on respondents and items that cannot be captured by the Rasch model, (2) the LSIRM can quantify and visualize potential between-person variations in item difficulty, (3) latent dimensions/clusters of persons and items can be detected or extracted based on their latent positions on the map, and (4) personalized feedback can be generated from person-item distances. We conclude with discussions related to the latent space modeling integrated with other psychometric models and potential future directions.
- Research Article
22
- 10.1007/s11336-018-9630-0
- Mar 15, 2019
- Psychometrika
Item response theory (IRT) is one of the most widely utilized tools for item response analysis; however, local item and person independence, which is a critical assumption for IRT, is often violated in real testing situations. In this article, we propose a new type of analytical approach for item response data that does not require standard local independence assumptions. By adapting a latent space joint modeling approach, our proposed model can estimate pairwise distances to represent the item and person dependence structures, from which item and person clusters in latent spaces can be identified. We provide an empirical data analysis to illustrate an application of the proposed method. A simulation study is provided to evaluate the performance of the proposed method in comparison with existing methods.
- Single Book
97
- 10.4324/9781315869797
- Jun 26, 2015
This book demonstrates how to conduct latent variable modeling (LVM) in R by highlighting the features of each model, their specialized uses, examples, sample code and output, and an interpretation of the results. Each chapter features a detailed example including the analysis of the data using R, the relevant theory, the assumptions underlying the model, and other statistical details to help readers better understand the models and interpret the results. Every R command necessary for conducting the analyses is described along with the resulting output which provides readers with a template to follow when they apply the methods to their own data. The basic information pertinent to each model, the newest developments in these areas, and the relevant R code to use them are reviewed. Each chapter also features an introduction, summary, and suggested readings. A glossary of the text’s boldfaced key terms and key R commands serve as helpful resources. The book is accompanied by a website with exercises, an answer key, and the in-text example data sets. Latent Variable Modeling with R: -Provides some examples that use messy data providing a more realistic situation readers will encounter with their own data. -Reviews a wide range of LVMs including factor analysis, structural equation modeling, item response theory, and mixture models and advanced topics such as fitting nonlinear structural equation models, nonparametric item response theory models, and mixture regression models. -Demonstrates how data simulation can help researchers better understand statistical methods and assist in selecting the necessary sample size prior to collecting data. -www.routledge.com/9780415832458 provides exercises that apply the models along with annotated R output answer keys and the data that corresponds to the in-text examples so readers can replicate the results and check their work. The book opens with basic instructions in how to use R to read data, download functions, and conduct basic analyses. From there, each chapter is dedicated to a different latent variable model including exploratory and confirmatory factor analysis (CFA), structural equation modeling (SEM), multiple groups CFA/SEM, least squares estimation, growth curve models, mixture models, item response theory (both dichotomous and polytomous items), differential item functioning (DIF), and correspondance analysis. The book concludes with a discussion of how data simulation can be used to better understand the workings of a statistical method and assist researchers in deciding on the necessary sample size prior to collecting data. A mixture of independently developed R code along with available libraries for simulating latent models in R are provided so readers can use these simulations to analyze data using the methods introduced in the previous chapters. Intended for use in graduate or advanced undergraduate courses in latent variable modeling, factor analysis, structural equation modeling, item response theory, measurement, or multivariate statistics taught in psychology, education, human development, and social and health sciences, researchers in these fields also appreciate this book’s practical approach. The book provides sufficient conceptual background information to serve as a standalone text. Familiarity with basic statistical concepts is assumed but basic knowledge of R is not.
- Book Chapter
- 10.36253/978-88-5518-461-8.09
- Jan 1, 2021
Evaluating learners' competencies is a crucial concern in education, and home and classroom structured tests represent an effective assessment tool. Structured tests consist of sets of items that can refer to several abilities or more than one topic. Several statistical approaches allow evaluating students considering the items in a multidimensional way, accounting for their structure. According to the evaluation's ending aim, the assessment process assigns a final grade to each student or clusters students in homogeneous groups according to their level of mastery and ability. The latter represents a helpful tool for developing tailored recommendations and remediations for each group. At this aim, latent class models represent a reference. In the item response theory (IRT) paradigm, the multidimensional latent class IRT models, releasing both the traditional constraints of unidimensionality and continuous nature of the latent trait, allow to detect sub-populations of homogeneous students according to their proficiency level also accounting for the multidimensional nature of their ability. Moreover, the semi-parametric formulation leads to several advantages in practice: It avoids normality assumptions that may not hold and reduces the computation demanding. This study compares the results of the multidimensional latent class IRT models with those obtained by a two-step procedure, which consists of firstly modeling a multidimensional IRT model to estimate students' ability and then applying a clustering algorithm to classify students accordingly. Regarding the latter, parametric and non-parametric approaches were considered. Data refer to the admission test for the degree course in psychology exploited in 2014 at the University of Naples Federico II. Students involved were N=944, and their ability dimensions were defined according to the domains assessed by the entrance exam, namely Humanities, Reading and Comprehension, Mathematics, Science, and English. In particular, a multidimensional two-parameter logistic IRT model for dichotomously-scored items was considered for students' ability estimation.
- Research Article
17
- 10.1177/0146621613491456
- Jul 29, 2013
- Applied Psychological Measurement
The violation of the assumption of local independence when applying item response theory (IRT) models has been shown to have a negative impact on all estimates obtained from the given model. Numerous indices and statistics have been proposed to aid analysts in the detection of local dependence (LD). A Monte Carlo study was conducted to evaluate the relative performance of selected LD measures in conditions considered typical of studies collecting psychological assessment data. Both the Jackknife Slope Index and likelihood ratio statistic G2 are available across the two IRT models used and displayed adequate to good performance in most simulation conditions. The use of these indices together is the final recommendation for applied analysts. Future research areas are discussed.
- Book Chapter
104
- 10.1002/9781118489772.ch30
- Feb 16, 2018
This chapter demonstrates how the Ising model can be estimated. It shows that the Ising model is equivalent to, or closely related to, prominent modeling techniques in psychometrics. The chapter introduces the general class of models used in network analysis called Markov random fields. It also shows that the Ising model is closely related to the modeling framework of item response theory (IRT), which is of central importance to psychometrics. The chapter illustrates how network models could be useful in the conceptualization of psychometric data. Latent variables have played a central role in psychometric models. As psychometrics starts to deal with network models, the Ising model offers a canonical form for network psychometrics, because it deals with binary data and is equivalent to well-known models from IRT. The chapter focuses on the equivalence between the Ising model and multidimensional item response theory.
- Ask R Discovery
- Chat PDF
AI summaries and top papers from 250M+ research sources.