Year Year arrow
arrow-active-down-0
Publisher Publisher arrow
arrow-active-down-1
Journal
1
Journal arrow
arrow-active-down-2
Institution Institution arrow
arrow-active-down-3
Institution Country Institution Country arrow
arrow-active-down-4
Publication Type Publication Type arrow
arrow-active-down-5
Field Of Study Field Of Study arrow
arrow-active-down-6
Topics Topics arrow
arrow-active-down-7
Open Access Open Access arrow
arrow-active-down-8
Language Language arrow
arrow-active-down-9
Filter Icon Filter 1
Year Year arrow
arrow-active-down-0
Publisher Publisher arrow
arrow-active-down-1
Journal
1
Journal arrow
arrow-active-down-2
Institution Institution arrow
arrow-active-down-3
Institution Country Institution Country arrow
arrow-active-down-4
Publication Type Publication Type arrow
arrow-active-down-5
Field Of Study Field Of Study arrow
arrow-active-down-6
Topics Topics arrow
arrow-active-down-7
Open Access Open Access arrow
arrow-active-down-8
Language Language arrow
arrow-active-down-9
Filter Icon Filter 1
Export
Sort by: Relevance
  • New
  • Research Article
  • 10.1177/07342829251414532
Examining the Measurement Invariance of TIMSS Science Self-Concept Scales Across Genders, Cultures, and Languages
  • Jan 3, 2026
  • Journal of Psychoeducational Assessment
  • Ying Liu + 1 more

This study employed item response theory (IRT), exploratory factor analysis (EFA), and confirmatory factor analysis (CFA) to examine the measurement invariance of TIMSS science self-concept (SSC) scales for grades 4 and 8 across genders, cultures, and languages in six countries and regions. IRT, and EFA and CFA were utilized to examine the reliability and validity of the SSC scales, respectively. The IRT results revealed substantial variation in each item and test information function values across the six societies. The EFA and CFA results identified two dimensions of SSC scales, that is, positive perception of SSC (SSCp) and negative perception of SSC (SSCn), with satisfactory fit indices. Measurement invariance was confirmed across genders, and cultural and linguistic groups. Recommendations are put forward for TIMSS program developers and researchers regarding the revision of the SSC scales.

  • Research Article
  • 10.1177/07342829251404011
Structural Validity of the Knowledge for Teaching Early Elementary Mathematics (K-TEEM) Assessment
  • Dec 2, 2025
  • Journal of Psychoeducational Assessment
  • Robert C Schoen + 1 more

This study examines the structural validity of the 2019 Teacher MKT assessment using empirical data from a sample of 645 elementary educators in Florida (USA). The examination of structural validity was supported through a series of complementary analyses, including evaluation of the dimensionality of the underlying latent trait, item performance, reliability, and the standard error of measurement. Evidence from dimensionality analysis pointed to a borderline unidimensional structure, implying that one factor or dimension may be sufficient for measuring the underlying construct. Based on the empirical results, the 2019 Teacher MKT appears to have adequate reliability and provides useful information about a relatively wide range of teacher abilities.

  • Research Article
  • 10.1177/07342829251382279
Overview of the Woodcock-Johnson V Tests of Cognitive Abilities and Virtual Test Library
  • Nov 29, 2025
  • Journal of Psychoeducational Assessment
  • Dawn P Flanagan + 2 more

The Woodcock-Johnson V (WJ V; McGrew, Mather, LaForte, et al., 2025) represents the most significant evolution in comprehensive cognitive assessment since the battery was introduced nearly 50 years ago. This article provides an overview of the WJ V Tests of Cognitive Abilities and Virtual Test Library, highlighting key revisions from the WJ IV and examining psychometric properties and interpretive options. The WJ V assessment system comprises 60 co-normed tests across cognitive, achievement, and language domains, making it the most comprehensive assessment battery currently available. Major innovations include a fully digital platform, contemporary post-pandemic norms, enhanced theoretical alignment with current Cattell-Horn-Carroll (CHC) theories, and expanded diagnostic capabilities through the new Virtual Test Library. The WJ V Tests of Cognitive Abilities contains 20 tests organized into 17 clusters, including three general intelligence measures, seven CHC broad ability clusters, and six narrow ability/clinical clusters. Significant structural changes include the creation of a new Retrieval Fluency (Gr) broad ability cluster, the removal of Auditory Processing (Ga) from the General Intellectual Ability (GIA) cluster, substantial revisions to the GIA composition, and the relocation of specialized tests to the Virtual Test Library, which incorporates 15 specialized tests, enabling the targeted evaluation of phonological awareness, rapid automatized naming, and memory functions critical for the identification of dyslexia. Extensive validity evidence supporting the battery’s structure was obtained through a sophisticated three-stage framework employing multiple analytical methods, including exploratory factor analysis, multidimensional scaling, cluster analysis, and psychometric network analysis. Clinical validity studies demonstrate the battery’s effectiveness in differentiating various neurodevelopmental conditions. The WJ V’s four-level interpretive framework emphasizes criterion-referenced proficiency measures alongside traditional norm-referenced scores, providing educationally relevant information for intervention planning. This comprehensive assessment system offers practitioners unprecedented flexibility for hypothesis-driven evaluation while maintaining the psychometric rigor essential for high-stakes educational and diagnostic decisions, positioning it as a transformative tool for evidence-based assessment practices in the digital age.

  • Open Access Icon
  • Research Article
  • 10.1177/07342829251395781
Test Review: Woodcock–Johnson V Tests of Cognitive Abilities
  • Nov 15, 2025
  • Journal of Psychoeducational Assessment
  • Laree B Foster + 2 more

The Woodcock–Johnson V Tests of Cognitive Abilities (WJ V COG), published in February 2025, offers the latest edition of the WJ family of tests alongside tests of academic achievement and oral language. The WJ V COG has changed substantially from previous editions regarding administration, which is now entirely digital. Administration and scoring are housed within the Riverside Insight’s online platform. The test battery features several changes, such as the addition of five new tests and the removal of three tests, including measures of Auditory Processing (Ga). The WJ V COG maintains a CHC-based theoretical framework, although updated to align with current theory. Psychometric evidence, including validity, reliability, and item-level analysis, is robust. Evidence is less convincing for children under six. The assessment was co-normed with measures of academic achievement, and the norm sample was gathered post-COVID. Although some may find requirements of digital administration limiting, the WJ V COG offers an engaging and psychometrically sound option for the assessment of intelligence.

  • Open Access Icon
  • Research Article
  • 10.1177/07342829251393579
Construct Validity of the Scandinavian Version of the Cognitive Assessment System
  • Nov 11, 2025
  • Journal of Psychoeducational Assessment
  • Niklas Fröst + 2 more

Purpose: This study aimed to evaluate the psychometric properties of the Cognitive Assessment System ver.2 in Swedish and Norwegian contexts (CAS-2). Methodology: A total of 614 students (aged 5 to 18 years) were assessed using the CAS-2. Confirmatory factor analysis was conducted to investigate the construct validity of three competing models of the Scandinavian version of the CAS-2: a four-factor, a three-factor, and a bifactor model. The measurement invariance was tested for age and sex. Key Findings: The results indicated an acceptable to good fit for all three models. In addition, measurement invariance across sex and age groups was also supported. However, with respect to the four-factor model, the two factors (Planning and Attention) were strongly correlated, raising a discussion on the empirical and theoretical distinctiveness of these factors. Conclusions/Implications: These findings suggest that all three tested models contribute uniquely to understanding the CAS-2 but also underscore the need for further refinement and caution when drawing conclusions about cognitive profiles in the present version.

  • Research Article
  • 10.1177/07342829251395913
Longitudinal Measurement Invariance Testing of the Model Minority Stereotype Response Scale Among Asian American STEM Students
  • Nov 10, 2025
  • Journal of Psychoeducational Assessment
  • Han Na Suh + 1 more

This study examined the temporal stability of the Model Minority Stereotype-Response Scale (MMS-RS) in assessing Asian American STEM students’ responses to the model minority stereotype (MMS) over time. Data from 188 Asian American STEM students were analyzed using longitudinal measurement invariance testing. The findings showed that the subscale scores of the MMS-RS, Negative Response to Stress (NRS) and Positive Response and Endorsement (PRE), reliably assess negative and positive responses across time, respectively, and can be meaningfully tracked for their changes over time, although one item in each subscale was not invariant. The residual invariance finding suggested that observed score differences reflect true changes. Overall, the MMS-RS subscale scores proved to be stable and reliable for measuring responses to the MMS. Specifically, the findings from the longitudinal measurement invariance suggest that the comparison of mean scores across time is viable. Thus MMS-RS subscales can be useful in tracking changes in response to interventions or policy changes that are aimed at supporting Asian American STEM students who experience the MMS.

  • Open Access Icon
  • Research Article
  • 10.1177/07342829251386972
The Winston Essential Skills Survey: A Comprehensive Social-Emotional Assessment for Students With Learning Disabilities
  • Nov 5, 2025
  • Journal of Psychoeducational Assessment
  • Amber Debono + 4 more

The Winston Essential Skills Survey (WESS) 3.0, the first social-emotional learning (SEL) assessment designed for students with learning disabilities (SwLD), has evidence for reliability and validity in a specific population of teachers. To build on this evidence, 180 educators responded to the WESS 3.0, along with new items and the Devereux Student Strengths Assessment (DESSA), to assess a sample of 397 SwLD. Minor modifications to the WESS 3.0 resulted in the WESS 4.0, which demonstrated internal consistency, an eight-factor structure, and convergent validity with the DESSA. The WESS 4.0 addresses a critical gap in SEL assessments for SwLD, providing researchers and educators with a valid and reliable method to assess SwLD's social-emotional skills. This survey may inform school interventions, policy decisions on SEL programming, and advance understanding of SEL in SwLD, potentially improving educational outcomes and well-being for this population.

  • Open Access Icon
  • Research Article
  • 10.1177/07342829251390697
A Scoping Review and Psychometric Assessment of Survey Instruments Used to Evaluate Post-secondary Student Stress
  • Nov 3, 2025
  • Journal of Psychoeducational Assessment
  • Brooke Linden + 1 more

Post-secondary students are exposed to a wide variety of stressors, and mental health-related concerns continue to increase among this population. The development and improvement of mental health services hinges on the use of valid and reliable measurement tools to assess student stress and evaluate the efficacy of interventions. To date, there has been no comprehensive overview of available instruments intended to evaluate post-secondary student stress. The purpose of this study was to conduct a rigorous scoping review of published works detailing the use of instruments to evaluate post-secondary student stress, and evaluate the psychometric properties of these tools as well as their suitability for use among post-secondary populations. Four academic databases were searched, including HaPI, CINAHL, Medline, and PsycINFO, including peer-reviewed records published in English between 2012 and 2024. Articles were independently screened for inclusion by title, abstract, and full-text review. A total of 62 records and 11 instruments were screened into the review. The majority of instruments included in our review did not meet the criteria for suitability, nor did they have published, comprehensive psychometric evaluations. The Post-Secondary Student Stressors Index emerged as the most appropriate instrument for evaluating post-secondary student-specific stress owing to both its suitability and strong psychometric properties. The Perceived Stress Scale and Stress Overload Scale were deemed to be the most appropriate instruments for evaluating general stress levels.

  • Research Article
  • Cite Count Icon 1
  • 10.1177/07342829251392959
Psychometric Properties of the Chinese Version of the Performance Failure Appraisal Inventory-Short Form (PFAI-SF)
  • Oct 28, 2025
  • Journal of Psychoeducational Assessment
  • Yikang Chen + 3 more

Fear of failure undermines individuals’ motivation and well-being globally. However, measuring fear of failure remains challenging due to the lack of brief questionnaires, and existing measures have primarily been developed and used in Western contexts. Hence, engaging in cross-cultural validation of fear of failure measures is needed. This study aimed to explore the psychometric properties of the Chinese version of the Performance Failure Appraisal Inventory-Short Form (PFAI-SF). Exploratory factor analysis (EFA), confirmatory factor analysis (CFA), concurrent validity, and convergent validity were investigated with 1670 Chinese students. Results of the EFA and CFA indicated satisfactory psychometric properties, suggesting a unifactorial model. The concurrent and convergent validity of the Chinese version of PFAI-SF was supported by positive correlations with the long-form Performance Failure Appraisal Inventory and its related psychological constructs. These findings provide substantial evidence for the psychometric properties of the Chinese version of the PFAI-SF, supporting its use as a succinct and culturally appropriate tool for assessing fear of failure among Chinese students.

  • Research Article
  • 10.1177/07342829251393575
Using the Bayesian Network Method to Evaluate the Effectiveness of College Students’ Mental Health Intervention Strategies and Their Impact on Academic Performance
  • Oct 28, 2025
  • Journal of Psychoeducational Assessment
  • Wang Xiaohui + 1 more

Mental health and academic success are increasingly interdependent challenges for university students worldwide. This study developed and validated dynamic Bayesian models to predict academic performance and psychological risk across semesters using probabilistic approaches. We analyzed a cohort of 3,276 undergraduates and externally validated findings against an independent cohort of 5,112 students. Dynamic Bayesian Networks (DBN) and Bayesian Networks (BN) were trained using psychological scores (PHQ-9, GAD-7, PSS-10, CD-RISC) to model psychological risk and academic records to model academic outcomes. Ten-fold temporal cross-validation was conducted internally, and comparative analyses involved Random Forests, XGBoost, Deep Neural Networks, and TabTransformer models. DeLong’s tests compared AUCs and permutation tests assessed Brier scores. Internally, BN achieved 91.0% accuracy, an AUC of 0.84 (95% CI 0.81–0.87), and a Brier score of 0.128, while DBN achieved 94.2% accuracy, an AUC of 0.86 (95% CI 0.84–0.89), and a Brier score of 0.124. In external validation, BN achieved 90.0% accuracy and an AUC of 0.88 (95% CI 0.85–0.90), and DBN achieved 92.0% accuracy and an AUC of 0.91 (95% CI 0.88–0.93). Top predictors included GPA, stress scores, depression scores, and intervention engagement. Posterior predictive p-values exceeded 0.44 across GPA and both outcome domains, indicating adequate calibration. Dynamic Bayesian modeling enables accurate, uncertainty-resilient prediction of both psychological risk and academic outcomes among university students.