Year
Publisher
Journal
1
Institution
Institution Country
Publication Type
Field Of Study
Topics
Open Access
Language
Filter 1
Year
Publisher
Journal
1
Institution
Institution Country
Publication Type
Field Of Study
Topics
Open Access
Language
Filter 1
Export
Sort by: Relevance
A structural equation investigation of linguistic features as indices of writing quality in assessed secondary-level EMI learners’ scientific reports

While inquiry into the relationship between linguistic features and L2 writing quality has been a long-standing line of research, little scholarly attention has been drawn to the predictive value of linguistic features in assessing the writing quality of English-medium scientific report writing. This study adds to the existing literature by examining the relation of lexical and syntactic complexity to writing quality, based on 106 scientific reports composed by Hong Kong Chinese learners of English in EMI secondary schools. Natural language processing tools were employed to extract computational indices of linguistic complexity features, followed by the use of a structural equation modeling (SEM) approach to investigate their predictive power. The validity of the anticipated construct was confirmed based upon several goodness-of-fit criteria. The SEM analysis indicated that writing quality was predicted by lexical sophistication (i.e., text-based complexity: word range and academic words; psycholinguistic complexity: word familiarity and age-of-acquisition ratings), lexical diversity (i.e., MTLD and VocD), and syntactic complexity (i.e., mean length of sentence and dependent clauses per T-unit). However, the relation of lexical diversity and syntactic complexity to writing quality was mediated by lexical sophistication. Implications for scientific report writing assessment and pedagogy in EMI educational contexts are discussed.

Read full abstract
A comparative study of voice in Chinese English-major undergraduates’ timed and untimed argument writing

As a somewhat elusive and occlusive concept, voice can be a challenging and formidable hurdle for second language (L2) writers. One area that exemplifies this struggle is timed argument writing, where authors must position claims, ideas, and individual perspectives to an existing knowledge base and scholarly community under the confines of time. To enrich our understandings of voice construction in L2 English writers’ timed writing, we explored how 41 Chinese English-major undergraduates deployed authorial voice in two prompt-based argument writing tasks (timed versus untimed). We also sampled their self-reported knowledge, use, and understanding of voice through a survey-based instrument. To compare the quantity and quality of voice construction between the two samples, we measured 10 voice categories, three voice dimensions, and overall voice strength. Results showed that only two categories displayed statistically significant differences in terms of frequencies, but all three voice dimensions and overall voice strength scored significantly higher in untimed writing samples. Based on the results of our text analysis and survey, we further highlight the complexities of voice in L2 writing, provide evidence in support of existing voice rubrics, and make practical suggestions for teaching and assessing voice in writing.

Read full abstract
Detecting and assessing AI-generated and human-produced texts: The case of second language writing teachers

Artificial intelligence (AI) technologies have recently attracted the attention of second language (L2) writing scholars and practitioners. While they recognize the tool’s viability, they also raised the potential adverse effects of these tools on accurately reflecting students’ actual level of writing performance. It is, therefore, crucial for teachers to discern AI-generated essays from human-produced work for more accurate assessment. However, limited information is available about how they assess and distinguish between essays produced by AI and human authors. Thus, this study analyzed the scores and comments teachers gave and looked into their strategies for identifying the source of the essays. Findings showed that essays by a native English-speaking (NS) lecturer and ChatGPT were rated highly. Meanwhile, essays by an NS college student, non-native English-speaking (NNS) college student, and NNS lecturer scored lower, which made them distinguishable from an AI-generated text. The study also revealed that teachers could not consistently identify the AI-generated text, particularly those written by an NS professional. These findings were attributed to teachers’ past engagement with AI writing tools, familiarity with common L2 learner errors, and exposure to native and non-native English writing. From these results, implications for L2 writing instruction and future research are discussed.

Read full abstract
The impact of task duration on the scoring of independent writing responses of adult L2-English writers

In writing assessment, there is inherently a tension between authenticity and practicality: tasks with longer durations may more closely reflect real-life writing processes but are less feasible to administer and score. What is more, given total testing time, there is necessarily a trade-off between task duration and number of tasks. Traditionally, high-stakes assessments have managed this trade-off by administering one or two writing tasks each test, allowing 20–40 minutes per task. However, research on second language (L2) English writing has not found longer task durations to significantly improve score validity or reliability. Importantly, very few studies have compared much shorter durations for writing tasks to more traditional allotments. To explore this issue, we asked adult L2-English test takers to respond to two writing prompts with either 5-minute or 20-minute time limits. Responses were then evaluated by expert human raters and an automated writing evaluation tool. Regardless of scoring method, short duration scores evidenced equally high test-retest reliability and criterion validity as long duration scores. As expected, longer task duration yielded higher scores, but regardless of duration, test takers demonstrated the entire spectrum of writing proficiency. Implications for writing assessment are discussed in relation to scoring practices and task design.

Read full abstract
Open Access
Validating an integrated reading-into-writing scale with trained university students

Integrated tasks are often used in higher education (HE) for diagnostic purposes, with increasing popularity in lingua franca contexts, such as German HE, where English-medium courses are gaining ground. In this context, we report the validation of a new rating scale for assessing reading-into-writing tasks. To examine scoring validity, we employed Weir’s (2005) socio-cognitive framework in an explanatory mixed-methods design. We collected 679 integrated performances in four summary and opinion tasks, which were rated by six trained student raters. They are to become writing tutors for first-year students. We utilized a many-facet Rasch model to investigate rater severity, reliability, consistency, and scale functioning. Using thematic analysis, we analyzed think-aloud protocols, retrospective and focus group interviews with the raters. Findings showed that the rating scale overall functions as intended and is perceived by the raters as valid operationalization of the integrated construct. FACETS analyses revealed reasonable reliabilities, yet exposed local issues with certain criteria and band levels. This is corroborated by the challenges reported by the raters, which they mainly attributed to the complexities inherent in such an assessment. Applying Weir’s (2005) framework in a mixed-methods approach facilitated the interpretation of the quantitative findings and yielded insights into potential validity threads.

Read full abstract
Exploring the use of model texts as a feedback instrument in expository writing: EFL learners’ noticing, incorporations, and text quality

Model texts as a feedback instrument (MTFI) have proven effective in enhancing L2 writing, yet research on this domain mainly focused on narrative compositions over a three-stage task: i) composing, ii) comparing, and iii) rewriting. The impact of MTFI on learners’ noticing, incorporations, and text quality in expository writing, especially in the Vietnamese context, remains underexplored. To address these gaps, this study aims to investigate the effect of MTFI on 68 Vietnamese EFL undergraduates’ expository writing following a process-product approach. The participants were divided into a control group (CG, N = 33) and an experimental group (EG, N = 35). Both groups attended stages one and three, but only the EG compared their initial writing with a model text in stage two. The results, derived from learners’ note-taking sheets, written paragraphs, and semi-structured interviews, revealed that despite the two groups’ comparability in stage one, the EG demonstrated significantly better text quality than the CG in stage three, particularly in content, lexis, and organization. Furthermore, while the EG largely encountered lexical issues at the outset, they primarily concentrated on content-related and organizational features in the subsequent stages. Based on the findings, recommendations for future research and implications for pedagogy were deliberated.

Read full abstract
L2 master’s and doctoral students’ preferences for supervisor written feedback on their theses/dissertations

The present study employed a qualitative research design to investigate possible differences between L2 master’s and doctoral students’ preferences for supervisor written feedback. Although the role of learners’ preferences, as a part of attitudinal engagement, has been emphasized in the literature on feedback, there are still niches in the literature that need to be occupied. One of these gaps is the examination of L2 master’s and doctoral students’ preferences for supervisor written feedback on their theses/dissertations. To bridge this research gap, the researcher interviewed 52 master’s and 21 doctoral Iranian English Language Teaching students. Thematic analysis of the interview data identified five main preferences: feedback that is clear, specific, encouraging, dialogic, and non-appropriative. The examination of interview data showed that both master’s and doctoral students expressed high levels of preference for receiving clear and encouraging feedback. A significantly higher percentage of master’s students expressed their preference for specific comments. In contrast, doctoral students exhibited heightened preferences for non-appropriative and dialogic feedback. The findings also provided insights into the underlying factors that can shape master’s and doctoral students’ preferences. Several practical implications and suggestions for further research are also discussed in this study.

Read full abstract