Multiple-choice test items with “none of the above” as an alternative

  • Abstract
  • References
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

Multiple-choice test items with “none of the above” as an alternative

ReferencesShowing 10 of 25 papers
  • Cite Count Icon 13
  • 10.1145/1189136.1189164
A singular choice for multiple choice
  • Jun 26, 2006
  • ACM SIGCSE Bulletin
  • Gudmund S Frandsen + 1 more

  • Cite Count Icon 2
  • 10.1080/00220973.2019.1671299
An Experimental Validation of Sequential Multiple-Choice Tests
  • Oct 11, 2019
  • The Journal of Experimental Education
  • Martin Papenberg + 2 more

  • Cite Count Icon 11
  • 10.1111/j.2044-8317.1993.tb01013.x
In defence of ‘none of the above’
  • Nov 1, 1993
  • British Journal of Mathematical and Statistical Psychology
  • Miguel A García‐Pŕrezt

  • Cite Count Icon 19
  • 10.2466/pr0.1965.16.3c.1193
Effect of Chance Success Due to Guessing on Error of Measurement in Multiple-Choice Tests
  • Jun 1, 1965
  • Psychological Reports
  • Donald W Zimmerman + 1 more

  • Cite Count Icon 17
  • 10.1177/0013164492052003006
A Meta-Analytic Review of Item Discrimination and Difficulty in Multiple-Choice Items Using "None-Of-The-Above"
  • Sep 1, 1992
  • Educational and Psychological Measurement
  • Susan L Knowles + 1 more

  • Open Access Icon
  • PDF Download Icon
  • Cite Count Icon 1
  • 10.1007/s40300-022-00237-w
Theoretical evaluation of partial credit scoring of the multiple-choice test item
  • Jan 6, 2023
  • METRON
  • Rasmus A X Persson

  • Cite Count Icon 19
  • 10.1080/00220973.1991.10806560
The Validity of Two Item-Writing Rules
  • Jan 1, 1991
  • The Journal of Experimental Education
  • Kevin Crehan + 1 more

  • Cite Count Icon 8
  • 10.1207/s15327574ijt0304_3
Multiple Evaluation: A New Testing Paradigm That Exorcizes Guessing
  • Dec 1, 2003
  • International Journal of Testing
  • Arie Dirkzwager

  • Cite Count Icon 7
  • 10.1002/j.0022-0337.1991.55.2.tb02500.x
The option “none of these” improves multiple‐choice test items
  • Feb 1, 1991
  • Journal of Dental Education
  • Rk Kolstad + 1 more

  • Cite Count Icon 12
  • 10.3102/10769986007004333
A Simulation Study of Reliability and Validity of Multiple-Choice Test Scores Under Six Response-Scoring Modes
  • Dec 1, 1982
  • Journal of Educational Statistics
  • Robert B Frary

Similar Papers
  • Research Article
  • Cite Count Icon 20
  • 10.1080/00220973.1974.11011479
Comparative Reliabilities and Difficulties of the Multiple-Choice and True-False Formats
  • Mar 1, 1974
  • The Journal of Experimental Education
  • Albert C Oosterhof + 1 more

The present study was initiated to investigate the comparability of multiple-choice and true-false item formats when the time necessary to respond to each type of item was equated empirically. Also investigated was the relative difficulty of multiple-choice (MC), true true-false (Tf), and false true-false (tF) items mea suring the same content. Results indicated that true-false items result in a less reliable test than one using a four-option MC format, even when empirically determined differences in time needed to answer the respective formats were taken into account. When scores were corrected for guessing, the MC items were significantly easier than the true-false format. CONSIDERABLE DISCUSSION has taken place among measurement specialists regarding the virtues of multiple-choice versus true-false test item for mats. Recent contrasting examples might include .. .the advantages attributed to (true-false) are not, unfortunately, very valid... (3:160), and .. .a few (test specialists) see special virtues of efficiency and ease of preparation in (true-false items) and advocate their wide use (2:1) The most obvious limitation of true-false relative to multiple-choice test items is the degree to which the former is subject to guessing. Several studies have shown that the reliability of a test is directly related to the number of choices per item (1, 4, 5, 6) . Similarly, it would be expected that a multiple choice test would have greater reliability than a true false test if the number of items were held constant. However, since a greater number of true-false items can be administered per unit time, it is possible that in a given amount of time, the increased number of true-false items administered would allow for greater reliability and more efficient sampling of content objectives than had a multiple-choice format been used. Using eighty-eight multiple-choice items from a published test in natural science, Ebel (2) compared formats by rewriting each multiple-choice item as a parallel true-false item. Two forms, each con sisting of forty-four multiple-choice and forty-four true-false items, were developed. Reliabilities (K.R. 20) were computed for the multiple-choice and true-false sections of both forms, and assuming that true-false items could be answered per multiple choice item, the Spearman-Brown formula was used to predict the reliability of an 88-item true false test. For the first form, this adjusted relia bility was greater than the reliability obtained for the multiple-choice section of the test; however, the inverse was true with respect to the second form. The present study was concerned with several currently unanswered questions. First, what is an empirically determined ratio of multiple-choice to equivalent true-false items which can be answered in a given amount of time? Second, for achieve ment test items administered within a classroom situation, which of the two formats under consider ation result in greater reliability per unit of testing time? Third, what is the relative reliability of true true-false and false true-false items when compared to multiple-choice items? Fourth, what ratio of multiple-choice to equivalent true-false items is necessary for producing equal reliability coefficients? Lastly, after equating for differences in the effect of guessing, what is the relative difficulty of the different formats?

  • Research Article
  • Cite Count Icon 33
  • 10.1080/00220671.1990.10885972
Multiple-Choice Test Items: What Are Textbook Authors Telling Teachers?
  • May 1, 1990
  • The Journal of Educational Research
  • Randy A Ellsworth + 2 more

The purposes of this study were to (a) Look at what teachers are being told about multiple-choice test item construction by introductory educational psychology textbook authors and (b) to evaluate the quality of test items that preservice teachers may be exposed to if their university instructors use unedited multiple-choice items provided by the publishers of their classroom text. A comprehensive survey of educational psychology texts was completed to identify the textbook authors’ recommended guidelines for teachers to follow when writing multiple-choice test items. Next, a reduced set of 12 guidelines was identified on the basis of the frequency of author recommendations. These 12 guidelines were used to evaluate 60 multiple-choice items (N = 1,080) that were randomly selected from 18 different instructor guides to introductory educational psychology texts. The results indicated that approximately 60% of the items violated one or more guidelines.

  • Research Article
  • 10.2139/ssrn.3634804
The Effectiveness of Multiple-Choice Test Items in Assessing the Clinical Judgment Abilities of Prelicensure Registered Nursing Students
  • Jul 10, 2020
  • SSRN Electronic Journal
  • Tracy Holt

The purpose of this study was to evaluate the effectiveness of multiple-choice questions in assessing the clinical judgment abilities of prelicensure registered nursing students. Nurse educators are responsible for ensuring fairness in testing and evaluating the development of clinical judgment abilities of students throughout the nursing program of study. This study compared multiple-choice test items developed using the current NCSBN® NCLEX-RN test item writing guidelines to multiple-choice test items developed using the new guidelines proposed by the NCSBN® Next Gen NCLEX-RN® project. The results indicated a statistically significant difference between the two types of multiple-choice items in the measurement of clinical judgment. Participants answered all the items developed according to the current NCSBN® NCLEX-RN test plan correctly within each of the five physiological categories 41% of the time, as compared to only 14% of the time on the items written according to the recommendations of the NCSBN® Nursing Clinical Judgment Model in the Next Gen Project. These findings indicate the participants understood the theoretical knowledge of the physiological concepts but could not apply the theoretical knowledge to the clinical scenario presented in the multiple-choice questions created using the NCSBN® Nursing Clinical Judgment framework. The information gained from this study provides evidence of a process that nurse educators can follow using the NCSBN® Nursing Clinical Judgment model as a guide to create multiple-choice test items to assess the clinical judgment level of nursing students.

  • Research Article
  • Cite Count Icon 7
  • 10.1177/0013164493053003006
Reliability of Comparably Wriften Two-Option Multiple-Choice and True-False Test Items
  • Sep 1, 1993
  • Educational and Psychological Measurement
  • Gregory R Hancock + 3 more

This study compares two-option multiple-choice vocabulary test items with comparably written true-false test items. Results indicate that the multiple-choice format provides a significantly more reliable measure than the true-false format. Results further indicate that the true and false items do not appear to be measuring the same abilities. Potential strategies are discussed that an examinee may use for solving two-option multiple-choice and true-false test items in an effort to explain these results.

  • PDF Download Icon
  • Research Article
  • 10.7176/jep/13-13-07
The Impact of the Number of Distractors in Multiple-choice Test Items on the Psychometric Characteristics of the Items and Item Information Function According to the Two-Parameter Logistic Model in the Item Response Theory
  • May 1, 2022
  • Journal of Education and Practice

The aim of the current study was to identify the impact of the number of distractors in multiple-choice test items on the psychometric characteristics of the test items and the item information function according to the two-parameter logistic model in the Item Response Theory (IRT). To answer the study questions, an achievement test was built covering all school tests course that measures teachers' degree of knowledge to build school tests. The test consisted of (42) items and was of 3 different forms varying in the number of options. The study sample consisted of 356 male and female teachers from Ma'an governorate and Wadi al-Sair district and was selected in a simple random manner. The statistical program (SPSS) and the (Bilog-mg3) program were used to analyse the responses of the study sample. The results of the study showed that there were statistically significant differences between the means of the values of the test items information function in favor of the four-distractors test form compared to the three-and-five- distractors test form. The results, furthermore, showed that there were statistically significant differences between the means of the values of the test items information function in favor of the five-distractors test form compared to the three-distractors test form.The results, further, showed that there were no differences between the means of the estimate of the items difficulty parameter due to the number of distractors of the item. It was indicated that there were differences between the means of the estimate of the items discrimination parameter concerning the three-and-four-distractors tests in favor of the four-distractors test. It was, besides, shown that there were statistically significant differences between the means of the estimate of the items discrimination parameter concerning the three-distractors and five- distractors tests’ forms in favor of the for the five- distractors test. Keywords: number of distractors, multiple-choice test, item response theory, items’ parameters, information function DOI: 10.7176/JEP/13-13-07 Publication date: May 31 st 2022

  • Research Article
  • 10.12816/0014338
تصحيح معاملات صعوبة الفقرات لأثر التخمين في أسئلة الاختيار من متعدد : صورة معدلة لمعادلة جيلفورد = Revision of Guilford Formula to Correct Item Difficulty for Guessing in Multiple Choice Test Items
  • Apr 1, 2014
  • Journal of Educational and Psychological Studies
  • أحمد سليمان عودة

صعوبة الفقرة | التصحيح لأثر التخمين | المخاطرة في الإجابة | فقرات الإختيار من متعدد، المعرفة الجزئية | معادلة التصحيح | Item Difficulty | Correction for Guessing | Risk - Taking | Multiple Choice Items | Partial Knowledge | Scoring Formula

  • Research Article
  • 10.18260/edgj.v82i2.686
Evolution of Test Items: From Open-ended to Multiple-Choice
  • May 19, 2019
  • The Engineering Design Graphics Journal
  • Mary A Sadowski + 1 more

Grading is often a faculty member’s least favorite chore, especially in engineering where open-ended problems prevail. For this reason, multiple-choice test items could be a popular alternative for assessing learning and understanding. In addition, most Learning Management Systems allow the instructor to cre¬ate multiple-choice questions to be scored automatically by the system. The use of multiple-choice items in engineering graphics education could increase efficiency, allowing instructors to focus on other aspects of their teaching rather than spending significant time grading open-ended problems. The authors of this paper have been involved in a project to develop a Concept Inventory for Engineering Graphics over the past several years. Since Concept Inventories typically consist of multiple-choice items, development of this instrument was reliant on the creation of numerous valid and reliable items. This paper will focus on the process employed in multiple-choice item creation with application to engineering graphics. The process will be illustrated through demonstration of item evolution through several iterations.

  • Research Article
  • 10.22111/ijals.2018.4187
The Effects of Presenting Multiple-Choice Test Items in Oral and Written Modes and Item Types on Advanced EFL Learners’ Listening Comprehension and Perception
  • Oct 1, 2017
  • Iranian Journal of Applied Language Studies
  • Mohammad Taghi Farvardin + 2 more

This quasi-experimental study aimed to compare the effect of different modes and item types of multiple-choice (MC) test items on advanced EFL learners’ listening comprehension and perception. To this end, 80 advanced EFL learners, aging 18 to 30, were selected. The participants took a listening test including dialogue-completion and question and answer multiple-choice items presented in written and oral modes. In addition, the participants were given a questionnaire on their perceptions of the oral and written modes. The results of two-way repeated measures ANOVA showed that there was no significant difference between the participants’ scores in the oral and written modes. Moreover, they received similar scores on the two item types. However, in their questionnaire responses, most of the participants preferred the written mode to the oral mode. The results also imply that although presenting the multiple-choice test items in the oral mode may form a pure test of listening comprehension, there are acceptable reasons for presenting the multiple-choice listening items in the written mode.

  • Research Article
  • 10.70382/tijerls.v07i8.045
PSYCHOMETRIC PROPERTIES OF MULTIPLE-CHOICE TEST ITEMS IN HISTORY SUBJECT FOR ASSESSING STUDENTS IN SECONDARY SCHOOLS IN NASARAWA STATE, NIGERIA
  • May 31, 2025
  • International Journal of Educational Research and Library Science
  • Daniel Edward Tanzegyo

The study focused on development and determination of psychometric properties of multiple-choice test items for students in history in secondary schools. Two research questions and one hypothesis were involved in the study It utilized the instrumentation research design and was carried out in Nasarawa State, Nigeria. The population for the study was 248 SSIII history students using three ability groups (36 high, 70 average and 42 low abilities). There was no sampling because the population was manageable. Sixty (60) multiple-choice test items were developed from the literature reviewed and WAEC history curriculum and used in the study. The test was validated by three experts, two in the Department of History Education, and one from the Department of Measurement and Evaluation, all in the University of Nigeria, Nsukka. The test was used in assessing students in the four secondary schools in the study area. The reliability of the items was determined using test-retest method and Kudder-Richardson (K-R21). This yielded co-efficient of .86 depicting that the instrument is reliable. The Analysis of variance (ANOVA) was utilized to test the hypothesis at 0.05 level of significant. It was found out that the items had difficulty indices of between 0.30 and 0.70. It was therefore recommended that the examination bodies (WAEC and NECO) should adopt the psychometric multiple-choice test items in their examination for certification of students at the secondary school level.

  • Research Article
  • Cite Count Icon 127
  • 10.3928/0148-4834-20010101-06
Writing multiple-choice test items that promote and measure critical thinking.
  • Jan 1, 2001
  • Journal of Nursing Education
  • Susan Morrison + 1 more

Faculties are concerned about measurement of critical thinking especially since the National League for Nursing Accrediting Commission cited such measurement as a requirement for accreditation (NLNAC, 1997). Some writers and researchers (Alfaro-LeFevre, 1995; Blat, 1989; McPeck, 1981, 1990) describe the need to measure critical thinking within the context of a specific discipline. Based on McPeck's position that critical thinking is discipline-specific, guidelines for developing multiple-choice test items as a means of measuring critical thinking within the discipline of nursing are discussed. Specifically, criteria described by Morrison, Smith, and Britt (1996) for writing critical-thinking multiple-choice test items are reviewed and explained for promoting and measuring critical thinking.

  • Research Article
  • Cite Count Icon 1
  • 10.21031/epod.1107597
Latent Growth Modeling of Item Process Data Derived From Eye-tracking Technology: An Experimental Study Investigating Reading Behavior of Examinees When Answering A Multiple-Choice Test Item
  • Sep 30, 2022
  • Eğitimde ve Psikolojide Ölçme ve Değerlendirme Dergisi
  • Ergün Cihat Çorbaci + 1 more

This study illustrates how eye-tracking data can be translated to “item process data” for multiple-choice test items to study the relationship between subjects’ item responses and choice reading behavior. Several modes of analysis were used to test the hypothesized added value of using process data to identify choice reading patterns of subjects. In addition to the cross-sectional analises of agggregate measurements derived from the time series eye tracking data, Latent Growth Curve Model Analises were undertaken to test if the the shape of change observed in the sequential choice reading patterns differed for subjects depending on their responses to the item being correct or incorrect. Application data were from an experimental study and included seventy-one subjects’ responses to two multiple-choice test items measuring reading comprehension ability in English as a second language. Analyses were carried out for one item at a time. For each item, first, each subject’s recorded eye movements were coded into a set of Area of Interests (AOIs), segmenting the lines in the stem and the individual choices. Next, each subject’s fixation times on the AOIs were time stamped into seconds, indicating when and in what order each subject’s gaze had fixated on each AOI until a choice was marked as the correct answer, which ended the item encounter. A set of nested Latent Growth Curve models were considered for the choice-related AOIs to deliniate if distinct choice-process sequences were evident for correct and incorrect respoders. Model fit indices, random intercepts, slopes, and residuals were computed using the mean log fixation times over item encounter time. The results show that the LGM with the best model fit indicies, for both items, was the quadratic model using response variable as a covariate. Albeit limited due to the two-item – seventy-one subjects experimental setting of the study, the findings are promising and show that utilizing item-level process data can be very useful for defining distinct choice processing (task-oriented reading) patterns of examinees. Over all, the results warrant further study of choice derived AOIs using longitudinal statistical models. It is argued that, the screening methodology desribed in this study can be a useful tool to investigate speededness, distractor functioning, or even to flag subjects with irregular choice processing behavior, such as providing a direct mark on a choice, without any significant reading activity on any of the choices presented (i.e., whether cheating might have occurred.)

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 51
  • 10.3389/fpsyg.2019.00825
Multiple-Choice Item Distractor Development Using Topic Modeling Approaches.
  • Apr 25, 2019
  • Frontiers in Psychology
  • Jinnie Shin + 2 more

Writing a high-quality, multiple-choice test item is a complex process. Creating plausible but incorrect options for each item poses significant challenges for the content specialist because this task is often undertaken without implementing a systematic method. In the current study, we describe and demonstrate a systematic method for creating plausible but incorrect options, also called distractors, based on students’ misconceptions. These misconceptions are extracted from the labeled written responses. One thousand five hundred and fifteen written responses from an existing constructed-response item in Biology from Grade 10 students were used to demonstrate the method. Using a topic modeling procedure commonly used with machine learning and natural language processing called latent dirichlet allocation, 22 plausible misconceptions from students’ written responses were identified and used to produce a list of plausible distractors based on students’ responses. These distractors, in turn, were used as part of new multiple-choice items. Implications for item development are discussed.

  • Research Article
  • Cite Count Icon 12
  • 10.1016/s2155-8256(15)30217-9
Best Practices for Writing Test Items
  • Jul 1, 2012
  • Journal of Nursing Regulation
  • Karen Sutherland + 2 more

Best Practices for Writing Test Items

  • Research Article
  • Cite Count Icon 1
  • 10.1080/00949659308811519
C102. Second attempts at multiple-choice test items
  • Jun 1, 1993
  • Journal of Statistical Computation and Simulation
  • T P Hutchinson

(1993). C102. Second attempts at multiple-choice test items. Journal of Statistical Computation and Simulation: Vol. 47, No. 1-2, pp. 108-112.

  • Research Article
  • 10.36941/mjss-2021-0013
Comparison of the Effectiveness of Four Scoring Techniques in Secondary School Multiple-Choice Test Items in Financial Accounting
  • May 17, 2021
  • Mediterranean Journal of Social Sciences
  • Olaitan Isaac Fadipe + 1 more

Poor academic performance of students in senior secondary school certificate examination in some commercial subjects may be attributed to ineffective scoring techniques used by classroom teachers and examination bodies. Therefore, this study compared the effectiveness of four scoring techniques in multiple-choice financial accounting in secondary schools of Osun state, Nigeria. The study adopted a descriptive survey and Quasi-experimental research designs. The sample consisted of 420 senior secondary schools (SSSII) Financial Accounting students selected using Multi-stage sampling technique. An instrument titled “Financial Accounting Multiple-Choice Test (FAMT)” with a reliability coefficient of .80 was administered to selected students. The marks obtained through standard conventional, logical weight, corrected and confidence scoring techniques were analyzed using descriptive and inferential statistics. Findings showed that the logical weight scoring technique had the highest mean value of 40.63 when compared with other scoring techniques. It was also revealed that there was a significant difference in the performance of students in Financial Accounting multiple-choice test items using the four scoring techniques (F(3,1676) =31.494, p<.05). Based on the findings, it was concluded that students performed better using the logical weight scoring technique in Financial Accounting multiple-choice test items and is the most effective scoring technique among others. The study recommended that a logical weight scoring technique should be encouraged and used in scoring Financial Accounting multiple-choice test items in secondary schools by examination bodies.
 
 Received: 27 March 2021 / Accepted: 29 April 2021 / Published: 17 May 2021

More from: METRON
  • New
  • Research Article
  • 10.1007/s40300-025-00299-6
Bayesian analysis and optimal life testing for new Pareto distribution under progressive censoring
  • Nov 25, 2025
  • METRON
  • Prakash Chandra + 2 more

  • Research Article
  • 10.1007/s40300-025-00298-7
Node splitting SVMs for survival trees based on an $$L_2$$-regularized dipole splitting criteria
  • Nov 9, 2025
  • METRON
  • Aye Aye Maung + 2 more

  • Research Article
  • 10.1007/s40300-025-00296-9
A decomposition of the changes in seasonal concentration by tourist segment: an analysis of the Veneto region
  • Aug 22, 2025
  • METRON
  • Mauro Mussini

  • Research Article
  • 10.1007/s40300-025-00295-w
Effect Size Estimation in Linear Mixed Models
  • Aug 11, 2025
  • METRON
  • Jürgen Groß + 1 more

  • Research Article
  • 10.1007/s40300-025-00294-x
Multiple-choice test items with “none of the above” as an alternative
  • Jun 16, 2025
  • METRON
  • Rasmus A X Persson

  • Open Access Icon
  • Research Article
  • 10.1007/s40300-025-00292-z
Finite mixture copulas for modeling dependence in longitudinal count data
  • Apr 22, 2025
  • METRON
  • Subhajit Chattopadhyay

  • Research Article
  • 10.1007/s40300-025-00293-y
Empirical Bayes methods in high dimensions: a survey and ongoing debates
  • Apr 21, 2025
  • METRON
  • Stefano Rizzelli

  • Research Article
  • 10.1007/s40300-025-00289-8
Discrete examples of sufficient or non-sufficient statistic $$\left| X\right| $$ for unknown $$\mathbf {\theta }$$ via conditioning
  • Mar 5, 2025
  • METRON
  • Nitis Mukhopadhyay

  • Open Access Icon
  • Research Article
  • 10.1007/s40300-025-00291-0
Editorial to the Special Issue on “New challenges for sample surveys: innovation through tradition”
  • Mar 3, 2025
  • METRON
  • Pier Francesco Perri + 2 more

  • Research Article
  • 10.1007/s40300-025-00290-1
Modeling heavy-tails with two-piece Burr distributions via conditional values-at-risk
  • Feb 25, 2025
  • METRON
  • Johan René Van Dorp + 1 more

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.

Search IconWhat is the difference between bacteria and viruses?
Open In New Tab Icon
Search IconWhat is the function of the immune system?
Open In New Tab Icon
Search IconCan diabetes be passed down from one generation to the next?
Open In New Tab Icon