You can view the full text of this article for free using the link below.

Title: Determining item position effects in a computer-based test

Authors: Gary Skaggs

Addresses: Virginia Tech, 145 Dixon Drive, Hardy VA 24101, USA

Abstract: If items become more or less difficult or discriminating depending on their position within a test form, and if different examinees respond to the same items in different positions, the fairness of test scoring for examinees is undermined. Such context effects have been investigated in the past with mixed results. This study investigates whether item position effects are present in a computer-administered certification test in which items are presented in random order. The results showed few consistent significant position effects for either item difficulty or discrimination for the entire test. However, there was a tendency for items to be slightly more difficult when they were administered in the last five positions at the end of each test form. For individual items, there was no obvious position effect for either item difficulty or discrimination. For practitioners working in computer-based testing programs, the findings support the use of random item ordering, but it is still recommended that each testing program conduct an item position investigation for items.

Keywords: item position effects; random item ordering; computer-based testing; certification tests.

DOI: 10.1504/IJQRE.2016.073673

International Journal of Quantitative Research in Education, 2016 Vol.3 No.1/2, pp.94 - 108

Available online: 15 Dec 2015 *

Full-text access for editors Access for subscribers Free access Comment on this article