We describe how an answer-until-correct multiple-choice (MC) response format allows for the construction of fully multiple-choice examinations designed to operate much as a hybrid between standard MC and constructed-response (CR) testing. With this tool—the immediate feedback assessment technique (IF-AT)—students gain complete knowledge of the correct answer for each question during the examination and can use such information for solving subsequent test items. This feature allows for the creation of a new type of context-dependent item set: the “integrated testlet.” In an integrated testlet, certain items are purposefully inter-dependent and are thus presented in a particular order. Such integrated testlets represent a proxy of typical CR questions, but with a straightforward and uniform marking scheme that also allows for granting partial credit for proximal knowledge. As proof-of-principle, we present a case study of an IF-AT-scored midterm and final examination for an introductory physics course and discuss specific testlets possessing varying degrees of integration. In total, the polychotomously scored items are found to allow for excellent discrimination, with a mean item-total correlation measure for the combined 45 items of the two examinations of r¯=0.41±0.13 (mean ± standard deviation) and a final examination test reliability of α = 0.82 (n = 25 items). Furthermore, partial credit is shown to be allocated in a discriminating and valid manner in these examinations. As has been found in other disciplines, the reaction of undergraduate physics students to the IF-AT is highly positive, further motivating its expanded use in formal classroom assessments.

1.
G. J.
Aubrecht
 II
and
J. D.
Aubrecht
, “
Constructing objective tests
,”
Am. J. Phys.
51
,
613
620
(
1983
).
2.
M.
Scott
,
T.
Stelzer
, and
G.
Gladding
, “
Evaluating multiple-choice exams in large introductory courses
,”
Phys. Rev. ST Phys. Educ. Res.
2
,
020102
1
(
2006
).
3.
S.
Tobias
and
J. B.
Raphael
, “
In-Class examinations in college-level science: New theory, new practice
,”
J. Sci. Educ. Technol.
5
,
311
320
(
1996
).
4.
T. M.
Haladyna
,
Developing and Validating Multiple-choice Test Items
, 3rd ed. (
Lawrence Erlbaum Assoc.
,
MahWah, NJ
,
2004
).
5.
D.
Hestenes
,
M.
Wells
and
G.
Swackhamer
, “
Force Concept Inventory
,”
Phys. Teach.
30
,
141
158
(
1992
).
6.
T. M.
Haladyna
, “
Context dependent item sets
,”
Educ. Meas. Issues Pract.
11
,
212
225
(
1992
).
7.
H.
Wainer
and
C.
Lewis
, “
Toward a psychometrics for testlets
,”
J. Educ. Meas.
27
,
1
14
(
1990
).
8.
S. G.
Sireci
,
D.
Thissen
, and
H.
Wainer
, “
On the reliability of testlet-based tests
,”
J. Educ. Meas.
28
,
237
247
(
1991
).
9.
P.
Baldwin
,
S. G.
Baldwin
, and
S. A.
Haist
, “
F-type testlets and the effects of feedback and case-specificity
,”
Acad. Med.
86
,
S55
(
2011
).
10.
M.
Peat
and
S.
Franklin
, “
Supporting student learning: The use of computer-based formative assessment modules
,”
Br. J. Educ. Technol.
33
,
515
523
(
2002
).
11.
M. L.
Epstein
,
A. D.
Lazarus
 et al., “
Immediate feedback assessment technique promotes learning and corrects inaccurate first responses
,”
Psychol. Rec.
52
,
187
(
2002
).
12.
D.
DiBattista
, “
The immediate feedback assessment technique: A learning-centred multiple-choice response form
,”
Can. J. High. Educ.
35
,
111
(
2005
).
13.
R. E.
Dihoff
,
G. M.
Brosvic
 et al., “
Provision of feedback during preparation for academic testing: Learning is enhanced by immediate but not delayed feedback
,”
Psychol. Rec.
54
,
207
231
(
2004
).
14.
D.
DiBattista
,
L.
Gosse
 et al., “
Grading scheme, test difficulty, and the immediate feedback sssessment technique
,”
J. Exp. Educ.
77
,
311
336
(
2009
).
15.
D.
DiBattista
,
J. O.
Mitterer
, and
L.
Gosse
, “
Acceptance by undergraduates of the immediate feedback assessment technique for multiple-choice testing
,”
Teach. High. Educ.
9
,
17
28
(
2004
).
16.
Eric
Mazur
,
Peer Instruction: A User's Manual
(
Prentice-Hall
,
Upper Saddle River, NJ
,
1997
).
17.
C. H.
Crouch
and
E.
Mazur
, “
Peer Instruction: Ten years of experience and results
,”
Am. J. Phys.
69
,
970
977
(
2001
).
18.
D. E.
Meltzer
and
K.
Manivannen
, “
Transforming the lecture-hall environment: The fully interactive physics lecture
,”
Am. J. Phys.
70
,
639
654
(
2002
).
19.
G. M.
Novak
,
E. T.
Patterson
 et al.,
Just-In-Time Teaching: Blending Active Learning with Web Technology
(
Prentice-Hall
,
Upper Saddle River, NJ
,
1999
).
20.
D.
DiBattista
and
L.
Kurzawa
, “
Examination of the quality of multiple-choice items on classroom tests
,”
Can. J. Schr. Teach. Learn.
2
,
Art
4
(
2011
).
21.

There is, however, a variety of IF-AT forms with variety of answer keys available from Epstein Educational—a fact that greatly aids in test security.

22.
L.
Ding
and
R.
Beichner
, “
Approaches to data analysis of multiple-choice questions
,”
Phys. Rev. ST Phys. Educ. Res.
5
,
020103
1
(
2009
).
23.
R. L.
Ebel
and
D. A.
Frisbie
,
Essentials of Educational Measurement
, 5th ed. (
Prentice-Hall
,
Englewood Cliffs, NJ
,
2011
).
24.
G. M.
Bonder
, “
Statistical analysis of multiple-choice exams
,”
J. Chem. Educ.
57
,
188
(
1980
).
25.
M. J.
Allen
and
W. M.
Yen
,
Introduction to Measurement Theory
(
Waveland Press
,
Long Grove, IL
,
2002
), pp.
95
115
.
26.
G. W.
Pfeiffenberger
and
C. C.
Modu
, “
A validity study of the multiple-choice component of the advanced placement physics C examination
,”
Am. J. Phys.
45
(
11
),
1066
1069
(
1977
).
27.
J.
Cohen
,
Statistical Power Analysis for the Behavioral Sciences
, 2nd ed. (
Lawrence Erlbaum Assoc.
,
Hillsdale, NJ
,
1988
).
28.
E. M.
Rogers
, “
Examinations: Powerful agents for good or ill in teaching
,”
Am. J. Phys.
37
,
954
962
(
1969
).
29.
C. B.
Spurgin
, “
What earns the marks?
,”
Phys. Educ.
2
,
306
310
(
1967
).
30.
R. L.
Johnson
,
F.
McDaniel
 II
, and
M. J.
Willeke
, “
Using portfolios in program evaluations: An investigation of interrater reliability
,”
Am. J. Eval.
21
,
65
80
(
2000
).
31.
S.-Y.
Lin
and
C.
Singh
, “
Can multiple-choice questions simulate free-response questions?
,”
AIP Conf. Proc.
1413
(
1
),
47
50
(
2012
).
32.
S.-H.
Chang
,
P.-C.
Lin
, and
Z.-C.
Lin
, “
Measures of partial knowledge and unexpected responses in multiple-choice tests
,”
Educ. Technol. Soc.
10
,
95
109
(
2007
).
33.
A.
Ben-Simon
,
D. V.
Budescu
, and
B.
Nevo
, “
A comparative study of measures of partial knowledge in multiple-choice tests
,”
Appl. Psychol. Meas.
21
,
65
88
(
1997
).
34.
J. D.
Jaradat
and
S.
Sawaged
, “
The subset selection technique foe multiple-choice tests: An empirical inquiry
,”
J. Educ. Meas.
23
,
369
376
(
1986
).
35.
M. C.
Rodrigues
, “
Construct equivalence of multiple-choice and constructed-response Items: A random effects synthesis of correlations
,”
J. Educ. Meas.
40
,
163
184
(
2003
).
36.
M. L.
Epstein
,
B. B.
Epstein
, and
G. M.
Brosvic
, “
Immediate feedback during academic testing
,”
Psychol. Rep.
88
,
889
894
(
2001
).
37.
R. E.
Bennet
, “
Formative assessment: A critical review
,”
Assess. Educ.
18
,
5
25
(
2011
).
38.

The top two scorers overall were excluded from this analysis because they scored 100% on the exam and thus had no opportunity for earning partial credit. The next fifteen top-ranked students were used to represent the top group.

39.

For non-integrated items, it is presumed that the order of question presentation is irrelevant. However, there is some indication that this may not be the whole story; see Ref. 44.

40.

This “Introductory Physics for the Life Sciences” is an algebra-based course that significantly differs in subject matter from the “Introductory Physics II: Electricity and Magnetism” course discussed throughout this article. While the topics between the two courses were different, the methods of instruction and assessment were nearly identical.

41.
L.
Ding
,
N.
Reay
 et al., “
Exploring the role of conceptual scaffolding in solving synthesis problems
,”
Phys. Rev. ST Phys. Educ. Res.
7
,
020109
1
(
2011
).
42.
C.
Singh
, “
Assessing student expertise in introductory physics with isomorphic problems. II. Effect of some potential factors on problem solving and transfer
,”
Phys. Rev. ST Phys. Educ. Res.
4
,
010105
1
(
2008
).
43.
H.
Kruglak
, “
Experimental Study of Multiple-Choice and Essay Tests. I
,”
Am. J. Phys.
33
(
12
),
1036
1041
(
1965
).
44.
K.
Gray
,
N.
Rebello
, and
D.
Zollman
, “
The Effect of Question Order on Responses to Multiple-choice Questions
,”
Presented at the Physics Education Research Conference 2002
,
Boise
,
Idaho
(
2002
).
AAPT members receive access to the American Journal of Physics and The Physics Teacher as a member benefit. To learn more about this member benefit and becoming an AAPT member, visit the Joining AAPT page.