Professional Documents
Culture Documents
Millman Et Al 1965 An Analysis of Test Wiseness
Millman Et Al 1965 An Analysis of Test Wiseness
Millman Et Al 1965 An Analysis of Test Wiseness
AN ANALYSIS OF TEST-WISENESS
1
ance in educational test scores other than item content and random
error. Test-wiseness was one suggested source.
1 This research was supported, in part, by the Cornell Social Science Re-
search Center through a grant from the Ford Foundation.
2 Now at
Michigan State University.
707
708 EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT
mann, 1962) have referred to test-wiseness as a potentially large
sourceof test variance. Almost all textbooks in educational measure-
ment provide rules intended to help the test constructor avoid giv-
ing away the answer. Numerous pamphlets on how to take tests are
available(e.g. Hook, 1958; Heston, 1953; and Pettit, 1960).
Although no comprehensive investigation of test-wiseness is
known to exist, data from a few empirical studies indicate that it is
a factor which may deserve attention.
Comparisons made by Vernon of British and American Students
on different types of examinations &dquo;illustrate the importance of
are usually less than the standard error of the test, but they are in
the positive direction and represent systematic bias.
Recent reviews of non-equivalence of measures of supposedly the
same traits (Campbell and Fiske, 1959; Smith and Kendall, 1963)
the answer they did at the time they responded to the item.) There
was a great range in the sophistication of reasons given by students
for responding as they did to test items in which they did not know
the answers.
The particular styles reported are highly specific to the nature of
the items being solved. However, the problem solving styles of high
and low test performers could be distinguished. Bloom and Broder
(1950) report that students trained in those general problem solv-
ing techniques (including the comprehension of test directions, the
understanding of the nature of the specific test questions, and the
ability to reason logically) used by high test performers, but not
additionally trained in subject-matter knowledge, made significant
gains in subsequent achievement test scores. French (1965) dem-
onstrated that the factor composition of a test frequently depended
upon the problem solving styles used in answering the test items.
The results of the references cited above and an analysis of the
literature dealing with principles of test construction or advice for
taking examinations were useful in developing the following out-
line.
JASON MILLMAN, ET AL. 711
An Outline of Test-Wiseness Principles
I. Elements independent of test constructor or test purpose.
A. Time-using strategy
1. Begin to work as rapidly as possible with reasonable as-
surance of accuracy.
2. Set up a schedule for progress through the test.
3. Omit or guess at items (see I.C. and II.B.) which resist
a quick response.
B. Error-avoidance strategy.
1. Pay careful attention to directions, determining clearly
the nature of the task and the intended basis for response.
2. Pay careful attention to the items, determining clearly
the nature of the question.
3. Ask examiner for clarification when necessary, if it is per-
mitted.
4. Check all answers.
C. Guessing strategy.
1. Always guess if right answers only are scored.
2. Always guess if the correction for guessing is less severe
than a &dquo;correction for guessing&dquo; formula that gives an
expected score of zero for random responding.
3. Always guess even if the usual correction or a more severe
penalty for guessing is employed, whenever elimination of
options provides sufficient chance of profiting.
D. Deductive reasoning strategy.
1. Eliminate options which are known to be incorrect and
choose from among the remaining options.
2. Choose neither or both of two options which imply the
correctness of each other.
3. Choose neither or one (but not both) of two statements,
one of which, if correct, would imply the incorrectness of
the other.
712 EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT
4. Restrict choice to those options which encompass all of
two or more given statements known to be correct.
5. Utilize relevant content information in other test items
and options.
B. Cue-using strategy.
1. Recognize and make use of any consistent idiosyncracies
crease their scores when they do (Berrien, 1939, Briggs and Reile,
1952). There is some evidence that persistence (i.e. using full time
on test) pays off (Briggs and Johnson, 1942).
JASON MILLMAN, ET AL. 715
3If a minimax decision function were used, that is, an examinee wanted to
minimize his maximum loss, he would never guess when a penalty for guessing
was employed. As another example, suppose the testee wished to guess only
when the probability of improving his score by guessing was greater than one-
half. If the usual correction for guessing formula was employed, he would
guess only when k +1)/
n
( = an integer ( >2),
k where n is the number of
items to be guessed and k the number of choices per item. (Graesser, 1958).
716~ EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT
for an incorrect answer (i.e. scored rights only). A person
who makes a pure guess is expected to earn
Example:
If, in the above illustration, 1/3 of a point were subtracted for
an incorrect answer, then:
E.V. = 1/4 (1) + (1 1/4) (-1/4) ;= 1/16 point.
-
Examples:
Which of the following men were presidents of the United
States?
a) George Washington
b)Andrew Jackson
c)Abraham Lincoln
*d)All of the above
A test-wise examinee who was sure that Washington and Lincoln
were presidents but was undecided about Jackson would neverthe-
c) kangaroo
d) whale
Assume the examinee knows that a bird is a warm-blooded ani-
mal, and that all animals are either warm-blooded or cold-blooded
but not both. He can then reason that a snake, which is an option
in both items, must be a cold-blooded animal and can, therefore,
answer the second question.
II.A. Intent consideration strategy.
It is possible that the examinee may receive a low score in an ob-
jective test merely because his views differ from the test construc-
tor’s or his level of knowledge is higher than that being tested. By
recognizing and acting upon the biases of the test constructor or the
intent of the test, the examinee may avoid loss of points due to mis-
interpretation, rather than lack of knowledge of the subject matter.
(The use of this strategy assumes that the goal of the examinee is
to earn the greatest possible number of points; i.e., he is not willing
to lose points on principle.)
II.A.l. Interpret and answer questions in view of previous emphasis
of the test constructor or in view of the test purpose.
It occasionally is difficult to determine what a question is &dquo;get-
ting at,&dquo; and as a consequence the answer is elusive. If the test is
taken with a set for the test constructor’s views or purposes, one
may be able to interpret the question more easily.
Example:
There is a test for normality. T F
If the examinee knows that the purpose of the test is to measure
statistical knowledge rather than knowledge of abnormal psychol-
ogy, the question may be interpreted easily.
II.A.2. Answer items as the test constructor intended.
If the examinee believes that the test constructor had a certain
answer in mind, but the examinee can think of a possible objection
to the answer, he should answer the item as he believes the test
constructor intended. That is, he should choose the option he be-
lieves has the greatest chance of being correct, even though others
have merits and even though the chosen option is not completely
satisfactory. This assumes that no explanation can be. communi-
cated to the grader; only the answer may be marked.)
Example:
Thomas Jefferson wrote the Declaration of Independence. TF
720 EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT
The examinee may object to answering &dquo;true&dquo; on the basis that
the statement is not complete since four other men were also ap-
pointed to the committee in charge of writing it. The test-wise per-
son, however, would answer &dquo;true&dquo; if he felt this was the answer
which the test constructor considered correct.
II.A.3. Adopt the level of sophistication that is expected.
An item (especially a true-false item) may have different correct
answers depending upon the &dquo;depth&dquo; which the constructor or pur-
pose demands. The test-wise person will recognize and adopt the
appropriate level.
Example:
Light travels in a straight line. T
F
This item may be keyed as true in a test given at an elementary
grade level, but may be false in an advanced college course in
physics.
II.A.4. Consider the relevance of specific detail.
Depending upon the test constructor or test purpose, the specific
detail in an item may or may not have bearing upon the answer.
Example:
A picture of George Washington, the most influential man in
shaping the history of the United States, appears on the one dol-
lar bill. T F
The test-wise person would have to decide whether the appositive
were a mere insertion by an enthusiastic admirer of Washington,
Examples:
The 13th, 14th, and 15 amendments to the Constitution are
REFERENCES
Ahmann, J. Stanley and Glock, Marvin D. Evaluating Pupil
Growth. Boston: Allyn and Bacon, Inc., 1963.
Anastasi, Anne. Psychological Testing. New York: The Macmillan
Company, 1961.
Berrien, F. K. "Are First Impressions Best on Objective Tests?"
School and Society, L. (1939), 319-20.
Bloom, Benjamin S. and Broder, Lois J. Problem-solving Proc-
esses of College Students: An Exploratory Investigation. Chi-
cago: The University of Chicago Press, 1950.
Briggs, Arvella and Johnson, D. M. "A Note on the Relation be-
tween Persistence and Achievement on the Final Examination."
Journal of Educational Psychology, XXXIII (1942), 623-27.
Briggs, Leslie J. and Reile, Patricia J. "Should Students Change
Their Initial Answers on Objective-Type Tests?: More Evi-
dence Regarding an Old Problem." Journal of Educational
Psychology, XLIII (1952), 110-15.
Campbell, Donald T. and Fiske, Donald W. "Convergent and Dis-
criminant Validation by the Multitrait-Multimethod Matrix."
Psychological Bulletin, LVI (1959), 81-105.
Connerly, John A. and Wantman, Morey J. "An Exploration of
Oral Reasoning Processes in Responding to Objective Test Items."
Journal of Educational Measurement, I (1964), 59-64.
Cook, Desmond. "A Comparison of Reading Comprehension Scores
Obtained Before and After a Time Announcement." Journal of
Educational Psychology, XLVIII (1957), 440-46.
Earle, Dotsie. "A Study of Problem-Solving Methods Used on
Comprehensive Examinations." Chicago: University Examin-
er’s Office, University of Chicago, 1950. (Mimeographed)
Ebel, Robert L. and Damrin, Dora E. "Tests and Examinations."
In Chester Harris (Editor) Encyclopedia of Educational Re-
search. New York: The Macmillan Company, 1960. (Pp. 1502-
517.)
French, John W. "The Relationship of Problem-Solving Styles to
the Factor Composition of Tests." EDUCATIONAL AND PSYCHO-
LOGICAL MEASUREMENT, XXV (1965), 9-28.
Gaier, Eugene L. "Students’ Perceptions of Factors Affecting Test
Performance." Journal of Educational Research, LV (1962),
561-66.
Graesser,R. F. "Guessing on Multiple-Choice Tests." EDUCATIONAL
AND PSYCHOLOGICAL MEASUREMENT, XVIII (1958), 617-20.
Gupta, R. K. and Penfold, D. M. Edwards. "Correction for Guess-
ing in True-False Tests: An Experimental Approach." British
Journal of Educational Psychology, XXXI (1961), 249-56.
726 EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT