You are on page 1of 16

1

3
Planning for !lssessmen.t

Studying this chapter should enable you to


1. List the types of learning outcomes to consider in planning for the assessment of
student achievement.
2. State instructional objectives as intended learning outcomes and define them in
performance terms.
3. Prepare a set of specifications for a test.
4. Describe:the relative merits of selection and supply type test items.
5. Match test items to the specific learning outcomes they measure.
6. Describe the factors to consider when preparing items for a test.
7. Describe how to arrange items in a test.
8. Write clear directions for a test.
9. Review 'md evaluate an assembled test.
10. Administer a test properly.
11. Make a simple item analysis.

Planning for the assessment of student achievement involves a consj,


of the following questions:

1. What do we expect students to learn?


2. "Y.h~twes of student perfoI'II1a!\cear~ \:.Ve
W~g ~().acc;:.ep!.!s
e
of.leaming? -
3. What <Jssessmentinstruments will bes~ evaluate the students'
mance? -. .

31
32 Chapter 3 Planning for Assessment 33

The first question focuses on thi\1oals of the curriculum as set by the V TABLE3.1 Some Types of Intended Learning Outcomes
school and influenced l?.Y. the.district and state. In recent years,states have Types Sample Categci,!!~ ::~''?';;'~~3
developedt~ontent standardsjthat provide guidelines for determining what
students snould know and be able to do. This has influenced districts and Facts
Knowledge of
schools to modify curriculums and put gr~ater emphasis on the more com- Concepts
plex learning outcomes. Comprehension of Principles
The second question focuses on the importance of specifying instruc- Methods
~~nal objectives in. Berformance terms and including all types Of desir~d Application of Process
learning outcomes. n\ese objectives must, of course, be in harmony with the Analyzing
goals of the curriculum and will reflect the state content standards to the Reasoning Comparing
same degree as the school curriculum. In final analysis, the assessment of Ability Inferring
student achievement is determined by what the students are expected Generalizing
to learn, and this is determined by the goals of the school curriculum. The Evaluating
~structional objectives.are s~mply a f!\eaI:lSof st.ating the curriculu~$oals Speaking
Observable Oral reading
?L~s.e~!.n.!!.5 s<!.-t!'a!!hey' are useful in both instruction 'and Skills
assessment. Laboratory skills
-'--rfiethird question focuses on the match between the instructional Psychomotor skills
Work-study skills
objectives and the procedures of assessment.tThe objectives spe~!!Y..!h~:
tended learning outcomes in performance. terms -and' -tne _ass~~t Writing
metli.odis Used:to det~1'ififfietJie"exlentto which students' performance is Drawing
Products Designing
satisfactorylFor some performance outcomes a test wil~ provide adequate
Constructing
evidence. In other cases, some type of observationel technique will be
Problem-solving
needed to evaluate the ongoing nature of the performance (e.g., giving a
speech) or to evaluate a res1;1ltingproduct (e.g., a written report). The key _

to effectiveasses~rnentis
avaiTable.
~ use the most direct and relevant meth~d
----.-
Role of Instructional Objectives

Typef of Intende_dJ~~arning Outcome~ ~ell-stated instructio~al objectives provide a description of Jhe intended
~earning outcomes in perfo~a,nce terms-that is, in terms of the types of
In planning for the assessment of student achievement, there are a number performance students can demonstrate to show that they have achieved the
of types of learning outcomes that might be considered. Although the goals knowledge, understanding, or skill described by the objective. By describing
of the school and the nature of the instructional content will determine what the performance that we are willing to accept as evidence of learning, we
specific types of learning outcomes are to be assessed, a review of the vari- provide a f9.9!sfor instmction, student learning, and assessment. Objectives
ous tyiPesof outcomes shown in Table 3.1 will prevent any serious omis- help keep all three in close harmony. For example, for an objective empha-
sions. trhe list is not exhaustive, but it makes clear the range of outcomes to sizing problem-solving strategies, we teach students how to apply problem-
consider beyond that of knowledge, when preparing objectives for instruc- solving strategies, they practice applying the strategies, and we assess their
tion and assessment. skill in applying problem-solving strategies with new problems. Thus, the
ell-designed classroom tests can be used to measure many of the out- instructional targets, the learning targets, and the assessment targets are all
come in the cognitive areas, but skills and products require the use of obser- the same. If the students are made aware of the objectives at the beginning of
vatio al techniques such as checklists, rating scales, or holistic scoring instruction, both teacher and students are working toward common goals
t
rubri s. and instruction and assessment are both part of the same process.
l
34 Chapter 3 t. . I'IIIIII/illg fllr As~:rss"/(,IIf 35

Stating
~
Instructional Objectives .~
BOX 3.1 · Stating Instructional Objectives
-- --
In stating instructional objectives, it is helpful to keep in mind that we are
The following steps provide guidelines for stating instru~ona1 qbjectives that
not describing the teaching procedures or the learning process~e ~re ~im- are usefulin instructionand assessment; ,
ply descr!b~Lthe stud~t perfo~~ce to be demonstrated at the ~~~...2f
ffierearnmg experience as evidence of learning:JThis permits us to use a vari- 1. State the general objectives as follows:
etyof procedures, materials, ani:lleamirig activities to achieve the desired 1.1 Write each as an intended learning outcome.
outcomes. 1.2 State each in performance terms.
A useful procedure is to first state the general objectives that focus on 1.3 Start each with a verb (e.g., "knows," "comprehends").
the broader learning outcomes" ana men define each general objective in 1.4 Write each so it is general enough to identify a domain of specific leam-
more specific terms, as follows. ing outcomes. '
2. List and state the specific learning outcomes as follows:
1. Understands Scientific Concepts 2.1 Clarify each general objective with a representative sample of specific
1.1 Describes the concept in his 01'her own words. learning outcomes. .
1.2 Describes the role of the concept in science. 2.2 Begin each with an action verb indicating observable student perfor-
1.3 Distinguishes the concept from similar scientific concepts. mance (e.g., "selects," "describes").
1.4 Uses the concept in writing about science. 2.3 Include a sufficient number to indicate attainment of the general objective.
1.5 Applies the concept in solving problems. 2.4 Check to be sure the specific learning outcomes are relevant to the gen-
eral objectives and that the general objectives are in harmony with the
school goals.
The specific learning outcomes listed under the general objective are
not meant to be exhaustive, but they should provide a representative sample
of the types of performance that we are willing to accept as evidence that
the objective has been achieved. The general procedure for stating instruc-
tional objectives is presented in Box 3.1. The specific application of these
steps is described and illustrated later for a unit on test planning. d. Analysis (breaking down material into its parts).
e. Synthesis (putting parts together into a whole).
f. Evaluation (judging the value of a thing using criteria).
Sources of Help in Locating Sample Objectives These categories are listed in.the 'order of increClsingcomplexit¥
In getting started with the identification and selection of instructiqnal objec- and are clarified in the taxonomy by subdivisions that further describe
tives, it helps to review various sources for ideas. Illustrative categories of the types of educational outcomes included. For detail~d descriptions of
intended learning outcomes and sample objectives can be found in the fol- all three domains of the taxonomy with illustrative objectives for each
lowing sources. category see Gronlund (2000)at the end of this chapt-:r.
2. Instructors' guides accompanying-student textbooks. These guides typ-
1. Taxonomy o/Educational Objectives. This is an older guide but it pro- ically contain objectives but they tend to focus on lower level outc0mes,
Vli:lesa classificationof objectives in the cognitive, psychomotor, and af- may be poorly stated, and are likely to lack close harmony with the
fective domains. Each domain includes categories and subcategories goals of the curriculum. However, they are worth a review for ideas on
that are designed to identify and classify all possible educational out- possible learning outcomes to consider. It may be possible to modify
comes. For example, the cognitive domain includes the following major some to put them in more usable form.
categories (Bloom, 1956): 3. Publications 0/ educational organizations. There are organizations in
a. Knowledge (remembering previous learning material). each of the main t~aching areas that publish useful material on in-
b. Comprehension (grasping the meaning of material). struction that contains suggested objectives. For exarltple, the National
c. Application (using information in concrete situation). Science Teachers Association, the National Council of Teachers of

"ii

"III
"
Chapter 3 Planningfor Assessment 37
H
I

Mathematics, the National Council of Teachers of English, and the Na-


Focus on Life Science
tional Council for the Social Studies periodically iSsue yearbooks and
other publications on instruction in their content area. The more recent Cell Biology
publications will focus on learning outcomes that are in harmony with 1. All living organisms are composed of cells, from just one to many trillions, whose
the newer curriculum emphasis on. reasoning and problem-solving details usually are visible only through a microscope. As a basis for understanding
skills. this concept:
4. State content standards. The majority of states have developed content a. Studentsknowcellsfunctionsimilarlyin all livingorganisms.
b. Students know the characteristics that distinguish plant cells from animal cells,
standards that provide guideiines-for'-aetemumng wnat-'students
including chloroplasts and cell walls.
should know and Dea6fi to do,in subject areas. TheSe standards 'typi- c. Students know the nucleus is the repository for genetic information in plant and
cally emphasize'complex learning outcomes that focus on understand- animal cells.
ing of concepts and problem-solving skills.~,us, they may not r~present d. Students know that mitochrondria liberate energy for the work that cells do and
a total curriculum i!1 Qte,content areas but serve as a fraIl}eworklqr that chloroplasts capture sunlight energy for photosynthesis.
shaping the goals of the districts and school,s.The intent is to move the e. Studentsknowcells diVide to increase their numbers through a process of mitosis,
scnool cUrriculum more tc:>warda'thinking curricu!um and t!'~reby i.m- which results in two daughter cells with identical sets of chromosomes.
~oving both instruction. and a~sessment. f. Studentsknowthat as multicellular organisms develop, their cellsdifferentiate.
'l wo InuStrafivecontent standards, developed by the California Investigation and Experimentation
Department of Education, are shown in Figure 3.1. The standards are 7. Scientificprogress is made by asking meaningful questions and conducting careful
followed by examples of how students will show that they have met the investigations. As a basis for upderstanding this concept and addressing the con-
standards. Although the standards are described as exemplary and tent in the other three strands, students should develop their own questions and
compliance with them is not mandatory, schools and districts are en- perform investigations. Students will:
couraged to use them in curriculum development. a. Select and use appropriate tools and technology (including calculators, com-
State content standards are typically reflected to some degree in puters, balances, spring scales, microscopes, and binoculars) to perform tests,
the school curriculum and thus provide a valuabie focus or tnemo(e collect data, and display data.
b. Use a variety of print and electronic resources (including the World Wide Web)
complex learning outcomes: Although they are usually too general for to collect information and evidence as part of a research project.
direct use in instruction and assessment, a review of these will provide c. Communicate the logical connection hypotheses, science concepts, tests con-
useful ide~s for preparing instructional objectives. In those schools that ducted, data collected, and conclusions drawn from the scientific evidence.
emphasize standard-based instruction and standard-based assessment, d. Construct scale models, maps, and appropriately labeled diagrams to commu-
it will be necessary to pay special attention to the content standards nicate scientific knowledge (e.g., motion of Earth's plates and cell structure).
prepared by the state when writing instructional objectives. e. Communicate the steps and results from an investigation in written reports
and oral presentations.

FIGURE3.1 Sample California Science Content Standards-Grade 7.


~reparing and Using an Achievement Test FromScienceContent Standardsfor CaliforniaPublicSchools,Copyright 2000,California
Department of Education. Used by permission.
l1e preparation and use of an achievement test that measures the intended
earning outcomes in a balanced manner involves the following steps:
vl. Specifying the instructional objectives.
\2. Preparing the test specifications.
\,3. Constructing r.elevanttest' items.
\4. Arranging the items in the test.
,,5. Preparing clear directions.
38 Chapter 3 Planning for Assessment

6. .Reviewing and evaluating the assembled test.


\ v'7. Administering the test and making an item analysis. ./ 1. 1.1
Knows the meaning of common-terms.
Identifies the correct definitions of terms.
1.2 Identifies the meaning of terms when used in context.
Each of these steps will be described in turn. Performance assessment 1.3 Distinguishes between terms on basis of meaning.
will be described in Chapters 7, 8, and 9. 1.4 Selects the most appropriate terms when describing testing proce-
./. dmes.
./I. Comp~ehends the relevant principles of testing.
Specifying the Instructional Objectives 4.1 Describes each principle in his or her own words.
4.2 Matches a specific example to each principle.
As noted earlier, the first step in preparing instructional objectives it to list 4.3 Explains the relevance of each principle to the major steps in test
the general objectives as intended learning outcomes. The following list planning.
illustrates the procedure for a unit on planning an achievement test. 4.4 P ,edicts the most probable effect of violating each of the principles.
4.5 Formulates a test plan that is in harmony with the principles.
At the end of this unit in achievement test planning the student will demon-
strate that he or she: Note that the terms used to describe the spe<#i~learning o.\l.tcomesin-
dicate student performance that can he..de.m.onstratedto an outsi~te9..Q~~rY~!:
-./ 1. Knows the meaning of common terms. Thatis, they are !!.~servableresp_~_that can be calledforth by test items.
",.2. Knows specific facts about test planning. The key terms are listed below to emphasize what is meant by defining learn-
./' 3. Knows the basic procedures for planning an achievement test. ing outcomes in sE~c:!£cP!.!i01'1!'anc!!...teT7!'~'
y 4. Comprehendsthe relevantprinciplesof testing. .
.' 5. Applies the principles in test planning. Identifies Matches
Distinguishes between Explains
These statements of general}e~g outcomes have be~ deliberately Selects Predicts
~ept free of specific course content so that with only slight modi~<:~.!ionth~y Describes Formulates
can be used with various units of study. As we shall see later, the test speci-
fications provide a means of relating intended outcomes to specific subject Action verbs such as these indicate prel)isely what the student is ab~ to
matter topics. do to demonstrate achievement. Such vague.and indefinite terms as "learns,"
This list of general outcomes could, of course, be expanded by making "sees," "realizes," and "is familiar with" should be avoided, since they do
the statements more specific, and in some cases it may be desirable to do so. not clearly indicate the terminal performance to be measUled.
The number of general learning outcomes to use is somewhat arbitrary, but Sample action verbs for stating specific learning outcomes at each level
somewhere between 5 and 15items provide a list that is both useful and man- of the cognitive domain of the taxonomy are presented in Table 3.2.
ageable. Typically,a shorter list is satisfactory for a unit of study, while a more Although certain action verbs may be used at several different levels (e.g.,
comprehensive list is needed for summative testing at the end of a course. "identifies"), the table provides a useful guide for defining intended out-
comes in performance terms. For more comprehensive lists of action.verbs,
see Gronlund (2000)listed at the-end ofthiScnapter.--......_--
Defining the General Outcomes in Specific Terms
In defining the general learning outcOi'i\eS-in specific performance
When a satisfactory list of general learning outcomes has been identified and terms, it is typically impossible to list all of the relevant types of perfor-
clearly stated, the next step is to list the specific types of student performance mance. The proportio~ that need be listed depends to a large extent on the
that are to be accepted as evidence that the outcomes have been achieved. For nature of the test. In planning a test that is to be used to describe which
example, what specific types of performance will show that a student "knows learning tasks a student has mastered (criterion. referenced test), we should
the meaning of common terms" or "comprehends the relevant principles of like as comprehensive a list as possible. For a test that is used to rank stu-
testing"? For these two areas the specific learning outcomes may be listed as dents in order of achievement (norm-referenced test), however, it is usually
follows: satisfactory to include a sufficient number of specific types of performance
40 Chapter 3 Planning for Assessment 41

TABLE3.2 Illustrative Action Verbsfor Defining Objectives in the Cognitive needs of the students, and a host of other local factors that have a bearing on
Domain of the Taxonomy the instructional program. Despite the variation from course to course, most
list~f instructional objectives will include learning outcomes in the foUOW:-
Taxonomy
Sample Verbsfor Stating Specific Learning Outcomes
ing areas: (1) knowledge, (2) intellectual abilities and skills, (3)$eneral skills
Categories
(laboratory, performance, communication, work-study), and (4)attitudes, ill-
Kllowledge Identifies, names, defines, describes, lists, matches, selects, ~~~st~,_~n..<:!-"~p.£r~~i~tions.
It is in the first two areas covered by the cogni~
outlines domain of the taxonomy that achievement testing is most useful. Learning
Comprehension Classifies, explains, summarizes, converts, predicts, distin- outcomes in the other areas are typically evaluated by rating scales, check-
guishes between lists, anecdotal records, inventories, and similar nontest evaluation proce-
dures. Thus, the first step is to separate from the list of learning outcomes
Application Demonstrates, computes, solves, modifies, arranges, operates,
relates those that are testable by paper-and-pencil tests. The selected list of learning
outcomes should, of course, be defined in specific terms, as described in the
Analysis Differentiates, diagrams, estimates, separates, infers, orders, previous section. Clarifying the specific types of performance to be called
subdivides
forth by the test will aid in constructing test terms that are most relevant to
Synthesis Combines, creates, formulates, designs, composes, constructs, the intended learning outcomes.
rearranges, revises
Evaluation Judges, criticizes, compares, justifies, concludes, discriminates, Outlining the Subject Matter. The stated learning outcomes specify how
supports students are expected to react to the subject matter of a course. Although it is
possible to include both the student performance and the specific subject mat-
tcr the student is to react toward in the same statement, 'itis usually desirable
to clarify what the typical student is like who has achieved the intended to list them separately. The reason for this is that the student can react in the
outcomes. same way to many different areas of subject matter, and he can react in many
different ways to the same area of subject matter. For example, when we state
that a student can "define a term in her own words," "recall a specific fact,"
Preparing the Test Specifications or "identify an example of a prmap!e," these types of performance can be ape-
plied to almost any area"of subject matter. Since particular types of student
The writing of test items should be guided by a carefully prepared set of test performance can overlap a variety of subject matter areas, and vice versa, it
specffications. The Junction of the specifications is to describe the achieve- is more convenient to list each aspect of performance and subject matter sep-
meqt: domain being measured and to provide guidelines for obtaining a arately and then to relate them in the table of specifications.
representative sample of test tasks. Although the nature and detail of test The c~tent ofa course may be outlined in detail for te~ching purposes,
specifications can be expected to vary considerably, here we shall describe but for test pfanning onl~th~-ID.aioJ:categodes...neeg,Q~)i§.!.ed.The following
one f the more commonly recommended procedures. In the construction of outline of subject matter topics based on a unit on achievement testing illus-

wa t
an chievement test, one of the most widely used devices has been a two-
chart called a. table of specifications.
trates sufficient detail for the test plan.

A. Roleof testing in the instructional process


1. Instructional decisions and test types

r
Building a Table of Specifications V 2. Influence of tests on learning and instruction
pre g a table of specifications involves (1) selecting the learning outcomes
' B. Principles of achievement testing
to b tested, ~outlining the su~ject matter, andl.~) making a two-way chart. 1. Relation to instructional objectives
The o-way chart describes the sample of items to be included in the test. 2. Representative sampling
3. Relevance of items to outcomes
1\sel ting the Learning Outcomes to Be Tested. The learning outcomes for a 4. Relevance of test to use of results
J pa cular course \~'illdepend on the specific nature of the course, the objec- s. Reliability of results
3
tive attained in previous courses, the philosophy of the school, the special 6. Improvement of learning
42 Chapter 3 43
Planning for Assessment

C. Norm-referenced versus criterion-referenced testing The numbersin each cellof the table indicatethe number of test items
D. Planning the test to be devoted to that area. For example, 15 items in the test will measure
1. Determining the purpose of the test knowledge of terms; 4 of them pertain to the "role of tests" in instruction," 4
2. Identifying the intended learning outcomes to "principles of testing," 4 to "norm-referenced versus criterion-referenced,"
3. Preparing the'test specifications .md 3 to "planning the test." The number of items assigned to each cell is
4. Constructing relevant test items determined by the weight given to each learning outcome and each subject
m.ltter area.
Making the Two-Way Chart. When the learning outcomes have been se- A number of factors will enter into assigning relative weights to each
lected and cleariy defil1edand the course content outlined, the two-way chart learning outcome and each content area. How important is each area in the
should be prepared. This is called a table of specifications. It relates out- total learning experience? How much time was devoted to each area during
comes to content and indicates the relative weight to be given to each of the instruction? Which outcomes have the greater retention and transfer value?
various areas. As noted earlier, the purpose of the table is to provide assur- What relative importance do curriculum specialists assign to each area?
ance that the test will measure a representative sample of the learning out- These and similar criteria must be considered. In the final analysis, however,
comes and the subject matter topics to be measured. the weights assigned in the table should faithfully reflect the emphasis given
An example of a table of specifications for a summative test on a unit during instruction. In Table 3.3, for example, it is assumed that twice as
on achievement testing is given in Table 3.3. Note that only the general much emphasis was given to "planning the test" (20 items) as was given to
learning outcomes and the major subject matter categories have been "norm-referenced versus criterion-referenced" (10 items). Similarly, it is
included. A more detailed table may be desirable for test purposes, but this assumed that knowledge outcomes were given approximately two-thirds of
is sufficient for illustration. the emphasis during instruction (40 items) and that comprehension and
. application outcomes were each given approximately one-sixth of the total
emphasis (10 items each).
TABLE3.3 Table of Specifications for a Summative Test on a Unit In summary, preparing a table of specifications includes the following
for Achievement Testing steps:- . ----.------.

1. Identify the learning outcomes and content areas to be measured by the


I Knof!Js Total
\/
test. ,
~ ComPJ'ehends . Applies Number . 2. Weight the learning outcomes and content areas' in teqns of their rela-
.
L.C())ltent Facts Procedures Principles Principles of Items tive importance.
Role of Tests 4 4 2 10
. 3. Build the table in accordance with these relative weights by distributing
in Instruction the test items proportionately among the relevant cells of the table.

Principles of 4 3 2 6 5 20 The resulting_:ay tabl_indiate$Jhe type of test needed to m


Testing sure the learning outcomes and course coment in a balanced manner. Thus,
Norm- 4 3 3 the table of specifications serves the test maker like a blueprint It specifies
10
Referenced the number and the nature of the items in the test, and it thereby provides a
versus guide for item writing.
Criterion-
Referenced

Planning 3 5 5 2 5 20 Considerations in Constructing


the Test Relevant Test Items
Total
Number The construction of a set of relevant test items is greatly simplified if the
of Items 15 15 10 10 intended learning outcomes have beE:,11.dearly de.firu:s:l.
and fue test specifica-
10 60
tions carefully prepared. The quality of the test will fuen depeoo on how
44 Chapter 3 ,
Planningfor Assessment 45
\
closely the test maker can match the specifications. Here we shall confine our
mine if students can spell, have them spell from dictation. If you want to
discussion to some of the general specifications in preparing test items. More
determine if students can solve mathematics problems, have them solve
detailed procedures and rules for item writing will be described in the chap- problems and supply the answers. If you want to determine if students can
ters that follow.
write, have them write something. Use selection-type items for supply-type
outcomes only if there is a compelling reason for doing so (e.g., electromc
Selecting the Types of Test Items to Use scoring) and then take into account, during interpretation of the results, that
a less direct measure has been used. In some cases, of course, both types of
The items used in achievement tests can be classified as either selection-type
items are useful in the same area. For example, a writing project may pro-
items or supply-typeitems.The selection-typeitem presents students with a
vide the best evidence of writing skill but a selection-type test would pro-
set of possible responses from which they are to select the most appropriate
vide the most systematic coverage of the elements of grammar needed for
answer. The supply-type item requires students to create and supply their effective writing.
own answers. These two major categories can be used to classify the most
There are a number of achievement areas where either selection-type
widely used item types as follows.
items or supply-type items would measure equally well. In these cases, our
choicebetween them must be based on other item characteristics. The prepa-
!t~ms
S!}ectio!!:..!1I.pe ration of good selection-type items is difficult and students can get a propor-
1. Multiple choice
2. True-false tion of answers correct by guessing. However, these disadvantages are offset o../,v
3. Matching . by the fact that (1) they can ~_~cor~d,g~ckJY and oJ2j~ctively(i.e.. scorerS'1 1cs.
agree on the answers), (2) They eliminate bluffing, (3) they eliminate the '
4. Interpretive exercise
infIuenceof writing skill, (4) fuey provIde an extensive sample of student
perfonnance'(Decauseof the large number of items used), and (5) they pro-
SuppIY::IY~ems
l=-Snort answer vide for the ide~tifica~~~,~f spec~!e~~g_en:o~s. In comparison, suppry-
2. Essay (restricted response) type Items are easier to construct (although harder than commonly believed)
, bilt more difficult to score. The scoring of short-answer items is contami-
3. Essay (extended response)
, nated by answ~~ _.<?! varying degrees of correc1i1eS-saItdby ~ustrnen~
n¥'cIed'~or-.~spellings.-Tfiesc6fing c5fesSc\'Y"teStfiS1edious, time consum-
These categories are sometimes referred to as --reco~ition . and ----
recall ing, and influenced b}' bluffin~! writin~ slsill, and 'the shifting of standard$.,
items. This is a case of mislabeling that confuses the method of responding
with the mental reaction needed to make the response. When measuring ~~s~@1g..Another:major
ite sample of learning tasks that shortcoming of supply-type
can be measured. items is the
The short-answer lim-
item is
knowledge of facts, test responses might be limited to either the recognition
rest.!!,ctedprimarily to measuring knowled,g~_utcomes.Atthough the essay
or recall of the answer. However, when measuring complex learning out- test is especIally sUltea1O"'ii\easunng complex leamii\g outcomes, its sam-
comes with selection-type items, the answer is not achieved through mere pling is limited by the relatively few questions that can be included in a test.
recognition of a previously learned answer. It typically involves some use of A summary comparison of the relative merits of selection-type and
higher mental processes to arrive at a solution (e.g., verbal or mathematical
supply-type items is presented iI\ T~le-.3.4. Although thISshould serve as a
reasoning) before the correct answer can be selected. Similarly, a short- gwdeuiselecnng the types of items to usema given test, as noted earlier, the
answer item may require reasoning or problem solving rather than simply mostimportant question to ask is: Doesthis item typeprovidethemostdirectmea-
recalling and supplying factual information. Essay answers, of course, typi- sureEflhe intendedlearningoutcome?If the vanous item types are eqUalm this
cally require analysis, synthesis, and evaluation skills in addition to recall. reg~rd,theseiection-typEfiferiiSwould be favored because of the broad sam-
Usingthe selection and supplylabelsmakesclearhow theresponsesare made pling,the objective scoring, and the pinpointing of specific learning errors.
but it does not imply limits on the types of learning outcomes that can be
measured with each. .
In deciding which item types to use in a test, a guiding principle Matching Items to Specific Le~ing Outcomes
shou;ld be: Use the item types that p~!!v.~4~_t~e.mo.st~irec.t ~~~!!!.C?1!t...uden.t pgr- Effectiveachievementtestingrequiresthat a set of test itemsbe constructed
. h12!}l!.'1CfLsp.~CJfi..ei[]Jy"
~h£i!Iteii4.~4..leqrni7JgQu.f.£Q!l!e,Thus, if you want to deter- that calls forth the performance described in the intended learning out-
--

46 Chapter 3
Planning for Assessment

TABLE3.4 Summary a/the Relative Merits a/Selection-Type Items and Supply- .C. Define the intended learnirig-outcomes.
Type Items
,.., - D. Prepare the test specifications.
I., '. , . .~.
I ",. "
CIiilraderlstic
'.'." Supply-Type Items 5,'('ciftcLeamingOl/tcome:Identifies the hierarchical order of the categories in the
cognitive domain of the taxonomy.
Selection-Type Items ShortAnswer Essay
1. Which one of the following categories in the taxonomy indicates the high-
Measures factual information Yes Yes est level of learning?
Yes(.)
Measures understanding Yes No( ) Yes A. Analysis
Measures synthesis No( ) No( ) Yes B. Application
Easy to construct No Yes Yes C. Comprehension
Samples broadly Yes Yes No .0. Synthesis
Eliminates bluffing y~~ No No
Eliminates writing skill Yes No No
Eliminates blind guessing No Yes Specific Learning Outcome: Distinguishes between sound and unsound principles
Yes
Easy to score Yes No No of achievement testing.
Scoring is objective Yes No No Directions:Read each of the following statements. If the statement indicates
Pinpoints learning errors Yes Yes No a sound principle of achievement testing, circle the S; if it indicates an unsound
Encourages originality No No Yes principle, circle the U.

(.)The essay test can measure knowledge of facts, but because of scoring and sampling ~ U 1. The specific learning outcomeS to be tested should be stated in
problems it probably should not be used for this purpose. terms of student performance.
(")These items can be designed to measure limited aspects of these characteristics. S .U 2. Achievement testing should be limited to outcomes that can be
measured objectively.
~ U 3. Each achievement test item should measure a clearly defined
type of student performance.
'I'

comes. While we can never be certain of a perfect correspondence between Specific Learning Outcome: Identifies examples of properly stated learning outcomes.
outcome and item, the following examples illustrate how items should be 1. Which one of the following learning outcomes is properly stated in perfor-
mance terms?
written to measure the specific type of performance stated in the specific
learning outcome. A. Student realizes the importance of tests in teaching.
B. Student has acquired the basic principles of achievement testing.
EXAMPLES C. Student demonstrates a desire for more experience in test construction
"D. Student predicts the most probable effect of violating a test construc-
Specific Learning Outcome: Defines terms in student's own words. tion principle.
Directions:Define each of the following terms in a sentence or two.
1. Taxonomy It should be noted in these examples that each specific learning out-
2. Cognitive come provides a precise definition of the student performance to be mea-
3. Measurement
sured, and the test item simply provides a task that makes measurement of
4. Evaluation
the specified performance possible.

SpedficLearn~'!..g
Outcome:Iden~~.pr~~u!al steEs in p'!~g for a t~st.
1. an
Which one oftl1eJoTIowmg steps should be completed first in planning for Improving the Functioning Content of Items
achievement test?\
A. Select the types of test items to use. If test items are to call fortl1the performance described in the intended learn-
B. Decide on the length of the test. ingoutcomes, great care must be taken in phrasing the items. We need to elim-
inateall barriers that might prevent a knowledgeable person from responding
and all clues that might lead the uninformed to the correct answer. Only those
~ho have achieved the outcome being mE;~~red should get the it~m -righi:.
lThe correct answer is indicated throughout this book by an asterisk.
A!l_others(no matter how Intelligent) sh<?uldmiss it.
48 Chapter 3 PlanningforAssessment 49

Some of the common barriersto be.avoided during test preparation are: In the testing of elementary school students, the testing time typically
should be no more than 20 or 30 minutes so that proper motivation is main- r
v Vocabulary that is unnecessarily difficult. tained. At the high school and college levels, students can be given tests last- I
Sentence structure that is unnecessarily complex. ing several hours (for example, a final examination), but most tests are
'~Statements containing ambiguity. limited to a testing time of 40 to 50 minutes because of the length of the typ-
Unclear pictorial materials. . ical class period.
Directions that are vague. In matching the number of items to available testing time, we are
'- Material reflecting race, ethnic, or sex bias. faced with the problem of estimating how many items students can com-
plete per minute. Unfortunately, there are no simple answers. The size of
Awareness of ~~h.barriers during the planning and preparation.of.!he the final product depends on the type of test item, the complexity of the
~st !!'!heJi!st step intheiI:elimination. Essentially, we can avoid 'these barri- learning outcome measured, and the age of the students. As a guideline,
ers by (J..rWritingeach test item so that it presents a clearly formulated task, high school and college students should be able to answer one multiple-
v(2) stating the items in simple, clear language,13) keeping the items free from choice item, three short-answer itemS, or three true-false items per minute
biased and nonfunctional material, and {4)using a test format and directions when the items are measuring knowledge outcomes. For measuring more
that contribute to effective test taking. Much of the material presented later complex learning outcomes such as comprehension and application, and
in this book is directed toward constructing tests that prevent extraneous for testing younger age groups, more time per item is needed. In estimating
factors from distorting the test results. the number of items to be used, keep in mind the slower students in the
Someof the commoncluesto be avoidedduring test preparationare: group, for it is desirable to give all students an opportunity to complete the
test. Experience in testing a given group of students is frequently the only
Verbal associations that give away the answer. dependable guide for determining proper test length.
Grammatical inconsistencies that eliminate wrong answers. In addition to our concern with total test length, consideration must be
Specific determiners that make certain answers probable (e.g., some- given to the number of test items needed for each type of interpretation to be
times) and others improbable (e.g., always). made. This issue is especially crucial in criterion-referenced interpretation,
where we want to describe student performance in terms of each intended
Stereotyped or textbook phrasing of correct answers.
learning outcome. For this purpose we should use at least 10 items per out-
Length or location of correct answers. come. Where practical constraints make it necessary to use fewer than 10
Material in an item that aids in answering another item. items for an. intended outcome, only tentative judgments should be made
and these should be verified by other means. In some cases it is possible to
Just as in controlling extraneous factors that provide barriers to the cor- combine items into larger item clusters for a more meaningful interpretation.
rect answer, clues such as these can be eliminated by being aware of them SeeBox3.2 for a summary checklist for evaluating the test plan.
and by following sound principles of test construction. Many of the rules for
item writing and test preparation in the chapters that follow provide guide-
lines for this purpose. GeneralGuidelines for Item Writing
There are a number of general suggestions that apply to the writing of all
~etermining the Number of Test Items to Use item types. These provide a general framework for writing items that func-
The number of items to use should be indicated in the test specifications, as tion as intended and that contribute to more valid and reliable results.
noted earlier, and is modified by a number of practical constrai!'~, such as
the following: 1. Select the type of test item that measures the intended learning outcome
most directly. Use a suppI:Y~t}'peitem if supplyingthe answer is an irilpor--
.,../ . tant element of the task (e.g., writing). Use a selection-type item if appropri-
1. Age of the students tested
v2. Time available for testing ate (e.g.,identification) or if both types are equally appropriate.
~..'3. Type of test items used _ 2. Write the test item so that the performance it elicits matches the per-
4. Type of interpretation to be made t
formance in the learn~~g..!!!.~~:.
The intended learning-outcome spedft~ the

. ~1;
,,*f
-_. ---

50 Chapter 3
I'I,,,,,,m,'fi"
A",,~'"'''''' Stl
BOX 3.2 · Checklist for Evaluating the Test Plan quires alertness during item writing and reviewing the items after setting
them aside for a while. The most common clues for each item type will be con-
sidered in the following chapters. It is also important to prevent the informa-
, ~1;IS'tlU!pUffiose pftheteSfclear? tion given in one item from providing an answer to another item in the test.
,;2];;~v~ th~,intended'l~n:ung outcomes been i~entified and defined?
:3..~ tl1~.~terided learmng outcomes stated m performance (measurable) 7. Write the test item so that the difficulty level matches the intent of the
terms? learninJ{outc~, t!!!.agegr!!~l!l.l!.lie!~!ed, and th~se to ~!!!!aile cj[!~e.rt!':
,( Have testiSpecifications been prepared that indicate the nature and distri- suits. When difficulty is being evaluated, check to be certain that it is relevant
bution of items to be included in the test?
iothe intended learning outcom~ and that the item is free from sources of ir-
S. Does'thespecified set of Items provide a representative sample of the tasks
relevant difficulty (e.g., obscure materials, overly fine discriminations).
, ,,' ,~~tailte4 in the achievementdomain?
'types of i tems ap prop ,,,, fu<th e learning outcomes to be mea- 8. Write the test item so that there is !!Q...di§..aKTeement~o.!!...":!!!!':~ns,.!h..e.
"~I:I. ,~ .-th ?e "
!!"swer. Typically the answer should be one that experts would agree is the
I ',7.,Is t1}.edi#icUltyof
,
,
, ,
" , the
, , items
,, appropriate
,
" " for, the students to be tested and
correct or best answer. Most problems arise here when students are to pro-
I . th~~JUl~e of the measureII\f:l)t (e.g., J;ruISteryorsurvey)? vide the best answer (best procedure, best explanation). This involves a mat-
l 8. Isavailiibie
the n~ ,of items appropriate for.the students to be tested, the time
far testing, and theiriteipretitions to Demade?
ter of judgment and to be defensible the answer must be clearly best and
identified as such by experts in the area. Where experts disagree, it may be
I '~9~Qoes tl1e~tpJan include built-in features that contribute to valid and re-
desirable to ask what a particular authority wo~d consider to be the best
liable sCoreS?
10~Have plans been made for arranging the\items in the test, writing direc- method, the best reason, and the like. When attributed to a source, the answer
tions, scoring, and using the results? can be judgeq as correct or incorrect. .

9. Write the test items far enough in advance that they can be (B,!£r..re.-
p,iewedatid modified as.needed.A good time to write test items is shortly after
the material has been taught, while the questions and context are still clearly
in mind. In any event, reviewing and editing items after they have been set
learning task in performance terms and the test task should call forth the aside for a while can detect flaws that were inadvertently introduced during
same performance. the original i~emwriting.
,
3. Write the test item so that the test task is clea.r..iIJ1JL4efinitt;.
Keep the 10. Write !!}.or.e~est items.JlJan called for by the test plan. This will enable
reading level low, use simple and direct language, and follow the rules for you to discard weak or inappropriate items during item review and make it
correct punctuation and grammar. easier to match the final set of items to the test specifications.
4. Write the test item so that it is ftee from nonflmc;tional tI!..,!~eri~.l.
Mate-
rial not directly relevant to the problem being presented increases the read-
ing load and may detract from the intent of the item. Use extraneous material Arranging the Items in the Test
only where its detection is part of the task (e.g., in math problems).
After the final selection of the items to be included in a test, a decision must
5. Write the test item so t~at irr~evant factors do not prevent an info!!!}ed be made concerning the best arrangement of items. This arrangement will
~tudent from ,responding cp'rrectly. Avoid trick questions that might cause a vary with the type of test being prepared, but the following guidelines
knowledgeable student to focus on the wrong aspect of the task. Use clear, should be helpful.
unambiguous statements that maximize the performance to be measured and
minimize all other influences. For example, word problems measuring math- 1. For instructional purposes it is usually) desirable to group tOS!!fr5r
ematical reasoning should keep reading level and computational demands items that measure the same outcome. Euch group of items can then be
simple if an uncontaminated measure of reasoning ability is desired. iClentifiedby an appropriate heading (e.g., knowl~dge, comprehension,
6. Write the test item so that irrelevant clues do not enable the uninformed application). The inclusion of the headings helps to identify the areas
student to respond correctly. Removingun-wanted clues from test items re- where students are having difficulty and to plan ior remedial action.
52 Chapter 3 Planning for Assessment S3

2. Where pos~le,Jl1ljJJ:ms_of th~.sa",e type should be grouped together.


This arrangement makes it possible to provide only one set of directions
BOX 3.3 · C;orrectinf{ for GuessinK-

for each item type. It also simplifies the scoring and the analysis of the
results. Use .thecorrectionf~r guessing,form~~.ortly\'!h~.sfudeJj~'hay~~u$ci~L'
timeto consideraUof theitemsinthetest (e.g.~sp~.fe!;tl.i@P#{~m't\ifS:tUY'
3. The items should be arranged in terms ofjncreasing difficul~. This dents that therewillbe a c~~on.tor:gUE!SS4).~;TJ:~:th,~fOJl§W#1~~~>tB..,
arrangement has motivational effect on students and will prevent them
from getting "bogged down" by difficult items ear~yin the test.

Because most classroom tests include only a few item types, it is usu-
=-.::~..- Wrootf.-l
In thisformula,n ".,.
", ,. "';;';'i¥§0~:iJ
als the IU1Jnberofalterriatives'iIt each:funei~d}'i',' .
ally possible to honor all three of these guidelines. When this is not feasible,
the item arrangement that best fits the nature of the test and use to be made
of the results should be used. fu~~_;~
Multiple-choice items
'9"'S=J<~W ',' ,"

S'= R.,.W /2 (3 a1fe~t.iyesJ


j(~~~,
'/.;::' ~~.j~/I
's :!:R - W/3(~ alterpati~~r'. . <:k\~:
Preparing Directions
--
S=.R-W/4(5al~tiv~), 1.;
'

';.

The directions for an achievement test should be simple and concise and yet
contain information concerning each of the following: (1)purpose of the test, answer sheets, and the specific instructions will have to be adapted to the
(2) time allowed to complete the test, (3) how to record the answers, and particular type used. Unless machine scoring is to be used, however, a
(4) whet1:lerto guess when in doubt about an answer. The following sample teacher-made answer sheet that simply lists the letters of the alternatives for
directions for a multiple-choice test cover these four points. each item is usually satisfactory.

EXAMPLE ~ The Problem of Guessing


Directions:This is a test of what you have learned during the first five weeks of Inour set ofsampledirections,the studentswere told, "Sinc;eyour scorewill
~'course. The res~~.'pf this test will be.used to clarr~ any pointS of difficulty
md thus help you complete the course successfql)j:
.
..., - "-
There are 6O'multiple-choice items, and you have one hour to complete
be the number of items answered correctly,be sure to answer every item."
Thisis an attemptto equalizethevariationamongstudentsin theirtendency
the test. to guesswhen in doubt about the answer. The directionsmake it unneces-
Foreachitem,selectthe answerthat bestcompletesthe statement,or an- sary to correct for guessing. Although a corrPrtinnfor...guessinemay be
swers the question, and circle the letter of that answer. aJ2P!OPriate.for..~~ttiizprl tests, they are n~~~.~or ~~m tests
Sinceyour scorewillbe the number of itemsansweredcorrectly,besure where the students have an op'portuni~_ to respond to all ~ It is only
toonswaeveryitem.. where students are unable to complete the test (e.g., speed test) that the
correction for guessing might be appropriate for classroom tests. Here, the
When two or m e item types are included in the same test, it is usually purpose is to prevent students from rapidly and randomly marking the re-
desirable to provide eneral directions for the test as a whole and specific maiItingitems just before time is up in an attempt to improve their score (see
directions for each rt. When this is done, the general directions should Box3.3for correction formula).
contain the informa on about purpose, time allowed, a6d what to do about
guessing, and the ecific directions should describe how to record the
answers for that par .cular part. Also, some items, such as keytype exercises, Reviewing and Evaluating
require special dire .ons for each item. the Assembled Test
The use of se arate answer sheets requires some elaboration of the
instructions for rec rding the answers. If students are J.1otfamiliar with the Afterassembling the items into a test, it is desirable to review the test as a
use of separate ans er sheets,-it might also be desirable to present a sample whole to be sure it meets the criteria of a good test. The types of questions to
item with the corre t answer properly marked. There is a variety of separate considerare listed in Box 3.4.

-.---
54 Chapter 3 1'1111/11illS for ASSt'SS"'I'1I1 55

,1crossthe hole, This indicates to the student the correct answer for each item
missed. If machine scoring is to be used, simply scan the students' papers to
IdJillimce Do 'theitems measure a representative sample of the learning make certain that only one answer was marked for each item.
tasks
, in the aciu~vt:Jlit:1\tdomain? Unless corrected for guessing, a student's score on an objective test is
. .'2; Relewince Do the test items present relevant tasks? typically the number of answers marked correctly. Thus, each test item is
,3; Conciseness Are the t~t tasks stated in simple, clear language? counted as one point. Although teachers frequently desire to count some
.4. .Soundness
itemsmore heavily than others because of their importance or difficulty, such
Are the items of proper difficulty, free of defects, and do
they have answers that are defensible? weighting of scores complicates the scoring task and seldom results in an
Are the items free from overlapping, so that one item does
improved measure of achievement. b _~~!t~!...~~yto increase th~r~JpJj.'y~
not aid in answering another? \~eightof an area is to consg-!!~t~reitems in.~~.area.
ii..{\frangement Are items measuring the same outcome grouped together?
Are items of the same type grouped together?
Analyzing the Effectiveness of Test Items
I ,..,.r Are items in order of increasing difficulty?

·
.",.7..'Nuft/!1eJ"il')g.
.. S. DireCtfinis
. .,Aretheitems"numbered in order throughout the test?
Are there directions for the whole test and each part?
After a test has been administered, the items should be analyzed for their
effectiveness.One method of evaluating the test is to review the test item-by-
Are the directions concise and at the proper reading level? item when it is handed back to students for a discussion of the results. Sig-
.I Do the directions include time limits and how to record nificant commenfs made during the discus~;ioncan be recorded on a blank
answers? copy of the test. The flexibility of this procedure makes it possible to'pursue
Do the directions tell what to do about guessing? comments about particular items, to ask students why they selected the
9. Spacing Does the spacing on the page contribute to ease of reading response they did (both correct and incorrect selections), and to spend as
and responding? much time on each item as seems warranted. This procedure may bring forth
. Is the final copy free of typographical errors? some unjustified criticisms of the test, but it will also help identify defective
items in need of revision. The extra time U!K.>d. in discussing the test results
will not be wasted because stuaentS will be obtaining a review of the course

Administering and Scoring the Test


?iaterial covered by the ~~st: ,. ~.. . . ,
Simple Item Analysis Procedure
The administration of a carefully prepared informal achievement test is
largely a matter of prov!.~gp.!2E.er ~orking S.Q.I).ditiQ..~LJ<eep~g interrup- The discussion of test results and review OF the items can be much more
tions to a_minimum,and ~ging enough space between s~d~~~ t~.prev~t effectiveif a simple tally of student respons~ is made on the master copy of
cheating. The written directions should be clear enough to make the test sell- .the test. By recording the results of the 10 highest scoring students (H) and
aCliiiiiUstering,but in some situations it may be desirable to give the directions 10lowest scoring students (L), like that in FiglU'e3.2, make the results easily
orally as well. With young students a 1?lackboardillustration may also be use- interpretable. The answer is circled and the nt"mbers to the left of each alter-
ful. Above all, make certain that all the students know exactly what to do, and native indicate how many students in each group selected that alternative.
then provide them with the most favorable conditions in which to do it. The result in Figure 3.2 provides all the information we need to esti-
Scoring is facilitated if all answers are recorded on the left side of each mate the following types of item analysis information.
test page, as we suggested earlier. Under this arrangement, sOOriilgis-simply
a matter of marking the correct answers on a copy of the test and placing it 1. The difficulty of the item (percentage of 3tudents answering the item
next to the column of answers on each student's paper. If a separate answer correctly).
sheet is used, it is usually better to punch out the letters of the correct 2. The discriminating power of the item (h~, well the item discriminates
answers on a copy of the answer sheet and use this as a scoring stencil. The between high and low scorers).
stencil is laid over each answer sheet and the correctly marked answers 3. The effectiveness of each alternative (all should be chosen and each one
appear through the holes. Where no mark appears, a red line can be drawn more frequently by the low scoring group:'.
r
56 Chapter 3 Planningfor AsseSsment 57

3. Obtain discriminatingpowerby subtracting the percentage correct in the


UNIT TEST low group trom the percentage correct in the high group. Add a deci-
Name Date mal point to the answer.
80 -40 =.40
Directions:
This test measures the knowledge, understandings, and applications you The description of the procedure makes it sound more complicated
have acquired during the first four weeks. than it is, but making a few mental calculations like this will reveal its sim-
There are 65 objective items, and you will have the full claf;speriod to com- plicity. Item difficulty typically uses the percent sign and, of course, can
plete the test. range from 0% to 100%~Ourdifficulty ind!.x is based on the high- and low-
Select the best answer for each item and circle the letter of that answer.
scoring groups only, but this provides a satisfactory approximation of item
Your score will be the nQIIlberof items answered correctly, so besure to an-
difficulty. The discrimination index typically uses the decimal point and
swer every item.
thus ranges from 0 to i:OO:-- -". -~ .
KNOWLEDGE OF TERMS This simple method <>.f analyzing the test results can be used as an aid
H L 1. An assessment instrument is properly classified as objective when in discussing the items when reviewing the test. Items that most students
1 3 A. the instrument uses objective-type questions. answer correctly can be skipped over or treated lightly, items missed by
1 2 B. the responses called forth by the instrument are free of opinion. most students can be discussed. in more detail, and defective items can be
0 1 C. those preparing the instrument used standard procedures of pointed out to students rather than defended as fair. Also, the frequency
construction. . with which each incorrect answer is chosen may reveal common errors and
8 4 @ there is agreement among scorers concerning the correctness misconceptions that can be corrected on the spot or serve as a basis for reme-
of the answers. dial work. In discussing the reasons students have selected the right or
wrong answer for an item can sometimes be revealing. We assume that if
FIGURE 3.2 Test copy with item analysis data to left of items (H = 10 highest scorers, they selected the correct answer they had learned the material. Not necessar-
L = 10 lowest scorers). ily so, when using selection-type items (see Box 3.5 for common reasons for
selecting alternatives). Having students discuss the reasons for their choices
By simply looking at the answer sheet in Figure 3.2, we can see that
8 students in the high-scoring group and 4 students in the low-scoring group
selected the correct answer. Also, more students in the low-scoring group
vBoX 3.5 · Common Reasonsfor Selecting Multiple-Choice
selected each wrong alternative than in the high-scoring group, indicating Alternatives
that the distracters seem to be distracting those students who haven't
learned the meaning of the term. '\1: ~ ~ ..:-,
Because we are using 10 students in the high-scoring group and 10 stu- Basesfor Correct Choice '\ "''Ii",.".,
J1.
' '
" ~""i.,f

dents in the low-scoring group, we can mentally calculate the usual indexes L Possesses required information o~skill.
2. Uses partial information that favors answer. 1. ~~.:;,~:':(
, , , 1
of difficulty and discriminating power by using the following simple steps.
3. Uses clues given In the item. . 1
4. Uses Information from other itemsiIithe test.. ;¥1rr'
1. Determine the percentage of high scorers ~nd low scorers passing the , -.';,:.;""...', r7:>\
item by adding a zero. - .
5. Makes a blind, lucky guess. \
Yr
I
r' I
H =8 out of 10 = 80 Basesfor Incorrect Choice .
L =4 out of 10'= 40 1. Lacks the information or skill required by the:~tem.
. I
2. Uses partial information that favors a distracter;,~
2. Obtain i!!!!!!!iffi~.l.!.Y-
by adding the percentage correct in the high and 3. Uses misinformation that favors a distracter. N 'j J
low groups and dividing by 2. Add a percent sign to the answer. 4. Makes a blind, unlucky guess.
S. Marks wrong answer through 'carelessness.
'Ti '"
4.j II
80+4°=60%
2
I'III////i//gfllr ASSl'ssII/l'IIt 59
58 Chapter 3

provides insights into student thinking that may contribute to both improved 9. The diffi':ulty of a test item should match the difficulty of the Jeaming
teaching and improved test construction skills. task to be measured. Beware of irrelevant sources of difficulty (e.g.,ob-
Although the method of item analysis presented here is a crude one, it scure material).
is probably satisfactory for most classroom tests. We are using it primarily 10. An achievement test should be short enough to permit all students to at-
as a basis for reviewing the items with students and for any insights the tempt all items during the testing time available.
results might give us for improving test items in the future. A more detailed 11. A test should contain a sufficient number of test items for each type of
description ~f_it~!l1analysis and cautions in interpreting ite:IT.i~aIysis data interpretation to be made. Interpretations based on fewer than 10 items
can be found in Linn and Gronlund (~OOO). should be considered highly tentative.
- Because'oHhe small iuimber of students involved in classroom testing, 12. Following a general set of guidelines during item writing will result in
item-analysis information must be interpreted with great caution. If a test higher quality items that contribute to the validity and reliability of the
item is reused with a new group, the results may be quite different, due to test results.
changes in the instruction, the study habits of students, or some other factor. 13. Item arrangement within the test will vary with the type of test used.
The tentative nature of the item-analysis data is not of great concern, how- Where possible, items should be grouped by major learning outcome
ever, when used to review tests with students for the purpose of improving (e.g., knowledge, comprehension, application); similar items should
learning. be grouped together and should be arra~\ged in order of increasing
difficulty. .
/
14. Test directions should dearly indicate the purpose of the test, the time
allowed, how to record the answers, and to answer every item.
/' Summary of Points 15. A correction for guessing should be used for speed tests only.
1. Assessment planning should be guided by what the students are ex- 16. Review and evaluate the assembled test before using.
pected to learn as specified by the school goals and the more specific in- 17. Administer the test under controlled conditions.
structional objectives. . 18. Item analysis provides a means of evaluating the items and is useful in
2. Assessment planning requires a consideration of all possible types of reviewing the test with students.
learning outcomes in a given content area, not just those that can be
measured by a test.
3. Assessment of performance skills and products typically requires some Referencesand Additional Reading.
type of observation procedure, such as a checklist, rating scale, or holis- Bloom, B.S., et al. (ed.). Taxonomy of Educa- (Upper Saddle River, NJ: Merrill/
tic scoring rubric (see Chapter 7). Prentice-Hall, 2000).
tional Objectives: Cognitive Domain
.~ j
4. Preparing and using an achievement test includes specifying the in- (New York: David McKay Co., 1956). Qosterhoff, A. c., Classroom Application of Ed-
ucational Measurement, 3rd ed. (Upper
structional objectives in performance terms, preparing a table of speci- Carey, L. M., Measuring and Evaluating School
Saddle River, NJ: Merrill/ Prentice-
fications, constructing relevant test items, arranging the items in the Learning, 3rd ed. (Boston: Allyn and
Bacon, 2001). Hall, 2001).
test, preparing clear directions, reviewing and evaluating the assembled Gronlund, N. E., How to Write and Use In- Stiggins, R. J., Student-Involved Classroom As-
test, administering the test,.;>nnm.a1<ingan item analysis. structional Objectives, 6th ed. (Upper sessment, 3rd ed. (Upper Saddle River,
5. Test specifications typically consist of a twofold table of specifications Saddle River, NJ: Merrill/Prentice NJ: Merrill/Prentice-Hall, 2001).
that indicates the sample of performance tasks to be measured. Hall, 2000).
6. The types of test items used in a test should be determined by how di- Linn, R. L., and Gronlund, N. E. Measurement
and Assessment in Teaching, 8th ed.
rectly they measure the intended learning outcomes and how effective
they are as measuring instruments.
7. Each test item should provide a task that matches the student perfor-
mance described in a specific learning outcome.
8. The functioning content of test items can be improved by eliminating ir-
relevant barriers and unintended clues during item writing.

,
Chapter 11

References and Additional Reading


American Educational Research Association, Lyman,H. B.,TestScoresand What They Mean,
Standardsfor Educational and Psycholog- . 6th ed. (Boston: Allyn and Bacon,
ical Testing (Washington, DC: AERA, 1998).
1999). Oosterhoff, A. C. Classroom
ApplicationsofEd-
Chase, C. I., Contemporary Assessmentfor Edu- ucationalMeasurement, 3rd ed. (Upper
cators (New York: Longman, 1999). Saddle River, NJ: Merrill/Prentice-
Linn, R..L., and Gronlund, N. E., Measurement Hall,2(01).
and Assessment in Teaching, 8th ed.
(Upper Saddle River, NJ: Merrill!
Prentice-Hall, 2000).

You might also like