Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
ANALYSIS
Technique to improve test
items and instruction
TEST DEVELOPMENT PROCESS
5. Construct Table of
Specifications
6. Develop Test
9. Assemble Operational Test Design
Forms
10. Produce Printed Tests Mat. 7. Develop New Test Questions
8. Review Test Questions
WHAT IS ITEM ANALYSIS ?
process that examines student
responses
to individual test items assessing quality
items and the test as a whole
valuable in improving items which will
be
used again in later tests and
eliminate ambiguous or misleading
items
valuable for increasing instructors' skills in
test construction, and
identifying specific areas of course content
which need greater emphasis or clarity.
SEVERAL PURPOSES
• Individual level:
• isolate specific errors the students made
2. Build future tests, revise
test items to make them
better
• know how much work in writing good
questions
• SHOULD NOT REUSE WHOLE TESTS -->
diagnostic teaching means responding to
needs of students, so after a few years a
test bank is build up and choose a
tests for the class
• can spread difficulty levels across your
blueprint (TOS)
3. Part of continuing professional
development
• doing occasional item analysis will help
become a better test writer
Characteristics:
1. The intercorrelations among the items --
the greater/stronger the relative number
of positive relationships are, the
greater
the reliability.
• Stability
1. Test – Retest
• Stability
2. Inter – rater / Observer/ Scorer
• applicable for mostly essay questions
• Use Cohen’s Kappa Statistic
• Equivalence
3. Parallel-Forms/ Equivalent
Used to assess the consistency of the results of
two tests constructed in the same way from the
same content domain.
• Internal Consistency
• Used to assess the consistency of results across
items within a test.
4. Split – Half
• 5. Kuder-Richardson
Formula 20 / 21
Correlation is determined from a
single administration of a test
through a study of score variances
• 6. Cronbach's Alpha (a)
Reliability
Interpretation
Indices
Excellent reliability; at the level of the best standardized
.91 and above
tests
.81 - .90 Very good for a classroom test
Item Difficulty
Percent answering correctly
Item Discrimination
How well the item "functions“
How “valid” the item is based on
the total test score criterion
WHAT IS A WELL-FUNCTIONING
TEST ITEM?
(DIFFICULTY
)
• which students got it
correct?
(DECRIMINATION
)
THREE IMPORTANT INFORMATION
ON QUALITY OF TEST ITEMS
• Item difficulty: measure whether an item
was too easy or too hard.
>=80
<= 30% > 30% AND < 80%
%
0 10 20 30 40 50 60 70 80 90
100
ITEM DIFFICULTY LEVEL: SAMPLE
QUESTIONS/DISCUSSION
• Is a test that nobody failed too
easy?
• Is a test on which nobody got 100%
too difficult?
• Should items that are “too easy” or
“too difficult” be thrown out?
ITEM
DISCRIMINATION
• Traditionally, using high and low scoring groups
(upper 27 % and lower 27%)
• Computerized analyses provide more accurate
assessment of the discrimination power of
items since it accounts all responses rather
than just high and low scoring groups.
• Equivalent to point-biserial correlation. It
provides estimate the degree an individual item is
measuring the same thing as the rest of the
items.
WHAT IS ITEM
DISCRIMINATION
?
• Generally, students who did well on the
exam should select the correct answer to
any given item on the exam.
• The Discrimination Index distinguishes for
each item between the performance of
students who did well on the exam and
students who did poorly.
INDICES OF DIFFICULTY AND
DISCRIMINATION
(BY HOPKINS AND ANTES)
Section 4 6 10 20
A (1,3,7,9)
Section 5 5 4 14
B (2,5,8,11,15)
Section 3 7 6 16
C (6,17,21)
Total 12 18 20 50
STEPS IN ITEM ANALYSIS
1 1 0 1 1 1 0 0 0 0 1 1 1 0 0 0 0 0 1 1
2 1 1 0 1 1 1 0 1 1 1 0 1 1 1 1 0 1 1 1
3 0 0 0 1 0 0 0 1 0 0 0 1 1 1 1 1 1 1 0
4 0 1 0 0 0 1 0 0 0 1 0 0 1 0 0 0 1 0 0
5 1 0 1 1 1 0 1 1 1 0 1 1 0 1 1 1 0 1 0
6 1 1 1 1 1 1 1 1 1 1 1 1 1 0 1 1 1 0 1
7 0 0 1 0 0 0 1 0 0 0 1 0 0 0 1 0 0 0 1
8 1 1 0 1 1 1 0 1 1 1 0 1 1 0 0 0 1 0 0
2. IN SPSS:
Analyze-Scale-Reliability
analysis – (drag/place variables
to Item box) – Statistics –
Scale if item deleted – ok.
• ****** Method 1 (space saver) will be used for this analysis ******
• R E L I A B I L I T Y A N A L Y S I S - S C A L E (A L P H
• Item-total Statistics A)
• Scale Scale Corrected
• Mean Variance Item- Alpha
• if Item if Item Total if Item
• Deleted Deleted Correlation Del
• VAR00001 14.4211 127.1053 .9401 eted
.9502
• VAR00002 14.6316 136.8440 .7332 .9542
A 4 1 6 3 10 3
B 5 3 5 3 4 2
C 3 2 7 4 6 3
Total 12 6 18 10 20 8
% 50% 56% 40%
Over 24/50 = 48%
all
• Realistically: Do item analysis to
your most important tests
• end of unit tests, final exams -->
summative evaluation
• common exams with other teachers
(departmentalized exam)
• common exams gives
bigger sample to work with, which
is good
• makes sure that questions other
teacher s prepared are
working for your class
ITEM ANALYSIS is one area where
even a lot of otherwise very
good classroom teachers fall
down:
• they think they're doing a good job;
• they think they've doing good
evaluation;
• but without doing item
analysis,
• They don’t really know.
ITEM ANALYSIS is not an
end in itself,
• no point unless you use it
to revise items, and
• helps students on the basis
of information you get out
of it.
END OF PRESENTATION…
THANK U FOR
LISTENING…
H AV E A RELIABLE A N D
ENJOYABLE DAY….