Sei sulla pagina 1di 9

Critical Thinking Assessment

Author(s): Robert H. Ennis


Source: Theory into Practice, Vol. 32, No. 3, Teaching for Higher Order Thinking (Summer,
1993), pp. 179-186
Published by: Taylor & Francis, Ltd.
Stable URL: http://www.jstor.org/stable/1476699
Accessed: 18-05-2015 22:26 UTC

REFERENCES
Linked references are available on JSTOR for this article:
http://www.jstor.org/stable/1476699?seq=1&cid=pdf-reference#references_tab_contents

You may need to log in to JSTOR to access the linked references.

Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at http://www.jstor.org/page/
info/about/policies/terms.jsp

JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of content
in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms of scholarship.
For more information about JSTOR, please contact support@jstor.org.

Taylor & Francis, Ltd. is collaborating with JSTOR to digitize, preserve and extend access to Theory into Practice.

http://www.jstor.org

This content downloaded from 50.155.233.239 on Mon, 18 May 2015 22:26:22 UTC
All use subject to JSTOR Terms and Conditions
Robert H. Ennis

Critical Thinking Assessment

LTHOUGHcritical thinking has often been urged Defining Critical Thinking


as a goal of educationthroughoutmost of this The upperthree levels of Blooms' taxonomy of
century (for example, John Dewey's How We Think, educationalobjectives (analysis, synthesis, and eval-
1910; and the EducationalPolicies Commission's The uation) are often offered as a definition of critical
Central Purpose of American Education, 1961), not a thinking.Sometimes the next two levels (comprehen-
great deal has been done about it. Since the early sion and application)are added. This conception is a
1980s, however, attentionto critical thinkinginstruc- good beginning, but it has problems. One is that the
tion has increased significantly-with some spillover levels are not really hierarchical,as suggested by the
to critical thinking assessment, an area that has been theory, but rather are interdependent.For example,
neglectedeven morethancriticalthinkinginstruction. although synthesis and evaluation generally do re-
Partly as a result of this neglect, the picture I quire analysis, analysis generally requires synthesis
paint of critical thinking assessment is not all rosy, and evaluation (Ennis, 1981).
though there are some bright spots. More explicitly, More significantly, given our concern here, the
my major theme is that, given our current state of three (or five) concepts are too vague to guide us in
knowledge, critical thinking assessment, albeit diffi- developing and judging critical thinking assessment.
cult to do well, is possible. Two subthemes are that Consider analysis, for example. What do you assess
(a) the difficulties and possibilities vary with the pur- when you test for ability to analyze? The difficulty
pose of the criticalthinkingassessmentand the format becomes apparentwhen we consider the following
used, and (b) thereare numeroustrapsfor the unwary. varietyof thingsthatcan be labeled"analysis":analysis
In pursuitof these themes, I consider some pos- of the political situationin the Middle East, analysis
sible purposes in attemptingto assess critical think- of a chemical substance,analysis of a word, analysis
ing, note some traps, list and comment on available of an argument,and analysis of the opponent'sweak-
critical thinking tests (none of which suit all of the nesses in a basketball game. What testable thing do
purposes), and finish with suggestions for how to de- all these activities have in common?None, except for
velop your own critical thinking assessment, includ- the vague principle that it is often desirable to break
ing a discussion of some major formats. But first, things into parts.
some attentionmust be paid to the definition of criti- A definition of critical thinking that I at one
cal thinking,becausecriticalthinkingassessmentrequires time endorsed is that critical thinking is the correct
thatwe be clearaboutwhat we are tryingto assess. assessing of statements (Ennis, 1962). If I had not
Robert H. Ennis is professor of education at the Univer- elaborated this definition, it would be as vague as
sity of Illinois at Urbana-Champaign. Bloom's taxonomy.But even whenelaborated,it suffers

THEORYINTO PRACTICE,Volume 32, Number 3, Summer 1993


Copyright 1993 College of Education, The Ohio State University

This content downloaded from 50.155.233.239 on Mon, 18 May 2015 22:26:22 UTC
All use subject to JSTOR Terms and Conditions
THEORYINTO PRACTICE/ Summer 1993
Teachingfor Higher Order Thinking

from excluding creative aspects of critical thinking, to test for such importantthings as being open mind-
such as conceiving of alternatives, formulatinghy- ed, and many even fail to test for judging the credi-
potheses and definitions, and developing plans for bility of sources. Without some defensible concep-
experiments.I now think the contemporaryconcep- tion of critical thinking, judgments about tests are
tion of critical thinking includes these things, so the likely to be erratic-or worse.
"correct assessing" definition is more narrow than Two other well-known definitions of critical
standardusage, and thus could interfere with com- thinkingare McPeck's "reflectiveskepticism"(1981,
municationamong proponentsof critical thinking. p. 7) and Paul's "strong sense" definition (1987).
The following definition seems to be more in Paul's definition is similar in broad outline to the
accord with contemporaryusage and thus, I hope, definitionproposedhere, but emphasizesmoreheavily
will minimize confusion in communication:"Critical being aware of one's own assumptions and seeing
thinking is reasonablereflective thinking focused on things from others' points of view. However, neither
deciding what to believe or do." As it stands,howev- of these definitionsprovidessufficient elaborationfor
er, this definition is also as vague as Bloom's taxon- developing critical thinking tests. Furthermore,
omy. It too needs elaboration.Here is an abridgment McPeck's definitionis negative.Criticalthinkingmust
of the elaborationsI have providedand defendedelse- get beyond skepticism.
where (Ennis, 1987, 1991, in press):
In reasonably and reflectively going about de- Purposes of Critical Thinking Assessment
ciding what to believe or do, a person characteristi- Not only must we have a defensible elaborated
cally needs to do most of these things (and do them definition of critical thinking when selecting, criti-
interdependently): cizing, or developing a test, we must also have a
1. Judgethe credibilityof sources. clear idea of the purpose for which the test is to be
2. Identifyconclusions,reasons,and assumptions. used. A varietyof possible purposesexist, but no one
3. Judgethe qualityof an argument,includingthe ac- test or assessmentprocedurefits themall. Herearesome
ceptabilityof its reasons,assumptions,andevidence. majorpossiblepurposes,accompaniedby comments:
4. Develop and defenda positionon an issue. 1. Diagnosing the levels of students' critical
5. Ask appropriate clarifyingquestions.
6. Planexperimentsandjudge experimentaldesigns. thinking. If we are to know where to focus our in-
7. Define termsin a way appropriate for the context. struction,we must "startwith where they are"in spe-
8. Be open-minded. cific aspects of critical thinking.Tests can be helpful
9. Try to be well informed. in this respect by showing specific areas of strength
10. Drawconclusionswhenwarranted, butwithcaution. and weakness (for example, ability to identify as-
This interdependentlist of abilities and disposi- sumptions).
tions can provide some specificity for guiding criti- 2. Giving studentsfeedback about their critical
cal thinking testing. The elaborations,of which the thinking prowess. If students know their specific
list is an abridgment,are more thorough,but the sim- strengthsand weaknesses, their attemptsto improve
plicity of this list can make it useful. It can serve as a can be better focused.
set of goals for an entire critical thinkingcurriculum 3. Motivating students to be better at critical
or as a partialset of goals for some subject-matteror thinking.Thoughfrequentlymisusedas a motivational
other instructionalsequence. It can be the basis for a device, tests can and do motivate students to learn
table of specificationsfor constructinga criticalthink- the materialthey expect to be covered on the test. If
ing test. (A table of specifications provides the areas critical thinking is omitted from tests, test batteries,
that a test is supposed to assess and indicates the or other assessment procedures,studentswill tend to
weighting assigned to each.) neglect it (Smith, 1991; Shepard, 1991).
The elaborationalso can be used as a guide in 4. Informingteachers about the success of their
judging the extent to which an existing critical think- efforts to teach students to think critically. Teachers
ing test is comprehensive, and whether it assesses can use tests to obtain feedback about their instruc-
critical thinking at all. One of my chief criticisms of tion in critical thinking.
most existing critical thinking tests is their lack of 5. Doing researchabout critical thinkinginstruc-
comprehensiveness.For example, they typically fail tionalquestionsand issues.Withoutcarefulcomparison

180

This content downloaded from 50.155.233.239 on Mon, 18 May 2015 22:26:22 UTC
All use subject to JSTOR Terms and Conditions
Ennis
CriticalThinkingAssessment

of a variety of approaches,the difficult issues in crit- are actually different tests. Comparabilityis always
ical thinking instructionand curriculumorganization suspect, since so much depends on the specific con-
cannot be answered. But this research requires as- tent of the test.
sessment, so that comparisonscan be made. 4. Most critical thinking tests are not compre-
6. Providing help in deciding whethera student hensive, especially those that are easiest to use, the
should enter an educationalprogram. People in some multiple-choicetests. These tests typically miss much
fields already use assessed critical thinking prowess that is importantin critical thinking.
to help make admissions decisions. Examples are 5. Another problem in the use of (especially)
medicine, nursing,law, and graduateschool in gener- multiple-choicetests lies in differencesin background
al. The idea seems good, but the efficacy of existing beliefs and assumptionsbetween test maker and test
efforts in selecting bettercriticalthinkershas not been taker. Since a critical thinkeremploys a grasp of the
established. Researchneeds to be done in this area. situation, different beliefs about the situation can
7. Providing informationfor holding schools sometimes result in justifiably different answers to
accountablefor the critical thinkingprowess of their test questions (see Norris & Ennis, 1989).
students. A currentlypopularpurpose for testing, in- 6. Significant results may be expected in too
cluding critical thinkingtesting, is to pressureschools short a time period. Learningto think critically takes
and teachers to "measure up" by holding them ac- a long time. Much reflective practice with many ex-
countable for the test results of their students. amples in a variety of situationsis required.
Purposes 6 and 7 typically constitute "high- 7. High-stakespurposes often interferewith the
stakes"testing, so called because much often depends validity of a test. This is partlybecause they motivate
on the results. The science reasoning section of the cram-schools, which teach students how to do well
American College Test (ACT),much of the new Med- on thetestswithoutthe students'havingthecriticalthink-
ical Colleges AdmissionsTest (MCAT),College Board ing prowessfor whichthe test is supposedlytesting.The
Advanced Placement (AP) tests, Iowa Test of Educa- studentsoften learntricksfor takingthe tests.
tional Development,and the analytic and logical rea- This interference with validity occurs also in
soning sections of the GraduateRecord Examination part because the high-stakes situation pressures the
(GRE) and the Law School Aptitude Test (LSAT)are test makers to avoid taking risks with items, the an-
examples of high-stakes critical thinking tests. swers to which might be subject to challenge. So the
pressureis for them to limit their testing to multiple-
Traps choice deductive-logic items of various sorts, that is,
In pursuingthe above purposes, educatorsneed items in which the conclusionnecessarilyfollows from
to be aware of several traps, including the following: the premises (thus limiting the test's comprehensive-
1. Test results may be compared with norms, ness and content validity). Deductive-logic items are
and the claim made that the difference, or similarity, the most immuneto complaintaboutthe keyed answer.
is the result of instruction. There are usually other 8. Scarce resources (indicated by low assess-
possible explanationsof the result, such as neighbor- ment budgets and overworkedteachers)often lead to
hood influences.Currently-popular accountabilitytest- compromisesthat affect test validity. Because of the
ing invites us into this trap. expense of, and/orteacher gradingtime requiredfor,
2. A pretestand a posttestmay be given without the tests necessary to assess critical thinking, many
comparingthe classto a controlgroup.The lackof a con- testingprogramshave resortedto multiple-choicetests
trolgrouprendersthe pretest-to-posttestresultsdubious, that are arguablyless valid than short answer, essay,
since many things otherthan the instructionhave hap- and performancetests of critical thinking.
pened to the students,and could accountfor the results.
3. The use of the same test for the pretest and Published Critical Thinking Tests
posttest has the problem of alerting the students to Although a numberof tests incorporatecritical
the test questions. On the other hand, the use of dif- thinking (including the high-stakes tests just men-
ferent forms of (allegedly) the same test for pretest- tioned), only a few have critical thinking (or some
posttest comparisons,given that the testing is for crit- aspect of critical thinking) as their primaryconcern.
ical thinking,is probablyworse, since differentforms None exist for studentsbelow fourth grade.

181

This content downloaded from 50.155.233.239 on Mon, 18 May 2015 22:26:22 UTC
All use subject to JSTOR Terms and Conditions
THEORYINTO PRACTICE/ Summer 1993
Teachingfor Higher Order Thinking

This dearthof critical thinking tests is unfortu- mal judgment about the validity of the content. Per-
nate; many more are needed to fit the various situa- sons who are seriously considering using any test
tions and purposesof critical thinkingtesting. In Ta- should take the test and score it themselves. There is
ble 1, I have attemptedto identify all currentlyavail- no better way to get a feel for the test's content va-
able publishedtests thathave criticalthinkingas their lidity. One should not depend solely on the name
primaryconcern. The tests are groupedaccordingto given to the test by the author and publisher. The
whether they aim at a single aspect of critical think- following questions should be considered:
ing or more than one aspect. The essay test is more 1. Is the test basedon a defensibleconceptionof criti-
comprehensivethan the others. cal thinking?
It would also make sense to group the tests ac- 2. How comprehensiveis its coverageof this concep-
cording to whether they are subject specific or gen- tion?
eral-contentbased. Subject-specific critical thinking 3. Does it seem to do a good job at the level of your
tests assess critical thinkingwithin one standardsub- students?
ject matterarea, whereas general-content-basedcriti- Though these questions might seem obvious, they
cal thinkingtests use content from a variety of areas are often neglected.
with which test takers are presumed to be already In varying degrees, all of the listed tests can be
familiar. A committee of the National Academy of used for the first five purposes specified earlier (all
Education has recommendedthat there be a strong but the high-stakes purposes). Their use for high
effort to develop subject-specifichigher orderthink- stakes is problematicfor two reasons:(a) There is no
ing tests (The Nation's Report Card, 1987, p. 54). A security on the tests, so prospective examinees can
full understandingof any subjectmatterarearequires secure copies, and (b) most of the tests are not suffi-
that the person be able to think well in that area. ciently comprehensive to provide valid results in a
Regrettably,I can find no subject-specificcriti- high-stakes situation. Let me elaborate on this sec-
cal thinking tests (that is, tests whose primarypur- ond problem.
pose is to assess critical thinking in a subject matter As indicated earlier, existing multiple-choice
area), althoughparts of some tests (such as the ACT tests do not directly and effectively test for many
section on science reasoning) fit this criterion. So
significant aspects of critical thinking, such as being
there is no subject-specificgroupingin this listing of
open mindedand drawingwarrantedconclusions cau-
tests primarilycommittedto critical thinking. All of
the tests listed here are general-content-basedtests. tiously. In response to this problem, some people
will hold that the various aspects of critical thinking
Unfortunately,the NationalAcademy committee are correlatedwith each other, so the lack of direct
also recommendedthe neglectof general-content-based
testing of specific aspects does not matter.For exam-
higherorderthinkingtests (p. 54). This is a mistake.We
need general-content-based tests to checkfor transferof ple, being open minded correlateshighly with judg-
criticalthinkinginstructionto everydaylife, regardless ing the credibilityof sources and identifyingassump-
of whetherthinkinginstructionis embeddedin subject tions, making all of these good indicatorsof the oth-
matterinstructionor whetherit is offeredin a separate ers, so the argumentgoes.
courseor unit, or some combinationof the two. However, when the stakes are high, people pre-
Since I am a coauthor of some of the listed pare the contentareasthatareexpectedto be on the
for
tests, my conflict of interest in presenting and dis- tests. Even thoughthese contentareasmight correlate
cussing this listing is obvious. I have tried not to let highly with other critical thinking aspects when the
it interfere with my objectivity, but do recommend stakes are low, specific preparationfor the expected
Arter & Salmon's Assessing Higher Order Thinking aspects in order to deal with them on the tests will
Skills: A Consumer's Guide (1987), which provides lower the correlations, destroying their validity as
more extensive coverage. A generaldiscussion of the indirect measures of the missing aspects of critical
problems,prospects,and methodsof critical thinking thinking. The danger is to accept correlations ob-
testing can be found in Evaluating Critical Thinking tained in low-stakes situations as representativeof
(Norris & Ennis, 1989). the correlationsobtainablein a high-stakessituation.
Since statistical informationabout tests can be A possible exception to this warning about the
misleading, it is importantto make one's own infor- use of the listed tests for high-stakes situationsis the

182

This content downloaded from 50.155.233.239 on Mon, 18 May 2015 22:26:22 UTC
All use subject to JSTOR Terms and Conditions
Ennis
Critical ThinkingAssessment

Table 1
An Annotated List of Critical Thinking Tests

Tests Covering More Than One Aspect of Critical assumptionidentification,word relationships,sentence se-
Thinking quencing, interpretinganswers to questions, information
The California Critical Thinking Skills Test: Col- sufficiency and relevance in mathematics problems, and
lege Level (1990) by P. Facione. The California Aca- analysis of attributesof complex stick figures.
demic Press, 217 LaCruz Ave, Millbrae, CA 94030. Test of Enquiry Skills (1979) by B.J. Fraser. Aus-
Aimed at college students, but probably usable with ad- tralian Council for Educational Research Limited, Fred-
vanced and gifted high school students. Incorporates in- erick Street, Hawthorn, Victoria 3122, Australia. Aimed
terpretation,argument analysis and appraisal, deduction, at Australian grades 7-10. Sections on using reference
mind bender puzzles, and induction (including rudimen- materials (library usage, index, and table of contents);
tary statistical inference). interpreting and processing information (scales, averag-
Cornell Critical Thinking Test, Level X (1985) by es, percentages, proportions, charts and tables, and
R.H. Ennis and J. Millman. Midwest Publications, PO graphs); and (subject-specific) thinking in science (com-
Box 448, Pacific Grove, CA 93950. Aimed at Grades 4- prehension of science reading, design of experiments,
14. Sections on induction, credibility, observation, de- conclusions, and generalizations).
duction, and assumption identification. Test of Inference Ability in Reading Comprehen-
Cornell Critical Thinking Test, Level Z (1985) by sion (1987) by L.M Phillips and C. Patterson. Institute
R.H. Ennis and J. Millman. Midwest Publications, PO for Educational Research and Development, Memorial
Box 448, Pacific Grove, CA 93950. Aimed at advanced University of Newfoundland, St. John's, Newfoundland,
or gifted high school students, college students, and oth- Canada A1B 3X8. Aimed at grades 6-8. Tests for ability
er adults. Sections on induction, credibility, prediction to infer information and interpretations from short pas-
and experimental planning, fallacies (especially equivo- sages. Multiple choice version (by both authors) and con-
cation), deduction, definition, and assumption identifica- structed response version (by Phillips only).
tion. Watson-Glaser Critical ThinkingAppraisal (1980)
The Ennis-Weir Critical ThinkingEssay Test (1985) by G. Watson and E.M. Glaser. The Psychological Cor-
by R.H. Ennis and E. Weir. Midwest Publications, PO poration, 555 Academic Court, San Antonio TX 78204.
Box 448, Pacific Grove CA 93950. Aimed at grades 7 Aimed at grade 9 through adulthood. Sections on induc-
through college. Also intended to be used as a teaching tion, assumption identification, deduction,judging wheth-
material. Incorporates getting the point, seeing the rea- er a conclusion follows beyond a reasonable doubt, and
sons and assumptions, stating one's point, offering good argument evaluation.
reasons, seeing other possibilities (including other possi-
ble explanations), and responding to and avoiding equiv- Tests Covering Only One Aspect of Critical Thinking
ocation, irrelevance, circularity, reversal of an if-then Cornell Class Reasoning Test (1964) by R.H. Ennis,
(or other conditional) relationship, overgeneralization, W.L. Gardiner,R. Morrow, D. Paulus, and L. Ringel. Illi-
credibility problems, and the use of emotive language to nois CriticalThinking Project, University of Illinois, 1310
persuade. S. 6th St., Champaign,IL 61820. Aimed at grades 4-14.
Judgment: Deductive Logic and Assumption Rec- Tests for a variety of forms of (deductive)class reasoning.
ognition (1971) by E. Shaffer and J. Steiger. Instruction- Cornell Conditional Reasoning Test (1964) by R.H.
al Objectives Exchange, PO Box 24095, Los Angeles, Ennis, W. Gardiner,J. Guzzetta, R. Morrow, D. Paulus,
CA 90024. Aimed at grades 7-12. Developed as a crite- and L. Ringel. Illinois Critical Thinking Project, Univer-
rion-referenced test, but without specific standards. In- sity of Illinois, 1310 S. 6th St., Champaign, IL 61820.
cludes sections on deduction, assumption identification, Aimed at grades 4-14. Tests for a variety of forms of
and credibility, and distinguishes between emotionally (deductive) conditional reasoning.
loaded content and other content. Logical Reasoning (1955) by A. Hertzka and J.P.
New Jersey Test of Reasoning Skills (1983) by V. Guilford. SheridanPsychological Services, PO Box 6101,
Shipman. Institute for the Advancement of Philosophy Orange, CA 92667. Aimed at high school and college
for Children, Test Division, Montclair State College, students and other adults. Tests for facility with class
Upper Montclair, NJ 08043. Aimed at grades 4 though reasoning.
college. Incorporates the syllogism (heavily represent- Test on Appraising Observations (1983) by S.P.
ed), assumption identification, induction, good reasons, Norris and R. King. Institute for Educational Research
and kind and degree. and Development, Memorial University of Newfound-
Ross Test of Higher Cognitive Processes (1976) by land, St. John's, Newfoundland, Canada, A1B 3X8.
J.D. Ross and C.M. Ross. Academic Therapy Publica- Aimed at grades 7-14. Tests for ability to judge the cred-
tions, 20 Commercial Blvd., Novato, CA 94947. Aimed ibility of statements of observation. Multiple choice and
at grades 4-6. Sections on verbal analogies, deduction, constructed response versions.

183

This content downloaded from 50.155.233.239 on Mon, 18 May 2015 22:26:22 UTC
All use subject to JSTOR Terms and Conditions
THEORYINTO PRACTICE/ Summer 1993
Teachingfor Higher Order Thinking

criticalthinkingessay test, which does test more com- less structured-in the form of naturalisticobserva-
prehensively than others. But it is not secure. Fur- tion of a student. Greater structureusually means
thermore,it is more expensive in time and/or money greatereffort beforehand,but also greaterassurance
than multiple-choice tests to administer and score. that there will be opportunitiesto assess specific as-
The problem is serious in high-stakestesting. We do pects of critical thinking.Less structuregenerally re-
not yet have inexpensive critical thinking testing us- quires greatereffort duringand after the observation,
able for high stakes. Research and development are and gives the opportunityfor more life-like situa-
needed here. tions, but provides less assurancethat a broad range
The listed multiple-choice tests can, to varying of specific aspects of critical thinking will be as-
degrees, be used for the first five listed lower stakes sessed. The sections thatfollow illustrateseveraltypes
purposes:diagnosis, feedback, motivation, impact of of open-endedcritical thinkingtests that teacherscan
teaching, and research. But discriminatingjudgment make themselves.
is necessary. For example, if a test is to be used for
diagnostic purposes, it can legitimately only reveal Multiple Choice With Written Justification
strengthsand weaknesses in aspects for which it tests. In the Illinois CriticalThinkingProject-in con-
The less comprehensivethe test, the less comprehen- junction with the Alliance for Essential Schools in
sive the diagnosis. Illinois-we are currently exploring the use of the
For comprehensiveassessment,unless appropri- multiple-choice-plus-written-justification format.We
ate multiple-choice tests are developed, open-ended have taken 20 items from the Cornell Critical Think-
assessment techniquesare probablyneeded. Until the ing Test, Level X, and requesteda brief writtenjusti-
published repertoireof open-ended critical thinking fication of the student's answer to each. In the fol-
tests increases considerably,and unless one uses the lowing example of an item focusing on judging the
publishedessay test, or partsof otheropen-endedtests, credibilityof a source, the situationis the exploration
such as College Board's advanced placement (AP) of a newly-discoveredplanet:
tests, it is necessary to make your own test. WHICHIS MOREBELIEVABLE? Circleone:
A. The healthofficer investigatesfurtherand says,
Making Your Own Test "Thiswatersupplyis safe to drink."
In making your own test, it is probably better B. Several others are soldiers. One of them says,
that it be at least somewhatopen ended anyway, since "Thiswatersupplyis not safe."
makinggood multiple-choicetests is difficult and time C. A and B are equallybelievable.
consuming, and requiresa series of revisions, tryouts, YOURREASON:
and more revisions. Suggestions for makingmultiple- One advantageof this promising format is that spe-
choice critical thinkingitems may be found in Norris cific aspects of critical thinking can be covered (in-
and Ennis (1989), but I do not present them here, cluding an aspect not effectively tested in existing
because open-ended assessment is better adapted to multiple-choicetests: being appropriatelycautious in
do-it-yourself test makers and can be more compre- the drawingof conclusions).Anotheradvantageis that
hensive. Norris and Ennis also make suggestions for answersthatdifferfromthosein the key, if well defend-
open-endedassessment,and the discussionhere grows ed, can receive full credit.Answersthatdifferfrom the
out of that presentation. key, as I notedearlier,are sometimesdefensible,given
Multiple-choiceassessment is labor intensive in thatthe test takerhas differentbeliefs aboutthe world
the constructionand revision of the tests. Open-end- than the test maker. We have found that high inter-
ed assessment is labor intensive in the grading, once rater consistency (.98) can be obtained if guides to
one has developed a knack for framing questions. scoringare carefullyconstructedand if the scorersare
One promising approachis to give a multiple-choice proficientin the same conceptionof criticalthinking.
item, thus assuring attentionto a particularaspect of I recommendthis approachto makingyour own
critical thinking, and to ask for a brief written de- test. It is fairly quick, can be comprehensive, pro-
fense of the selected answer to the item. vides forgivenessfor unrefinedmultiple-choiceitems,
As in the previous example, open-endedassess- and allows for differencesin studentbackgroundsand
ment can be fairly well structured.Or it can be much interpretationof items.

184

This content downloaded from 50.155.233.239 on Mon, 18 May 2015 22:26:22 UTC
All use subject to JSTOR Terms and Conditions
Ennis
Critical ThinkingAssessment

Essay Testing of Critical Thinking We developeda six-factoranalyticscoring system, an


Several approachesto making one's own essay adaptationof scoring guides developedby the Illinois
test of critical thinking are viable, depending on the StateBoardof Education,and have securedhigh inter-
purpose. raterconsistency(.94). This approachalso looks prom-
High structure. The use of the argumentative ising. Gradingtakes us about 5 minutesper essay for
essay to assess critical thinking can vary consider- essays writtenin 40 minutesof class time.
ably in degree of structure.The Ennis-WeirCritical
ThinkingEssay Test is an example of a highly struc- Performance Assessment
turedessay test. It provides an argumentativepassage Performanceassessment is the most expensive
(a letter to an editor) with numberedparagraphs,most of all, since it requires considerableexpert time de-
of which have specific built-in errors. Students are voted to each student.It has the greatestface validity
asked to appraisethe thinking in each paragraphand for whateveris revealed,since the situationsare more
the passageas a whole, and to defendtheirappraisals. realistic-possibly real life situations. However, the
A scoring guide assigns a certainnumberof pos- greaterthe realism, the less assuranceof comprehen-
sible points to the appraisal of each paragraphand siveness. In real life situations, people generally re-
the passage as a whole, and provides guidance for the veal only what the situation requires, and most ob-
grader. But the grader must be proficient in critical servable situationsdo not requireall aspects of criti-
thinkingin orderto handle responses that differ from cal thinking.So real-life performanceassessment en-
standardresponses in varyingdegrees. Responses that countersa difficulty similarto one found in multiple-
are radically different, if well defended, receive full choice assessment: reduced comprehensiveness.An-
credit. Scoring by proficient graders takes about 6 other possible danger in performanceassessment is
minutes per essay. excessive subjectivity.
Medium structure. Structurecan be reduced by The least structuredperformanceassessment is
providing an argumentativepassage and requesting naturalistic observation, as in a case study (Stake,
an argumentativeresponse to the thesis of the pas- 1978). Here, a trainedobservertakes extensive notes
sage and its defense-without specifying the organi- describing a series of events and focuses on the ac-
zation of the response. College Board AP tests use tivities of one person or group.Interpretationis inev-
this approach. itable, but "rich"descriptionis the goal.
Scoring can be either holistic (one overall grade An example of slightly more structuredperfor-
for the essay), or analytic (a grade for each of several mance assessment is the use of a student's portfolio
criteria).Holistic scoring is quicker and thus less ex- of work to determine graduationfrom high school
pensive. Proficient graderstake roughly 1 or 2 min- (recommended, for example, by Sizer in Horace's
utes for a two-pageessay. Analytic scoringgives more Compromise,1984). Validity of this type of assess-
information and is more useful for most purposes. ment is not yet established. It is an attractiveidea,
Proficientgraderstakeroughly3 to 6 minutesfor a two- but many problems exist, including probablelack of
page essay, dependingon how elaboratethe criteriaare. comprehensivenessof critical thinking assessment.
Minimal structure. Structurecan be furtherre- A more structuredperformance assessment is
duced by providingonly a question to be answeredor exemplified by an exploratoryassessment effort by
an issue to be addressed.The Illinois Critical Think- the NationalAssessmentof EducationalProgress(Blum-
ing Essay Contest uses this approach(Powers, 1989). berg, Epstein,MacDonald,& Mullis, 1986). A student
In one year, studentswere asked to take and defend a is given a varietyof materialsand asked to investigate
position about the possible regulation of music tele- what factors affect the rate at which sugar cubes dis-
vision, a topic of great interest to students. Reduced solve. The observerasks questionsand watchesto see
structuregives students more freedom, but provides whetherthe studentgoes aboutthe taskscientifically.In
teachers with less assurance of diagnostic informa- this kind of performanceassessment, structureis pro-
tion, not a problem for the essay contest. Again, ei- vided by the assignmentof a task, which is designed
ther holistic or analytic scoring is possible. to check things of interest.
At Illinois we are also using this formatfor the Performanceassessmentseems valid on the face
developmentof the IllinoisCriticalThinkingEssay Test. of it. Expense, possible lack of comprehensiveness,

185

This content downloaded from 50.155.233.239 on Mon, 18 May 2015 22:26:22 UTC
All use subject to JSTOR Terms and Conditions
THEORYINTO PRACTICE/ Summer 1993
Teachingfor Higher Order Thinking

possible excessive subjectivity, and lengthy reports Dewey, J. (1910). How we think. Boston: D.C. Heath.
are dangers. Educational Policies Commission. (1961). The central
purpose of American education. Washington, DC:
National Education Association.
Summary Ennis, R.H. (1962). A concept of critical thinking. Har-
Critical thinkingtesting is possible for a variety vard Educational Review, 29, 128-136.
of purposes.The higher the stakes and the greaterthe Ennis, R.H. (1981). Eight fallacies in Bloom's taxono-
budgetaryrestraints,the fewer the purposes that can my. In C.J.B. Macmillan (Ed.), Philosophy of edu-
be served. In particular,comprehensivenessof cover- cation 1980 (pp. 269-273). Bloomington, IL: Phi-
losophy of Education Society.
age of aspects of critical thinking is threatened in Ennis, R.H. (1987). A taxonomy of critical thinking dis-
high-stakes testing. positions and abilities. In J. Baron & R. Sternberg
A number of published tests focus on critical (Eds.), Teaching thinking skills: Theory and prac-
thinking. Almost all are multiple-choicetests, an ad- tice (pp. 9-26). New York: W.H. Freeman.
vantage for efficiency and cost, but currentlynot for Ennis, R.H. (1991). Critical thinking: A streamlined con-
comprehensiveness.More research and development ception. Teaching Philosophy, 14(1), 5-25.
Ennis, R.H. (in press). Critical thinking. Englewood
are needed.
Cliffs, NJ: Prentice-Hall.
Viable alternativesinclude the additionof justi- McPeck, J.E. (1981). Critical thinking and education.
fication requests to multiple-choice items, essay test- New York: St. Martin's Press.
ing with varyingdegreesof structure,andperformance The nation's report card. (1987). Cambridge, MA: Na-
assessment.All are considerablymore expensive than tional Academy of Education, Harvard Graduate
School of Education.
multiple-choice testing when used on a large scale, Norris, S.P., & Ennis, R.H. (1989). Evaluating critical
but on a small scale, they offer a feasible alternative
thinking. Pacific Grove, CA: Midwest Publications
in terms of validity and expense. However, grading Paul, R.W. (1987). Dialogical thinking: Critical thought
them does take more time than grading a prepack- essential to the acquisition of rational knowledge
aged multiple-choice test. and passions. In J. Baron & R. Sternberg (Eds.),
Teaching thinking skills: Theory and practice (pp.
Note: The author deeply appreciates the helpful com- 127-148). New York: W.H. Freeman.
ments of Michelle Commeyras, Marguerite Finken, Ste- Powers, B. (Ed.). (1989). Illinois critical thinking annu-
phen Norris, and Amanda Shepherd. al. Champaign, IL: University of Illinois College of
Education.
References Shepard, L.A. (1991). Will national tests improve stu-
Arter, J.A., & Salmon, J.R. (1987). Assessing higher or- dent learning? Phi Delta Kappan, 73, 232-238.
der thinking skills: A consumer's guide. Portland, Sizer, T. (1984). Horace's compromise. Boston: Hough-
OR: Northwest Regional Educational Laboratory. ton-Mifflin.
Blumberg, F., Epstein, M., MacDonald, W., & Mullis, I. Smith, M.L. (1991). Put to the test: The effects of exter-
(1986). A pilot study of higher-order thinking skills nal testing on teachers. Educational Researcher,
assessment techniques in science and mathematics. 20(5), 8-11.
Princeton, NJ: National Assessment of Educational Stake, R.E. (1978). The case study method in social in-
Progress. quiry. Educational Researcher, 7(2), 5-8.

Tip

186

This content downloaded from 50.155.233.239 on Mon, 18 May 2015 22:26:22 UTC
All use subject to JSTOR Terms and Conditions

Potrebbero piacerti anche