Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
***no right or wrong answer. Provide an argument and make an informed decision.
1. Is it an assessment or an evaluation?
Think about significance = evaluation
Your assessment strategy feeds into your evaluation
2. Consider the 3 steps to the decision-making process
Conditions (for collected info - physical, emotional context, letting
students show their typical behaviour, quality of instruction)
o Which (of the 3 primary methods) is used to collect data (paper
and pencil, observation techniques, asking oral qs)
Quality of instruments
Objectivity of info, including unbiased scoring
3. Is it performance (actively engaged in the task) or authentic (replicate real-world
contexts) assessment?
4. Is it a standard (under identical conditions or scoring, for fairness) or nonstandard (created by teacher and do not compare students assessment?
5. Is it norm-referenced (comes from within the group, your own criteria in the
classroom, bell-curve) or criterion referenced (designed outside of the gropu. E.g.
criteria made up by the ministry?
6. It is a pluralistic assessment: being responsive to cultural diversity in the
classroom?
7. Does it demonstrate a fixed or growth mindset?
This is why one assessment is not reliable for all
Fixed mindset ex is lack of motivation/attitude. Can represent a random
error. Norm ref can lead to fixed mindset where teachers teach to the test
rather than skills and high order thinking. Teachers are not encouraging
students to learn beyond what is needed. This also effects teachers e.g. low
passion towards job and pedagogy if teaching a low achieveing school
Praising students for their intelligence puts students into a fixed mindset
Growth = encourages teamwork and collaboration rather than an
individualist mindset
8. What is the washback effect?
The impact on on-intended and unintended impacts of assessment
strategies and tools. What does this mean for the organization? Its the test
on teachers and students actions. Washback can be positive (expected) or
negative (unexpected, harmful)
4 aspects
o Practicality
pilot testing
training
double-checking data
multiple measures of the same construct (using different tools, e.g. not all
oral qs)
reliability is dependent on the sample (e.g. for test-retest its dependent upon how
the student performs)
test and sample dependent
problem is it doesnt differentiate between test takers ability
problem is its based on validity and reliability
classical analyses are done on the whole rather than individual
although item statistics can be generated, they apply to the whole group.
In CTT we assume that error is: normally distributed, uncorrelated with true
score, has a mean of zero
o
o
o
o
Evaluation:
what
Factor Analysis
- statistical model that tries to say what does income/diversity/socioeconomic (any
general term) mean?
- These qs are not as simple as they seem
- Looks at concept to see what factors come together to make that global concept so
then you can make 30 qs and see which ones are close together
- Regression vs factor analysis
IRT
-
Paradigm = use this word to refer to your philosophy, how you do things, the tools you
choose. Also try not to evaluate certain philosophies. You can evaluate them, but you
cant really say that one is better than the other.
Pragmatist = person who says depends
Test: Technique or tool that helps obtain (obtain indicates that it is a test) info about
characteristics of a person. A test is not an assessment!