Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
8/02/2011
Theory
We have theories that explain the way things are (not just
descriptions)
Th
Theory
and
d data
d t are inter-twined
i t t i d
Models
BUT
8/02/2011
Models are rejected if they do NOT have close fit to the data
Models are NOT accepted if they have close fit to the data
8/02/2011
WHY?
CHANCE.ERROR.DEFICIENCIES IN STIMULI
Observed behaviour is not perfectly controlled or reflective of our
TRUE intelligence, attitude, etc.
8/02/2011
linear regression
Changes in XXX cause a
linear change (increase or
decrease) in YYY
Formula:Y= m*X + b
m=slope [standardised
beta = a proportion of
standard deviation]
b=intercept [starting point
of equation; represents all
the unknown stuff]
Y variablee
b
intercept
X
variable
i bl
Interpretations:
1. For every 1 SD change in X, you
will get m*SD change in Y.
2. This relationship explains x% of
variance in Y
Variables
8/02/2011
Paths
Grades
Ticks
Well-being
Evaluative
Praise
Stickers
Answers
1
e12
1
e13
1
e14
1
e15
1
e16
8/02/2011
Estimation
Maximum likelihood
The parameter values in the data set (a sample) are the most likely
values in the population (not present, but to which we wish to
generalise)
l )
Hence, procedure attempts to maximise the input values (means,
standard deviations, covariances) when estimating the solution
Hence, it matters that the sample reflects the population and is
sufficiently large that parameters are likely to apply to population
The chi-squared
chi squared (2) test is the fundament of model evaluation
2 test: difference between Observed (model) and Expected
(Data) adjusted by number of parameters and cases (degrees
of freedom)
However, 2 penalises falsely large N (i.e., >100) and large
number of manifest variables
So it is a poor test
test, notwithstanding vehement objections by
some researchers
8/02/2011
p of
2/df
Badness of fit
CFI
gamma hat
h t
RMSEA
SRMR*
Good
>.05
>.95
<.05
<.06
Acceptable
>.05
>.90
<.08
<.08
Marginal
>.01
.85-.89
<.10
Reject
<.01
<.85
>.10
>.08
Note.
Report multiple indices but beware..
beware
CFI punishes falsely complex models (i.e., >3 factors)
RMSEA rewards falsely complex models with mis-specification
See Fan & Sivo, 2007
*AMOS only generates SRMR if NO missing data;
thus, important to clean up missing values prior to any analysis. Recommend
expectation maximization (EM) procedure
RMSEA
Model
Default model
Independence
RMSEA LO 90
.048
.045
.127
.124
HI 90
.051
.129
PCLOSE
.899
.000
8/02/2011
Structural model:
multiple predictors
of performance
8/02/2011
Have
a beginning
H
b i i and
d an end
d which
hi h are nott the
th same
NOT circular
origins
endings
Make it longitudinal
Time 2
A2 B2C2
Time 1
A1B1C1
AM1 BCAM2
10
8/02/2011
Interpreting a Model
Evaluating Results
If p>.05 path
not stat sig
Note. Fixed
paths have no
probability.
11
8/02/2011
Evaluating Results
Equivalent to R2
effect
ff t size
i f2 = R2 /(1 - R2)
f2 =.19/.81=.23 (medium)
.19
e42
MQQ44
14
e41.14
MQQ63
.23
e40
.17
.48
MQQ5
.41
MQQ25
e38.29
MQQ8
.34
e48
.37
e39
e37
.44
.08
Evaluation
-.54
MQQ23
-.28
.58
58
Developing a Model
12
8/02/2011
My research questions:
Th
Theoretical
i l fframework:
k
Outcomes
Criterion of effectiveness
EFA to CFA
Statement
29.Assessmentfostersstudents'character.
22.Assessmentcultivatesstudents'positiveattitudestowardslife.
20.Assessmentisusedtoprovokestudentstobeinterestedinlearning.
14.Assessmenthelpsstudentssucceedinauthentic/realworldexperiences.
13.Assessmentensuresstudentspayattentionduringclass.
34.Assessmentmeasuresstudents'higherorderthinkingskills.
g
g
27.Assessmentallowsdifferentstudentstogetdifferentinstruction.
24.Assessmentstimulatesstudentstothink.
49.Assessmentforcesteacherstoteachinawayagainsttheirbeliefs.
31.Assessmentinterfereswithteaching.
10.Assessmenthaslittleimpactonteaching.
26.Assessmentisanimpreciseprocess.
23.Assessmentresultsarefiled&ignored.
45.Teachersconductassessmentsbutmakelittleuseoftheresults.
13
8/02/2011
1
MQQ69
1
MQQ70
MQQ57
Student
Involvement
e43
1
MQQ40
1
MQQ68
1
MQQ59
MQQ50
1
MQQ58
e42
MQQ44
e41
e48
48
MQQ63
e40
MQQ45
Well-being
Evaluation
1
MQQ35
1
MQQ24
MQQ8
e47
MQQ41
MQQ10
e34
MQQ20
e33
MQQ29
MQQ6
MQQ15
e45
1
e30
e29
MQQ42
e28
MQQ21
e27
MQQ32
Growth
e46
46
MQQ1
MQQ11
1
MQQ4
1
MQQ22
1
MQQ12
1
MQQ16
MQQ43
e7
e8
e9
e10
e11
e12
e13
e16
e17
e18
e19
e20
e21
e22
1
MQQ55
Irrelevance
e23
1
MQQ2
e24
1
e6
e15
1
TCoF e49
Timeliness
MQQ61
e5
e14
e31
e4
MQQ31
1
1
e3
MQQ23
e35
e2
MQQ18
e36
e32
32
MQQ25
e38
e37
MQQ49
MQQ66
MQQ5
e39
e44
1
e1
MQQ53
e25
1
MQQ17
e42
e26
MQQ44
e41
MQQ63
e48
1
e40
MQQ5
Evaluation
e39
MQQ25
e38
38
e37
MQQ8
1
MQQ23
Check that it is
IDENTIFIED.
OOOPS!
Seed value omitted
14
8/02/2011
15
8/02/2011
e42
e34
e33
MQQ20
MQQ17
.49
MQQ16
.81
.65
.80
Strategy
Development
.49
.60
e32
e41
MQQ19
e30
MQQ14
e29
MQQ38
e28
MQQ42
e27
MQQ39
e43
e4
e43
e5
e42
MQQ44
MQQ32
e6
.13
e41
MQQ63
MQQ2
e7
e48
Organised
Planned
e9
.45 MQQ4
e10
.51
MQQ60
Conceptions
C
i
of Feedback
.70
MQQ58
e23
MQQ25
e22
MQQ24
e44
.71
.70
.68
e47
e46
.70
.75
MQQ55
.63
.71
.77
Encourage
Improvement
MQQ53
e40
mqq56
MQQ21
e13
MQQ34
e14
MQQ28
e16
MQQ8
.43
.32
.69
Independence
.57
.81
e59
e17
MQQ43
e18
MQQ64
e35
MQQ50
e36
MQQ48
e37
.22
e47
e46
e45
e44
e43
MQQ38
MQQ37
e35
MQQ10
MQQ30
MQQ44
MQQ63
e38
e41
MQQ70
e49
e40
MQQ5
e39
MQQ25
e37
MQQ23
e36
MQQ41
e35
MQQ10
e34
MQQ20
e33
MQQ29
e31
e30
MQQ11
e28
MQQ21
e27
MQQ32
e51
.80
.52
.30
.69
.43
.25
.20
.15
-.22
.34
.52
.35
.50 Accuracy of
Grading
.41
.69
.05
.22
-.12
.65
.67
.83
.42
.12
Timeliness
.65
.55
Student
Well-being
-.14
.62
.54
.59
.34
.39
.15 -.02
.01 .10
-.46
.63
.57
.64
.72
.41
.55
.22
Tchr-Only
Valid
.58
.58
.28
.42
.34
.57
.54
-.02
.03
MQQ8
MQQ26
MQQ64
MQQ67
.67
.57
.64
.60
.53
Ineffective Grading
-.07
.70
.51
.66
.53
-.39.80
.43
Student
Learning
Growth
.63
.55
.55
.46
Interactive
MQQ26
.53
.46
.57
.59
.49
.56
MQQ64
.29
e1
e52
MQQ67
MQQ70
e2
e51
MQQ65
MQQ57
e3
MQQ40
e4
-.39
.63
.59
MQQ58
.86
.69
MQQ49
e60
.30
e7
.16
e5
e6
e6
e5
.57
.52
.40
MQQ50
MQQ59
MQQ68
e7
e4
MQQ40
e8
e3.43
MQQ57
.58
MQQ49
e9
e2
MQQ70
MQQ45
e10
e1.27
MQQ69
MQQ66
e11
MQQ35
e12
MQQ24
e13
MQQ31
e15
e16
e20
MQQ16
e21
MQQ43
e22
MQQ55
e23
MQQ2
e24
MQQ53
e25
MQQ17
e26
MQQ65
.30
.97
MQQ24
.50
Learning
.21
.66
.89
.94
.55
.45
Growth
.45
e65
.55
.40
.72
Student
Involvement
.64
.66
.39
.72
.68
.76
e58
MQQ22
MQQ12
MQQ16
.57
MQQ2
MQQ53
e15
.29
e16
.25
e19
e20
.30
e21
.21
e22
.33
e23
e24
.24
e25
.33
MQQ17
MQQ37
MQQ38
.51
e13
.18
.33
.57
.80
e12
.34
.43
MQQ55
.80
.56
.52
MQQ6
.57
Student
Response
e57
MQQ31
MQQ43
.49
.83
e11
.32
MQQ35
.42
.54
Interactive
MQQ66
.58
.25
e64
-.20
e9
e10
MQQ36
e26
e45
e46
.64
.63
.26
e47
e50
e49
e19
.32
.64
e61
e63
MQQ58
MQQ6
MQQ45
.57
Well-being
.65
.54
MQQ50
MQQ12
.45
e8
.48
.43
.66
.75
MQQ59
MQQ22
.35
Evaluation
.41
.66
e31
MQQ61
e62
Timeliness
.53
.20
MQQ68
MQQ3
.53
.22
.20 .11
-.46
.42
e54
MQQ69
MQQ54
.70
.21-.19-.31
-.10 .54
MQQ29
.32
Expected
MQQ61
e54
e53
e52
-.12
MQQ20
e33
e53
.66
Student
Led
Process
Oriented
e34
e30
.40
.46
Grading
.62
.59
.44
.51
.61
MQQ9
e42
e55
.52
.79
MQQ36
.39 MQQ47
Louisiana: 7 Hierarchical
factors, marginal fit
But this is not what we
really expectedsample or
model?
MQQ41
e28
.43
MQQ11
.55
.39
e36
.28
.76
e15
.63
.66 MQQ22
Make Ss
Feel Good
e45
.66
e12
.75
MQQ23.64
e19
.66
.61
.34
NZ: 10 inter-correlated
factors, acceptable fit
But its a mess
e11
MQQ1
MQQ40
.42
e27
.25
MQQ32
MQQ21
.65
Irrelevance
.48
.93
.42
MQQ23
.50
e8
MQQ7
.43
.42
MQQ25
.37
.25
.44
.50
.17
MQQ5
e39
-.23
.96
.73
.36
.20
e56
Tchr-Only
Valid
.52
e37
MQQ3
.40
.61
MQQ41
.27
.18
.63
.57
Irrelevance
.57
MQQ30
.18
Expected
.57
.58
MQQ9
.54
.70
MQQ3
MQQ54
.33
e44
e40
.54
.73
e24
e20
e3
MQQ30
.71
.78
e21
MQQ37
e49 .33
e50
.48
e31
e25
e2
.64 MQQ36
.72
e26
e1
MQQ45
.72
.88
Required
Expected
.49
MQQ46
Factors Items
df
1. LAhierarchical
40
1758.12
733
2. NZinter
correlated
3. NZ
intercorrelated,
bifactor
hierarchical
10
48
2444.97
1035
10
46
2378.58
1019
2/df,
p
2.40,
.12
2.36,
.12
2.33,
.13
CFI
.78
.79
.90
.051
.061
.79
.90
.051
.063
16
8/02/2011
No EFA needed
Just run the model, does it fit?
If NOT, then EFA must begin again
17
8/02/2011
Correlation analysis
Regression analysis
18
8/02/2011
Hypothesis
19
8/02/2011
CFA/SE
M: Belief
to Belief
((Brown, 2009))
Theory
Inventory development
20
8/02/2011
Summary
21
8/02/2011
Summary
Brown, G. T. L., Harris, L. R., & Harnett, J. (2010, July). Teachers conceptions of feedback:
Results from a national sample of New Zealand teachers. Paper presented at the
International Test Commission biannual conference, Hong Kong.
Brown, G. T. L., Harris, L. R., OQuinn, C., & Lane, K. E. (2011, April). New Zealand and
Louisiana practicing teachers
teachers conceptions of feedback: Impact of Assessment of Learning
versus Assessment for Learning policies? Paper accepted for presentation to the
Classroom Assessment SIG at the annual meeting of the American Educational
Research Association, New Orleans, LA.
Brown, G. T. L., & Hirschfeld, G. H. F. (2008). Students conceptions of assessment:
Links to outcomes. Assessment in Education: Principles, Policy and Practice, 15(1), 3-17.
Brown, G. T. L., Irving, S. E., Peterson, E. R., & Hirschfeld, G. H. F. (2009). Use of
interactive-informal assessment practices: New Zealand secondary students
conceptions of assessment. Learning & Instruction, 19(2), 97-111.
Brown, G. T. L., & Michaelides, M. (2011). Ecological rationality in teachers
teachers
conceptions of assessment across samples from Cyprus and New Zealand. European
Journal of Psychology of Education. doi:10.1007/s10212-010-0052-3
Brown, G. T. L., Peterson, E. R., & Irving, S. E. (2009). Self-regulatory beliefs about
assessment predict mathematics achievement. In D. M. McInerney, G. T. L. Brown, &
G. A. D. Liem (Eds.) Student perspectives on assessment: What students can tell us about
assessment for learning (pp. 159-186). Charlotte, NC: Information Age Publishing.
22
8/02/2011
References: Authorities
Ajzen, I. (2005). Attitudes, personality and behavior (2nd ed.). New York:
Open University Press.
Byrne, B. M. (2001). Structural Equation Modeling with AMOS: Basic
Concepts, Applications, and Programming. Mahwah, NJ: LEA.
Fan, X., & Sivo, S. A. (2007). Sensitivity of fit indices to model
misspecification and model types. Multivariate Behavioral Research,
42(3), 509529.
Marsh, H. W., Hau, K.-T., & Wen, Z. (2004). In search of golden rules:
Comment on hypothesis-testing approaches to setting cutoff values
for fit indexes and dangers in overgeneralizing Hu and Bentler's
(1999) fifindings.
di
S
Structural
lE
Equation
i Modeling,
M d li 11(3),
11(3) 320-341.
320 341
Marsh, H. W., Hau, K.-T., Balla, J. R., & Grayson, D. (1998). Is more
ever too much? The number of indicators per factor in confirmatory
factor analysis. Multivariate Behavioral Research, 33(2), 181-220.
23