Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
ABSTRACT
While the most common way of evaluating a computational
model is by showing a good fit with the empirical data,
recently the literature on theory testing and model selection
criticizes the assumption that this is actually strong
evidence for a model. This paper explores the possibilities
of developing a method selection technique that can serve
as an alternative to a goodness-of-fit (GOF) measure. This
alternative, a measure of surprise, is based on the common
idea that a model gets more support from the correct
prediction of an unlikely event than the correct prediction
of something that was expected anyway.
1.
2.
3.
4.
5.
Keywords
Cognitive science, computational
selection, rhythm perception.
modeling,
6.
model
INTRODUCTION
38
ICMPC9 Proceedings
computational models has become a more important issue
than before.
APPROACH
While the most common way of evaluating a computational
model is to see whether it shows a good fit with the
empirical data, recently the literature on theory testing and
model selection criticizes the assumption that this is
actually strong evidence for the validity of a model. Some
authors consider a fit between a theory and the empirical
observations a necessary starting, but clearly not the end
point of model selection or verification (e.g., Jacobs &
Grainger, 1994; Desain, Honing, Thienen, & Windsor,
1998; Rodgers & Rowe, 2002). Others suggest alternatives
to a goodness-of-fit (GOF) measure, such as preferring the
simplest model, in terms of both its functional form and the
number of free parameters (e.g., Pitt & Myung, 2002; Pitt,
Myung, & Zhang, 2002). These authors identified serious
shortcomings of GOF as a model selection method
(summarized in Figure 1). Yet others have indicated a
preference for theories that predict an empirical
phenomenon that was least expected, as they consider a
good fit to be of less relevance or even misleading (e.g.,
Roberts & Pashler, 2000).
39
ICMPC9 Proceedings
tempo values (e.g., a horizontal line would indicate a
constant tempo, a vertical line an instant tempo change).
DISCUSSION
While for most scientists the limitations of GOF might be
clear (cf. Pitt & Myung, 2002), the recent discussion in the
cognitive science literature (summarized here in the
Approach section) shows that this selection method is still
(or again) in the center of scientific debate. However,
Honing (2006) can be seen as a proof of concept with
regard to the applicability of the element of surprise to
theory testing and model selection in music cognition
research. It demonstrates that a measure of surprise an
index of the limited range, non-smooth and surprising
predictions of a model could potentially serve as an
alternative to more common model selection techniques,
including GOF and measures of simplicity. Of course, the
challenge is to elaborate, formalize and evaluate such a
measure of surprise. This is the topic of current research.
ACKNOWLEDGMENTS
Thanks to Leigh M. Smith and Olivia Ladinig for their
comments on an earlier version of this paper.
REFERENCES
Desain, P., & Honing, H. (2003). The formation of
rhythmic categories and metric priming. Perception, 32,
341-365.
Desain, P., & Honing, H. (2004). Final Report NWOPIONIER Project Music, Mind, Machine. ILLC, X-200402. [ http://dare.uva.nl/en/record/117783 ]
40
ICMPC9 Proceedings
Desain, P., Honing, H., Thienen, H. van, & Windsor, W. L.
(1998) Computational Modeling of Music Cognition:
Problem or Solution? Music Perception, 16(1), 151-166.
41
ICMPC9 Proceedings
42