Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Module_A Page 1
Now, we have got two Gaussian models, each of them representing the class
conditional likelihood.
Module_A Page 2
In our two class classification problem, if we are given an sample the probability
of the class label can be formulated as
Module_A Page 3
=
Note that, using two multivariate Gaussian distributions with shared diagonal
covariance matrix we get an equivalent model to logistic regression in binary
classification problems. The converse is not true, being logistic
regression does not imply is multivariate Gaussian. Using Gaussian
distribution to model implies a strong assumption of that the class
conditional data is indeed Gaussian. If this assumption is correct, then GDA is
better than logistic regression. On the contrary, logistic regression does not
require our data to be in any parametric form. It uses less assumptions, and hence
it is more robust, i.e. it works better with a wide-range of types of data. It turns
out, if the class-conditional data is modelled by a Poisson distribution we get a
logistic regression model as well.
Module_A Page 4
Then, using class priors, say, and the
corresponding likelihood functions can be plotted:
Pick if otherwise:
Module_A Page 5
The value of the likelihood ratio can range between zero and infinity. Decision can
be made as well on the likelihood ratio as:
Pick if otherwise:
Since the denominator would be the same in all discriminant functions it can be
omitted:
Module_A Page 6
Or for simpler and faster calculation its logarithm can be used
For multi-class classification using SVMs, for each class one can define a
discriminative functions such that the class is marked as a
positive class, while all the other classes are merged and marked as negative
class. Then the classification can be carried out as above. In fact, this approach
can be applied on all binary classification methods, such as logistic regression.
Module_A Page 7