Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Bayesian Learning
Dr. Hasan Sajid
Bayes Theorem
• In machine learning we use a model to describe
the process that results in the data that are
observed.
• Formally, we are often interested in determining
the best hypothesis from some space H, given
the observed training data D.
• One way to specify what we mean by the best
hypothesis is to say that we demand the most
probable hypothesis, given the data D plus any
initial knowledge about the prior probabilities of
the various hypotheses in H.
Bayes Theorem
• Bayes theorem provides a way to calculate the
probability of a hypothesis (P(h|D)) based on its
prior probability (P(h)), the probabilities of
observing various data given the hypothesis (P(D|
h)), and the observed data (P(D)) itself.
Bayes Theorem
• Important Observations:
– P(h|D) increases with P(h) and with P(D|h)
– P(h|D) decreases as P(D) increases, because the more
probable it is that D will be observed independent of h, the
less evidence D provides in support of h.
Likelihood Estimate or Probability of
Observing data given a hypothesis
Prior/Our Belief
Posterior Probability
P(h|D)=P(D|h)P(h)
P(cancer|positive) = P(positive|cancer)P(cancer)
= (0.98)(0.008) = 0.0078
Normalized Probabilities
P(cancer|positive) = 0.0078/(0.0078+0.0298) = 0.21
P(no cancer|positive) = 0.0078/(0.0078+0.0298) = 0.79
Bayes Theorem Revisited
• What are the probability terms?
Likelihood Estimate or Probability of
Observing data given a hypothesis
Prior/Our Belief
Posterior Probability
Parameters
Maximum Likelihood Estimation
• Different values of the parameters m and c results
in different lines (hypothesis)