Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
html
2. Agentt Types?
14. Propert es of Complete - Yes f step cost > 23. Greedy Search expand the node wh ch s closest
Un form-Cost ? eps lon (otherw se t can stuck n ? to goal n terms of heur st cs
nf n te loops) - t only takes the heur st cs not
T me - b[C*/e] # of nodes w th path the path cost
cost =< cost of opt mal solut on
Space - b[C*/e] # of nodes w th 24. Propert es of Complete - No(can stuck n loop)
path cost =< cost of opt mal Greedy search T me - O(bm) good heur st cs
solut on mprove t dramat cally
Opt mal - Mostly Yes for any step Space - O(bm) keep all nodes n
cost >= memory
Opt mal - NO
15. Depth-f rst L fo - last n , f rst out
Search ? Expand the deepest unexpanded 25. A* Search ? Evaluat on funct on f(n) = g(n) +
node
h(n)( g= cost to node, h=est mated
cost from node to goal, f=
16. Propert es of Complete - No ( h gh probab l ty t est mated total cost through node
Depth-f rst falls nto loops) to goal
search T me - O(bm) m=max mum depth Algor thm selects the m n mum
f m>>d (d= depth of least cost f(n) and expand t
solut on) -same as un form goal test
Space - O(bm) performed when node s selected
Opt mal - NO for expans on
17. Depth-l m ted depth-f srt search w th depth l m t l 26. Propert es of A* Complete - Yes, unless there are
search? Search nf n tely many nodes w t f(n) =<
f(G)
18. Iterat ve w th l m ted depth 0,1,2,3,...,d T me - Exponent al bd
deepen ng terate over d to f nd solut on Space - Keeps all nodes n
search IDS? -prefered when search space s memory
large and solut on depth s Opt mal - Yes
unknown (
- terat ve deepen n may seem A* expands all nodes w th f(n) <
wasteful because t generate early C*
nodes nult ple t mes but for deep A* expands some nodes w th f(n)
searches t s l ttle amount = C*
A* expands no nodes w th f(n) >
C*
19. Propert es of Complete - Yes )
IDS? T me - (d + 1)b0 + (d)b1 + (d-1)b2
+ ... +bd = O(bd) 27. Opt mal ty of A*
Space - O(bd)
Opt mal - Yes f step cost = 1 or
ncereas ng funct on of depth
20. Summary of
algor thms
29. Dom nance of If h2(n) >= h1(n) for all n(both 34. Grad ent Assume whe have some cost-
Heur st cs needs to be adm ss ble) Descent ? funct on : C(x1,...,xn)(C(x,y) =
then h2 dom nates h1 and s beter sqrt((x-a)2+(y-b)2)
for search. 1. Compute the grad ent
- And h(n) = max(h1(n),h2(n)) h(n)
s also adm ss ble and dom nates
h1, h2
33. H ll - Cl mb ng
?
S mulated anneal ng search ng for
a max mum. The object ve here s
to get to the h ghest po nt;
however, t s not enough to use a
s mple h ll cl mb algor thm, as
there are many local max ma. By
cool ng the temperature slowly the
global max mum s found.
45. How do we Standard approach: 49. Does False ⊨ Read as : False log c ally enta ls
make α–β • cutoff test: (where do we stop True ? True f all models that evaluate
pract cal? descend ng the tree) False to True also evaluate True
– depth l m t to True.
– better: terat ve deepen ng no model makes "false" true,
– cutoff only when no b g changes "False ⊨ True" s actually true! It's
are expected to occur next an nstance of vacuous
(qu escence search). mpl cat on: th nk of t as be ng true
• evaluat on funct on for the same reason the statement
– When the search s cut off, we https://math.stackexchange.com/q
evaluate the current state by uest ons/1721140/does-false-
est mat ng ts ut l ty. Th s est mate enta l-true-and-v ce-versa
s captured by the evaluat on
funct on. 50. Declarat ve Tell t what t needs to know
– Run α-β prun ng m n max w th
approach to Ask t what to do -> answers
these est mated values at the
buld an agent: should follow from the Knowledge
leaves nstead.
Base
Evaluat on funct ons:
• For chess, typ cally l near
we ghted sum of features Eval(s) 51. Why Do We • Problem-solv ng agents were
= w1 f1(s) + w2 f2(s) + … + wn Need Log c? very nflex ble: hard code every
fn(s) poss ble state.
• e.g., w1 = 9 w th f1(s) = (number • Search s almost always
of wh te queens) – (number of exponent al n the number of
black queens), etc. states.
• Problem solv ng agents cannot
46. Knowledge
nfer unobserved nformat on.
-a l st of facts that are known to
• We want an algor thm that
base the agent.
reasons n a way that resembles
- set of sentences n a
reason ng n humans.
formallanguage
48. Enta lment ? Means that one th ng follows from 53. Models n log c? In log c, models are formally
another
structured worlds w th respect to
KB ⊨𝛼
w ch truth can be evaluated
Knowledge base KB enta ls M s a model of sentence 𝛼 f
sentence 𝛼
sentence 𝛼 s true n M
f and only f a s true n all worlds
KB ⊨𝛼 f and only f M(KB) ⊆ M(𝛼)
where KB s true
( f there s any knowledge n KB
wh ch makes 𝛼 false then we say
KB ⊨𝛼 s false
f there s not enough knowledge
to make 𝛼 always true n KB we
say KB ⊨𝛼 s false)
Note f 𝛼 ⊨ 𝛽 then 𝛼 s a stronger
assert on than 𝛽
https://www.youtube.com/watch?
v=2M-K5OjVgYQ
54. Inference ? KB ├ α => sentence α(alfa) can 58. Propos t onal Syntax
be der ved from KB by procedure log c Propos t onal log c s the s mplest
Consequences of KB are log c - llust rates bas c deas
haystack; α s a needle
Enta lment = need n haystack;
nference = f nd ng t
79. How mutex İf they negate each other: 85. What s In AI, propos t on s a set of
between l terals Have(Cake) and ¬Have(Cake) propos t on n poss ble worlds( events) where
occur? If any two act on comes to them AI? the propos t on holds
(y eld them) are mutex
¬Have(Cake) and ¬ Eaten(Cake)
86. P(a v b) = ?
are mutex because no act on and
Eat(cake) act ons are mutex
94. How jo nt
d strub ton
90. repsentat on
change when
cond t onal
=? ndependence
ntroduced?
91. What s cha n
rule?
92. P(toothache)
g ven the jo nt
probab l ty table 95. Bayes' rule ? Product rule P(a ∧ b ) =
what are P(a|b)P(b)
P(toothache)=?
P(cav ty V P(toothache) = 0.108 + 0.016 +
toothacge) = ? 0.012 + 0.064 = 0.2
P(¬cav ty|tootha P(cav ty ∨ toothache) =
che) = ?
P(Cav ty|tootha
che) = ?
96. naïve Bayes
model?
P(cav ty ∨ toothache) = 0.108 +
0.012 + 0.016 + 0.064 + 0.072 +
0.008 = 0.28
P(¬cav ty|toothache) = ? 97. What s Markov Each node s cond t onally
Blanket n ndependent of all others g ven ts
Bayesnet: Markov blanket:
parents + ch ldren 9 ch ldren's
parents
P(Cav ty|toothache) = ?
100. What s S mple nference task of P(X|e). 109. How Inference Suppose we have query of P(B |
poster or P(X|e) s called poster or by var able j,m)
marg nal? marg nal. It s poster or el m nat on Inference by enumerat on does
d strubut on of subset of works?
var ables, n th s part cular case w th var able el mant on we turn
th s subset s only one var able X probab l t es to the factors and
103. Opt mal Dec s on networks nclude ut l ty and comput ng r ght to left starts
Dec s on ? nformat on. requ red for here are the steps
P(outcme | act on, ev dence)
108. Ma n dea of El m nat ng the repated 110. What s bel ef Ingerence by local mesage
nference by calculat ons by carry ng out propagat on? pass ng among ne ghbor ng
var able summat ons r ght-to-left (bottom- nodes
el m nat on? up) stor ng ntermed ate results The message can loosley
for later use nterpreted as "I(node ) th nk that
you (node j) are that much l kely
Does a and b
are a and b are ndependent g ven c
ndependent ev dence
g ven c
ev dence n th s
d vergent graph
Does th s graph
s a val d
bayes an
network?
129. What s
explan ng
away?
H = Happy
S = Sunny
R = Ra se (money)
When you see happy person and
you see the sunny weather t s
less l ke to person got a ra se
When you see heppy person and
you see and t s not sunny
weather and t s more l kely to
have a ra se
https://www.youtube.com/watch? 132. Sum-product -Exact nference n trees and
v=pyxyYWNo8Qw algor thm cha ns
Expla n ng away s a common -Qu te good approx mate
pattern of reason ng n wh ch the nference on other graph cal
conrmat on of one cause of an models
observed or bel eved event -Eff c ent, fast computat on
reduces the need to nvoke -Reus ng computat ons for
alternat ve causes. The oppos te d fferent nodes n the graph
of expla n ng away also can -Bel ef propagat on algor thm's
occur, n wh ch the conrmat on of general zat on
one cause ncreases bel ef n
another. We prov de a general
133. Bel ef -Sum-product
qual tat ve probab l st c analys s
Propagat on -Max-sum
of ntercausal reason ng
Algor thms? -Loopy-bel ef propagat on
130. In wh ch graphs
134. Loopy bel ef Sum-product and max-sum
the exact
propagat on vers on are there
nference v a
algor thm
message
vers ons?
pass ng s
poss ble
135. Dynam c Every kalman f lter model s DBN
Bayes an but few DBNs are KFs; real world
Networks vs requ res non-Gauss an poster ors
Kalman f lters ?
137. What are the F lter ng: P(Xt|e1:t) 143. Representat on Feed ach ne w th raw data
nference tasks bel ef state - nput to the dec s on (feature) automat cally d scover
n temporal process of a rat onal agent learn ng representat ons
probab l ty Pred ct on: P(Xt+k|e1:t) for k> 0 To do that deep learn ng:
models? evaluat on of poss ble act on representat on learn ng w th
sequences l ke f lter ng w thout mult ple layers.
the ev dence s mple but non-l near modules at
Smooth ng: P(Xk|e1:t) for 0 < k < each level
1
better est mate of past states, 144. Key deas of Layers are not des gned by
essent al for learn ng representat on(f programmers
Most l kely explanat on: argmax eature) learn ng Learn ng data from data w th
x1:tP(x1:t| e1:t) general method
speech recogn t on, decod ng
w th no sy channel 145. Is t true that Yes. It s true and t s because
s ngle layer The perceptron s s mply
138. V terb Ma n dea explanat on: perceptrons separat ng the nput nto 2
Algor thm? https://www.youtube.com/watch? can solve only categor es based on the
v=6JVqutwtzmo l nearly threshold funct on at the f nal
separable node
cases? w1I1 + w2I2 < t
139. V terb algor tm -It s var ant of max-product
propert es? product alog rthm w th back- f the LHS s < t, t doesn't f re,
track ng. otherw se t f res. That s, t s
-It max m zes over the ncom ng draw ng the l ne:
messages to each node w1I1 + w2I2 = t
-In pract ce, comput ng products And because the we ghts and
of small numbers may become thresholds can be anyth ng, th s
nstable, so often the problem s s just any l ne across the 2
reformulated as f nd ng the d mens onal nput space
max mum of the logar thm of the
jo nt probab l ty, wh ch results n 146. Can we use No because XOR problem s not
replac ng the products by s ngle layer l nearly separatable
summat ons, and y elds the “max- perceptron for
sum” algor thm. The back- XOR problem?
track ng g ves the same
sequence as w th the “max-
product” algor thm
174. How to avo d F n te hor zon: 181. Bas c dea of -Rece ve feedback n the form of
deal ng w th Term nate after T steps Re nforcement rewards
nf n te rewards D scount ng 0 < y < 1 Learn ng -Agent's ut l ty s def ned by the
smaller y means smaller hor zon reward funct on
absorb ng state: guarantee that -Must (learn to) act so as to
for every pol cy a term nal state max m ze expected rewards
w ll be reached -All learn ng s based on
observed samples of outcomes