Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
• AI vs. NN
– ‘Perceptrons’ and the XOR problem
• 1980s
– Back-propagation, Kohonen nets
– Applications
– Availability of computing power
??
“…is not applying some set formula, but balancing her experience
and knowledge of sale prices of similar houses…her knowledge about
housing prices is not static...fine tuning her calculation to fit the latest data”
Loan Prospector – HNC/Fair Isaac
Model defined by
interconnection weights
NN can have multiple output neurons
Neural Network Training
• Training - process of setting the best weights on the
edges connecting all the units in the network
1
y = −Net
1+e
Hard limiter Linear threshold Squashing function
(Sigmoid)
Common transfer functions
Single ‘layer’ NN
W: weight matrix
w11 w11 w12 …w1n
x1 y1
w21 w22 …w2n
……
x2 y2 wm1 wm2 …wmn
x: input vector
Net = x.W (dot product)
xm wmn wk
yn
Net = |wk||x| (cos A)
Inputs Outputs x
m Weights n
mxn
Assume no transfer function
w1
Let only one neuron with
w2
highest output ‘fire’
yk = w1kx1 + w2kx2 +…+wmkxm
w5 w3 = x.wk
= x ( W)
yp
xm ie. equivalent to single layer
Outputs
Weights Weights p
Inputs W1
m W2 Advantage arises from transfer
function (non-linearity)
Perceptrons and XOR
x1 Rosenblatt (1962): Perceptron can learn
w1 t
w2 anything that it can represent
w3
Representation – ability of a NN to simulate
wm
xm Net = ∑ wi xi a particular function
i
If Net >= threshold, output 1, else output 0
x2 (0,1) (1,1)
Can we linearly separate the red from yellow?
Output is 1
only is this
region
Training a NN
• Adjust weights such that the application of inputs produce desired
outputs (as close as possible)
• Input data is continuously applied, actual outputs calculated, and
weights are adjusted
• Weights should converge to some value after many rounds of training
• Supervised training
– Adjust weights such that differences between desired and actual outputs
are minimized
– Desired output: dependent variable in training data
– Each training example specifies
{independent variables, dependent variable}
• Unsupervised training
– No dependent variable specified in training data
– Train the NN such that similar input data should generate same output