Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
YouTube
hEp://www.youtube.com/watch? v=b1rHS3R0llU&annota)on_id=annota)on_882813&feature=iv
Business
Applica)ons
Complex
scheduling
problems
Resource
op)miza)on
in
large
factories
Op)mizing
weights
of
neural
networks
(will
learn
about
neural
networks
in
this
or
next
session)
Op)miza)on
of
Making
predic)ons
in
nance
(Well
get
to
this!)
Darwins
Insight
evolu)on
is
a
series
of
single
steps
each
single
step
is
simply
rela%ve
to
its
predecessor,
arising
randomly...
but
a
sequence
of
small
steps,
cumula%ve
selec%on
is
not
a
random
process!
explains
how
order
and
complexity
can
happen
and
thrive
Key
Observa)on
As
long
as
you
can
es)mate
the
quality
of
a
proposed
solu)on,
search
does
a
lot
of
work
for
you!
Dene
the
performance
variables
Dene
the
features
REMEMBER: OPIM 621 and OPIM 101? Example: A Linear Discontinuous Terrain of Possible Solutions X < 50 Y
B C
Y < 75
Components
of
a
GA
A
problem
to
solve,
and
...
Encoding
technique
(gene,
chromosome)
Ini)aliza)on
procedure
(crea%on)
Evalua)on
func)on
(environment)
Selec)on
of
parents
(reproduc%on)
Gene)c
operators
(muta%on,
recombina%on)
Parameter
sekngs
(prac%ce
and
art)
modica)on
modied
children
popula)on
deleted
members
evaluated children
evalua)on
discard
Popula)on
popula)on
Arrays
Trees
<
0.8
AND
Boolean
Expressions
I.e.
X1
between
0
and
0.8
Reproduc)on
parents
popula)on
Parents
are
selected
at
random
with
selec)on
chances
biased
in
rela)on
to
chromosome
evalua)ons.
reproduc)on
children
Chromosome
Modica)on
children
modica)on
modied
children
Modica)ons are stochas)cally triggered Operator types are: Muta)on Crossover (recombina)on)
A
Simple
Example
The
Traveling
Salesman
Problem:
Representa)on
Representa)on
is
an
ordered
list
of
city
numbers
known
as
an
order-based
GA.
CityList1
(3
5
7
2
1
6
4
8)
CityList2
(2
5
7
6
8
1
3
4)
80
y 60
40
20
0 0 10 20 30 40 50 x 60 70 80 90 100
44 62 69 67 78 64 62 54 42 50 40 40 38 21 35 67 60 60 40 42 50 99
80
y 60
40
20
0 0 10 20 30 40 50 x 60 70 80 90 100
80
y 60
40
20
0 0 10 20 30 40 50 x 60 70 80 90 100
100
80
y 60
40
20
0 0 10 20 30 40 50 x 60 70 80 90 100
Overview of Performance
TSP30 - Overview of Performance
1800 1600 1400 1200 e c 1000 n a t s 800 i D 600 400 200 0 1 3 5 7 9 11 13 15 17 19 21 Generations (1000) 23 25 27 29 31
Best Worst Average
Local maxima
20 Generations
50
From Seven Methods for Transforming Corporate Data Into Business Intelligence, by Vasant Dhar and Roger Stein, Prentice-Hall, 1997.
Breeds
Breeds
Breeds
Breeds
Worse part
Worse part
Worse part
Worse part
Worse part
Drops Out
Drops Out
Drops Out
Drops Out
Drops Out
stop when acceptable solu)ons are reached, or solu)ons are not improving, or aier the algorithm has run for a specied amount of )me or for a specic number of genera)ons
Krome 5
Krome 1
The image cannot be displayed. Your computer may not have enough memory to open the image, or the image may have been corrupted. Restart your computer, and then open the le again. If the red x still appears, you may have to delete the image and then insert it again.
The image cannot be displayed. Your computer may not have enough memory to open the image, or the image may have been corrupted. Restart your computer, and then open the le again. If the red x still appears, you may have to delete the image and then insert it again. The image cannot be displayed. Your computer may not have enough memory to open the image, or the image may have been corrupted.The image cannot be displayed. Your computer may not have enough memory to open the Restart your computer, and then open the le again. If the red x still appears, you may have to delete the image and then insert it again. image, or the image may have been corrupted. Restart your computer, and then open the le again. If the red x still appears, you may have to delete the image and then insert it again.
Krome 2
Krome 4 Krome 3
The image cannot be displayed. Your computer may not have enough memory to open the image, or the image may have been corrupted. Restart your computer, and then open the le again. If the red x still appears, you may have to delete the image and then insert it again.
The image cannot be displayed. Your computer may not have enough memory to open the image, or the image may have been corrupted. Restart your computer, and then open the le again. If the red x still appears, you may have to delete the image and then insert it again.
Process
in
Finance
Establish
model
basis
(I.e.
fundamentals,
anomalies?
Momentum?
Asset-specic
versus
general?)
Feature
engineering
(construc)ng
good
technical
and
fundamental
indicators)
Search
Selec)on,
crossover,
muta)on
Aier
Product
Type
0
0
1
Price
0.6
0.9
68
0.1
0.8
0.6
0.9
84
0.1
0.8
0.7
0.9
92
Crossover Point
Before
Frequency
0.3
0.9
0
Product
Type
0
0
1
Price
0.7
0.9
Frequency
0.3
0.9
0
Aier
Product
Type
0
0
1
Price
0.5
0.9
0.5
0.8
0.6
0.9
Long
0.84
PaEern: GO LONG WHEN Asset Trend is between 0.5 and 0.8 Asset Type is of type 2 or type 4 Rela)ve Strength is between 0.6 and 0.9
Based on mul)ple objec)ves, I.e. consistency of returns over )me consistency of returns across assets, etc
Return
0.02 0
14 20 26 32 38 44 50 56 62 68 74 -4 -3 -2 -2 -1 -1 80
-4
Worst Surprise
Announcement
Day
Days
The Problem: How Do You Figure Out the Outcome X Days in Advance?
The
Data
In
sample
training
14,490
records
In
sample
tes)ng
9,650
records
(split
randomly
into
mul)ple
test
sets)
Out
of
sample:
12,164
records
(con)guous
periods
chosen
manually)
The
Objec)ve
Balance
between
generality
and
accuracy
Balance
condence
and
support
50 40 30 20
10 0
64.39
63.49
65.53
Mean
Sigma
5.41 TI:conf
3.34 RL:conf
1.39 GA:conf
80 60
40 20
64.24
65.02
71.48
Mean
5.39 TI:conf
4.38 RL:conf
2.09 GA:conf
Sigma
Strengths
Flexible
Mul)ple
objec)ves
encoded
into
1
Can
model
rela)onships
between
aEributes
Searches
a
large
space
of
possible
solu)ons
Weaknesses
You have to know your objec)ve/tness Time Heuris)c doesnt nd the op)mal solu)on
A chromosome can be a whole path or even a tree. The GA can therefore generate and test in parallel
Termina)on Criteria Performance, scalability Solu)on is only as good as the evalua)on func)on (oien hardest part)
A chromosome can be a whole path or even a tree. The GA can therefore generate and test in parallel
When
to
Use
a
GA
Alternate
solu)ons
are
too
slow
or
overly
complicated
Need
an
exploratory
tool
to
examine
new
approaches
Problem
is
similar
to
one
that
has
already
been
successfully
solved
by
using
a
GA
Want
to
hybridize
with
an
exis)ng
solu)on
Benets
of
the
GA
technology
meet
key
problem
requirements
Ques)ons?
Neural
Networks
h7p://www.youtube.com/watch? v=FZ3401XVYww&feature=related
h7p://www.youtube.com/watch? v=AjxJabpjDGo
Applied Problems:
Image, Sound, and Pa7ern recogniQon Decision making Knowledge discovery Context-Dependent Analysis
- modern theory about principles and new mathemaQcal models of informaQon processing, which based on the biological prototypes and mechanisms of human brain 57 acQviQes
NEUROINFORMATICS
Massive
parallelism
Brain
computer
as
an
informa%on
or
signal
processing
system,
is
composed
of
a
large
number
of
a
simple
processing
elements,
called
neurons.
These
neurons
are
interconnected
by
numerous
direct
links,
which
are
called
connec%on,
and
cooperate
which
other
to
perform
a
parallel
distributed
processing
(PDP)
in
order
to
soJ
a
desired
computa%on
tasks.
ConnecQonism
Brain
computer
is
a
highly
interconnected
neurons
system
in
such
a
way
that
the
state
of
one
neuron
aects
the
poten%al
of
the
large
number
of
other
neurons
which
are
connected
according
to
weights
or
strength.
The
key
idea
of
such
principle
is
the
func%onal
capacity
of
biological
neural
nets
determs
mostly
not
so
of
a
single
neuron
but
of
its
connec%ons
58
59
central body in which almost all the logical func)ons of the neuron are realized.
Biological
Neurons
Synapses
to the soma which can serve as a nal output channel of the neuron. An axon is usually highly branched. branching tree of bres. These long irregularly shaped nerve bres (processes) are aEached to the soma.
Axon
from
other
neuron Soma
Axon
neuron which are the termina)on points for the axons from other neurons. Dendrites
60
A large number of neurons (informa)on processing units) connected together A neurons response depends on the states of other neurons it is connected to and to the strength of those connec)ons. The strengths are assigned based on experience.
62
Biological
Learning
n n
A neuron: many-inputs / one-output unit. Output can be excited or not excited based on signals from incoming neurons. Resistance in the synapses, changes output.
Hebbs
Rule:
If
an
input
of
a
neuron
is
repeatedly
and
persistently
causing
the
neuron
to
re,
a
metabolic
change
happens
in
the
synapse
of
that
par%cular
input
to
reduce
its
resistance
63
64
65
w1
Ac)va)on func)on
Input values
x2
w2
Summing func)on
()
Output y
xm
wm
weights
66
A
node
(neuron)
is
the
basic
informa)on
processing
unit
of
a
neural
net.
It
has:
n
A set of inputs with weights w1, w2, , wm along with a default input called the bias An adder func)on (linear combiner) that computes the weighted sum of the inputs
u = w jx j
j =1
n
An AcCvaCon funcCon (squashing func)on) that limits the amplitude of the neuron output.
y = (u + b)
67
+ 1 if v 0 (v ) = 1 if v < 0
x1
x2
xn
b
(bias)
w1
w2
wn
v
y
(v)
68
69
n n
70
input layer
data
CONFUSION!!!
sum
NO!
1
Umm...
?
0
-
sum
+
0
-
sum
+
0
-
sum
+
...Where
the
Input
Levels
Themselves
are
Determined
By
weighted
connec)ons
From
Other
Neurons!
Nega%ve
weights
0
output
layer
1
sum
-
-
-
-
-1-1-1-1=-4
input
layer
1
1
1
1
sum
+
+
+
+
input
layer
1
1
1
1
1+1+1+1=4
The
Net
as
a
Whole
Tries
to
Find
Connec)on
Weights
Such
That
its
Error
in
Minimized
Ini)al
weights
Error
w1
w2
1.5
1
0.8
0.6
0.4
0.2
0
0.3
0.6
0.9
1.2
1.5
-0.75
1
0.8
0.6
0.4
0.2
0
1.8
2.1
2.4
2.7
3
0.3
0.6
0.9
-2
0.5
1.2
1.5
1.8
2.1
-2.7
-2.4
-2.1
-1.8
-1.5
-1.2
-0.9
-0.6
-2.7
-2.4
-2.1
-1.8
-1.5
-1.2
-0.9
-0.6
-0.3
-0.3
1.527E-15
-1
-1
1 -0.5 -0.5 3
1.527E-15
-0.5 1 0.25
2.4
-3
-3
Image
Recogni)on
Given
noisy
and
incomplete
input
data
about
an
object
(I.e.
a
submarine),
iden)fy
the
type
of
the
object
I
N
P
U
T
S
O
U
T
P
U
T
S
Why are weights typically set around zero prior to training? Why are values normalized between zero and 1 prior to training?
Protability
Financial
InformaQon
Working Capital / Total Assets Leverage RaCo Firm Size (log(Total Assets)) Stock VolaClity and Return Market Value of Equity Distance to Default Default Probability
Market InformaQon
Risk Components:
Sub-Model Scores:
. . . . . . . . . . . . . .
Yn
+ + + + +
Y1 Y2 Yn
Y1 Y2 Yn
Financial Score
Equity Market
Risk Components:
Internal Components:
. . . . . . . . . . . .
Y1 Y2 Yn
Y1 Y2 Yn
Financial Score
Equity Market
Actual weight value Error surface for different neural network weights
Learning rate step size too big Weight bounces around but cant settle down to find minimum
Figure 5.11 Too big a step in the learning rate parameter can prevent a neural network from finding good weight settings
output layer
guesses
hidden layer
internal processing
input layer
data
o u t p u t sum +
Figure 5.4 The inside of a neurode. The thickness of each input is proportional to its weight. A solid denotes a positive weight and an unfilled one a negative weight
YES!
NO!
Umm...
0
sum +
0
sum +
0
sum +
Figure 5.5 How the sum of inputs affects the ouptut of a neurode. Notice how the rate of change of the output is nonlinear across the input range
1 output layer
input layer 1 1 1 1
Figure 5.7 A very simple neural network with no hidden layer. The numbers state that if the inputs to each neurode are 1 the output should be 1.
sum
input layer 1
-1-1-1-1=-4
sum
+
input layer 1
+
1+1+1+1=4
1 output layer
sum
input layer 0
+
0+0+1+1=2
0 output layer
sum
input layer 1
+
-1-1+0+0=-2
Figure 5.9 A more involved example of setting weight values to get desired output
98
99