Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
During the past several years, the techniques developed from deep learning
research have already been impacting a wide range of signal and information
processing work within the traditional and the new, widened scopes including
key aspects of machine learning and artificial intelligence.
Three important reasons for the popularity of deep learning today are
These advances have enabled the deep learning methods to effectively exploit
complex, compositional nonlinear functions, to learn distributed and
hierarchical feature representations, and to make effective use of both labeled
and unlabeled data.
Deep learning typically uses artificial neural networks. The levels in these
learned statistical models correspond to distinct levels of concepts, where
higher level concepts are defined from lower level ones, and the same lower
level concepts can help to define many higher level concepts.
Using hidden layers with many neurons in a DNN significantly improves the
modeling power of the DNN and creates many closely optimal configurations.
Even if parameter learning is trapped in to a local optimum, the resulting DNN
can still perform quite well since the chance of having a poor local optimum is
lower than when a small number of neurons are used in the network. Using
deep and wide neural networks, however, would cast great demand to the
computational power during the training process.
Most machine learning and signal processing techniques had exploited shallow
structured architectures. These architectures typically contain at most one or
two layers of nonlinear feature transformations. Examples of the shallow
architectures are Gaussian mixture models (GMMs), Support vector machines
(SVMs), logistic regression, kernel regression, multilayer perceptrons (MLPs)
with a single hidden layer including extreme learning machines (ELMs). For
instance, SVMs use a shallow linear pattern separation model with one or zero
feature transformation layer. Shallow architectures have been shown effective
in solving many simple or well constrained problems, but their limited
modeling and representational power can cause difficulties when dealing with
more complicated real world applications involving natural signals such as
human speech, natural sound and language, and natural image and visual
scenes.
Deep learning algorithms are contrasted with shallow learning algorithms by
the number of parameterized transformations a signal encounters as it
propagates from the input layer to the output layer. Here parameterized
transformation is a processing unit the has trainable parameters, such as
weights and thresholds.
For simplicity, we can think DNNs as decision-making black boxes. They take
an array of numbers (that can represent pixels, audio waveforms, or words),
run a series of functions on that array, and output one or more numbers as
outputs. The outputs are usually a prediction of some properties we’re trying
to guess from the input, for example whether or not an image is a picture of a
cat.
The functions that are run inside the black box are controlled by the memory
of the neural network, arrays of numbers known as weights that define how
the inputs are combined and recombined to produce the results. Dealing with
real-world problems like cat-detection requires very complex functions, which
mean these arrays are very large, containing around 60 million numbers in the
case of one of the recent computer vision networks. The biggest obstacle to
using neural networks has been figuring out how to set all these massive
arrays to values that will do a good job transforming the input signals into
output predictions.
One of the theoretical properties of neural networks that has kept researchers
working on them is that they should be teachable. It’s pretty simple to show
on a small scale how you can supply a series of example inputs and expected
outputs, and go through a mechanical process to take the weights from initial
random values to progressively better numbers that produce more accurate
predictions.
Problems with deep neural networks If DNNs are naively trained, many
issues can arise. Two common issues are overfitting and computation time.
There are many training parameters to be considered with a DNN, such as the
size (number of layers and number of units per layer), the learning rate and
initial weights. Sweeping through the parameter space for optimal parameters
may not be feasible due to the cost in time and computational resources.