Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
1 Introduction
Many pattern recognition approaches have been applied to electronic nose [1] [2].
However, until now, robustness and accuracy is still the main weakness for pattern
analysis on electronic nose. Problems such as drift compensation, mixture
separation and identification against complex odor backgrounds are still challenges.
In contrast with artificial electronic nose’s limitation, the mammalian olfactory
system can detect and interpret the information from volatile molecules in the
environment with a high degree of sensitivity, selectivity and stability. Therefore,
many researchers begin to pay more attention to biologically-inspired odor
processing models [3], to conquer difficulties in machine olfactory.
Milk classification using electronic nose is an especially difficult task due to the
heterogeneous nature of dairy product [4]. Difference in heat treatment, protein and
fat concentration can all affect aroma of milk [5]. Thereby, classification of milk
always deals with noisy data with complicated structure. This paper investigates
the performance of a novel classification approach by cascading bionic KIII model
and SVM in classifying reconstituted milk and fresh milk.
In olfactory system, each odorant activates a subset of the receptor cells and
initiates a spatial pattern of action potentials. This pattern initiates another spatial
pattern of activity in the outer layer of the olfactory bulb. These spatial patterns
have no specific topographic relation to the stimulus input pattern [7].
Discrimination among odors is a problem of spatial pattern recognition.
Manuscript by Nai Ding
Mimicking olfactory system, KIII model receives stimulus at receptor level and
transforms this stimulus pattern to amplitude modulation at olfactory bulb level. If
the standard deviations of responses of mitral cells are viewed as the output, KIII
model with N channels can be regarded as a nonlinear multiple-input
multiple-output system mapping input vector in its N-dimensional input space onto
its N-dimensional output space. KIII model aims at stimulating the information
processing phase in olfactory system but it does not model the decision making
function realized by higher level neural systems.
Learning Rule
Associative learning in olfactory bulb is by enhancement of mutually excitatory
synapses among the mitral level, creating a landscape of chaotic attractors. Each
attractor, formed in a self-organized way, represents a class.
Habituation is also an essential part of discrimination of sensory stimuli. It takes
place at the synapses of the excitatory neurons onto both inhibitory and other
excitatory neurons. A modified Hebbian learning rule with habituation is employed
to train KIII model [8].
2.2 SVM
Basically the SVM [9] [10] is a linear machine that applies a kernel method to map
the data into a higher dimensional space where a hyperplane can be used to do the
separation. It hinges on two mathematical operations [11]. One is to map the input
vector into a higher dimensional feature space in which non-linear separable
patterns become highly probable to be linearly separable. The other is to construct
an optimal hyperplane in the new feature space. The nonlinear map in operation
one is achieved by a kernel method. Each kernel function, satisfying Mercer’s
Theorem, corresponds to a space where the function is defined as an inner product.
In the new space a hyperplane is constructed in such a way that the margin of
separation between different classes is maximized. Different from
back-propagation algorithm devised to train a multilayer perceptron with
artificially designed structure, SVM automatically determines the required number
of hidden units (the number of support vectors (SV)). The decision function of
SVM can be expressed by equation (1)[12],
⎛ ⎞
I ( x ) = sign ⎜ ∑ α i K ( xi ⋅ x ) + b0 ⎟ (1)
⎝ SV in training set ⎠
K ( x . x ) represents the kernel function.
i
preprocessed by KIII.
other being ST − S K . S K is used as the training set for KIII and obtain the trained
network ϕ ( i, S ) , the whole training set S T is used as training set for SVM after
K
3 Experimental Results
3.1 Data Acquisition
Our experiments employ static head-space analysis. A tin oxide gas sensor array
with 8 sensors (TGS880 (2×), TGS813 (2×), TGS822 (2×), TGS800, TGS823)
from Figaro Engineering Inc. is mounted into a chamber. After equilibrated air
above a milk sample is injected into the chamber, sensors begin to be heated. The
step response of the gas sensor array is recorded. Diverse methods to extract
features from dynamic response of gas sensor array exist [16], but in our
experiment maximum response is adopted as the single feature for each sensor. As
mentioned in [17], when the ratio of samples to variables is low many erroneous
classifications may be made. Thereby, if too many features are extracted, a large
amount of measurements have to be conducted to make classification result
convincing, which brings difficulty to experiment. Classification rate when
maximum response is adopted as the single parameter is higher than those when
rise time, maximum slopes and stable response are adopted in our pretest.
Since sensor array’s response to milk is not strong, the impact of humidity and
temperature is not neglectable in our experiment. Equation (5) is used for the
purpose of drift compensation, with sensor response in fresh air as the baseline. To
compensate for differences in concentration between samples, each feature vector
is normalized using equation (6).
Robserved − Rbaseline R
R= (5) R= (6)
Rbaseline
∑ ( R (i ))
2
3.3 Experiments
Four experiments are conducted with different purposes. Experiment I and
Experiment II measure classifiers’ ability to classify fresh and reconstituted milk
from different perspective. Experiment III is designed to measure classifier’s
ability to detect reconstituted milk adulterated into fresh milk. Experiment IV is to
evaluate the impact of fresh milk’s concentration on adulteration detection.
For the first two experiments, all twelve brands of dairy are considered. The six
brands of fresh milk are regarded as one class and the six brands of reconstituted
milk are regarded as another. Difference between the two experiments lies in the
methods to select training set for classifiers. Experiment I uses ‘non-blind’
Manuscript by Nai Ding
In Experiment III, one brand of fresh milk and one brand of reconstituted milk
randomly selected are mixed together with different ratio. Samples with
reconstituted milk’s ratio under one value are deemed as one class and other
samples are deemed as another class. In Experiment IV, the two brands of dairy
used in Experiment III are used again. With reconstituted milk’s concentration
fixed, different amount of water is added into fresh milk. Classifiers trained in
Experiment III are employed to classify the diluted fresh milk and reconstituted
milk.
3.4 Results
Four classifiers’ performances in Experiment I with different training set size are
shown in fig.2. Result of Experiment II is shown in Fig.3 and Fig.4. Noting that
KIII-SVM outperforms SVM with 30-50 training samples, the size of KIII’s
training set for KIII-SVM-modified and KIII-SVM-ensemble are chosen to be 48
or 36. In Fig.4 approaches with KIII trained by 36 samples are noted as
KIII-SVM-modified(1) and KIII-SVM-ensemble(1) while approaches with KIII
trained by 48 samples are noted as KIII-SVM-modified(2) and
Manuscript by Nai Ding
Experiment III’s results are in Table 1. When 80% or more reconstituted milk is
adulterated into fresh milk, classifiers can achieve fairly satisfying classification
rate. Classifiers trained by all samples considered in this condition are tested in
Experiment IV. Different approaches’ performance with different levels of water
added into fresh milk is shown in Fig.5. The normalized Euclidean distances
between cluster center of reconstituted milk and cluster centers of milk water
mixture with 100%, 90%, 80%, 70%, 60% milk are 0.09 0.18 0.44 1.00
respectively. The Euclidean distance grows longer as the proportion of water grows
higher, but SVM’s classification rate dose not fall gradually.
SVM performs well for both data with simple or complex structure. KIII can not
enhance SVM’s accuracy directly. However, the involvement of KIII makes it easy
to construct ensemble classifiers. The size of training set of KIII does not affect
classification result significantly and the differences in training set for KIII result
in differently transformed data. Maintaining the size of training set for SVM large
enough enables SVM to make as accurate prediction as possible for data
transformed in different ways. The KIII-SVM-ensemble is shown to be the most
accurate approach for complex data in Experiment II.
set is one problem, but finding the optimum hyperplane reflecting the intrinsic
differences of different classes is another, when the training set does not carry
comprehensive information.
4 Discussion
Pattern recognition involves a process to transform a realistic scene to a feature
vector that is easy to deal with. It is a process to reduce dimensionality, to
strengthen differences between classes, and to make patterns more separable. This
process may be accomplished explicitly in signal processing phase or implicitly by
the classifier with different orientations. For example, discriminant function
analysis (DFA) and the kernel method used in SVM are oriented to map the data
into a space where the patterns become more separable. Isometric feature mapping
(ISOMAP), locally linear embedding (LLE) and other manifold learning method
aim at reducing dimension by learning the underlying global geometry of a data set.
Principle component analysis (PCA) and independent component analysis (ICA)
are designed to exploit the statistical structure of data. However, information
processing in biology system is highly complicated whose mechanism is still
unknown but is definitely not optimized according to one specific mathematical
object. As mentioned in [20], we need to characterize how representations are
transformed such that they are useful. KIII network can stimulate EEG waveform
in electrophysiological experiments and perhaps can partly stimulate the signal
processing mechanism in mammalian olfactory system. It may transform data to
strengthen the useful information in them rather than simply make them more
separable. The information processing mechanism of KIII is still unknown too
and is worth studying.
5. Conclution
This article demonstrates that KIII-SVM-ensemble is the best approach to classify
reconstituted milk from fresh milk considering its robustness and accuracy shown
in Experiment II and IV. Preprocessing feature vectors with KIII model is a choice
when data is complex and may suffer from unexpected noise.
References
(To be added)
Manuscript by Nai Ding
fresh: ensemble
reconstituted
86
84
85
84
82
correction rate %
correction rate %
83
80
82
81 KIII-SVM(ensemble)1
78
KIII-SVM(ensemble)2
KIII-SVM 80 SVM
SVM KIII-SVM(modified)1
76
minimum distance 79 KIII-SVM(modified)2
KIII-minimum distance KIII-SVM
78
74 60 65 70 75 80 85
10 20 30 40 50 60 70 80 90
size of training set
size of training set
Fig.2. Performance measures for different classifying Fig.4. Performance measures for different approaches
approaches with ‘non-blind’ training set selection cascading KIII and SVM with ‘non-blind’ training set
70
78
correction rate %
correction rate %
60
76
50
74
40
72
30
70
SVM
20
68 KIII-SVM(ensemble)
KIII-SVM
10
66 0.6 0.65 0.7 0.75 0.8 0.85 0.9 0.95 1
10 20 30 40 50 60 70 80 90
milk’s proportion in the milk-water mixture
size of training set
Fig.3. Performance measures for different Fig.5. Performance measures for different classifying
classifying approaches with ‘blind’ training set approaches with different proportion of water added
The training phase of KIII-SVM-ensemble is illustrated in fig.1.
The testing phase of KIII-SVM-ensemble is illustrated in fig.2.
The testing and training phases of KIII-SVM are illustrated in fig.3.