Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
MATHEMATICS
All the copyright of the present book belongs to the editors. All rights reserved. No part of this publication
may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, electronic,
mechanical, photocopying, recording, or otherwise, without the prior written permission of the editors.
All papers of the present volume were peer reviewed by no less than two independent reviewers.
Acceptance was granted when both reviewers' recommendations were positive.
ISSN: 2227-4588
ISBN: 978-1-61804-240-8
ADVANCES in APPLIED and PURE
MATHEMATICS
Program Chairs
• Prof. Filippo Neri
Dipartimento di Informatica e Sistemistica
University of Naples "Federico II"
Naples, Italy
• Prof. Constantin Udriste,
University Politehnica of Bucharest,
Bucharest
Romania
• Prof. Marcia Cristina A. B. Federson,
Universidade de São Paulo,
São Paulo, Brazil
Tutorials Chair
• Prof. Pradip Majumdar
Department of Mechanical Engineering
Northern Illinois University
Dekalb, Illinois, USA
Publication Chair
• Prof. Gen Qi Xu
Department of Mathematics
Tianjin University
Tianjin, China
Publicity Committee
• Prof. Vjacheslav Yurko,
Saratov State University,
Astrakhanskaya, Russia
• Prof. Myriam Lazard
Institut Superieur d' Ingenierie de la Conception
Saint Die, France
International Liaisons
• Professor Jinhu Lu, IEEE Fellow
Institute of Systems Science
Academy of Mathematics and Systems Science
Chinese Academy of Sciences
Beijing 100190, P. R. China
• Prof. Olga Martin
Applied Sciences Faculty
Politehnica University of Bucharest
Romania
• Prof. Vincenzo Niola
Departement of Mechanical Engineering for Energetics
University of Naples "Federico II"
Naples, Italy
• Prof. Eduardo Mario Dias
Electrical Energy and Automation
Engineering Department
Escola Politecnica da Universidade de Sao Paulo
Brazil
Steering Committee
• Prof. Stefan Siegmund, Technische Universitaet Dresden, Germany
• Prof. Zoran Bojkovic, Univ. of Belgrade, Serbia
• Prof. Metin Demiralp, Istanbul Technical University, Turkey
• Prof. Imre Rudas, Obuda University, Budapest, Hungary
Program Committee for PURE MATHEMATICS
Prof. Ferhan M. Atici, Western KentuckyUniversity, Bowling Green, KY 42101, USA
Prof. Ravi P. Agarwal, Texas A&M University - Kingsville, Kingsville, TX, USA
Prof. Martin Bohner, Missouri University of Science and Technology, Rolla, Missouri, USA
Prof. Dashan Fan, University of Wisconsin-Milwaukee, Milwaukee, WI, USA
Prof. Paolo Marcellini. University of Firenze, Firenze, Italy
Prof. Xiaodong Yan, University of Connecticut, Connecticut, USA
Prof. Ming Mei, McGill University, Montreal, Quebec, Canada
Prof. Enrique Llorens, University of Valencia, Valencia, Spain
Prof. Yuriy V. Rogovchenko, University of Agder, Kristiansand, Norway
Prof. Yong Hong Wu, Curtin University of Technology, Perth, WA, Australia
Prof. Angelo Favini, University of Bologna, Bologna, Italy
Prof. Andrew Pickering, Universidad Rey Juan Carlos, Mostoles, Madrid, Spain
Prof. Guozhen Lu, Wayne state university, Detroit, MI 48202, USA
Prof. Gerd Teschke, Hochschule Neubrandenburg - University of Applied Sciences, Germany
Prof. Michel Chipot, University of Zurich, Switzerland
Prof. Juan Carlos Cortes Lopez, Universidad Politecnica de Valencia, Spain
Prof. Julian Lopez-Gomez, Universitad Complutense de Madrid, Madrid, Spain
Prof. Jozef Banas, Rzeszow University of Technology, Rzeszow, Poland
Prof. Ivan G. Avramidi, New Mexico Tech, Socorro, New Mexico, USA
Prof. Kevin R. Payne, Universita' degli Studi di Milano, Milan, Italy
Prof. Juan Pablo Rincon-Zapatero, Universidad Carlos III De Madrid, Madrid, Spain
Prof. Valery Y. Glizer, ORT Braude College, Karmiel, Israel
Prof. Norio Yoshida, University of Toyama, Toyama, Japan
Prof. Feliz Minhos, Universidade de Evora, Evora, Portugal
Prof. Mihai Mihailescu, University of Craiova, Craiova, Romania
Prof. Lucas Jodar, Universitat Politecnica de Valencia, Valencia, Spain
Prof. Dumitru Baleanu, Cankaya University, Ankara, Turkey
Prof. Jianming Zhan, Hubei University for Nationalities, Enshi, Hubei Province, China
Prof. Zhenya Yan, Institute of Systems Science, AMSS, Chinese Academy of Sciences, Beijing, China
Prof. Nasser-Eddine Mohamed Ali Tatar, King Fahd University of Petroleum and Mineral, Saudi Arabia
Prof. Jianqing Chen, Fujian Normal University, Cangshan, Fuzhou, Fujian, China
Prof. Josef Diblik, Brno University of Technology, Brno, Czech Republic
Prof. Stanislaw Migorski, Jagiellonian University in Krakow, Krakow, Poland
Prof. Qing-Wen Wang, Shanghai University, Shanghai, China
Prof. Luis Castro, University of Aveiro, Aveiro, Portugal
Prof. Alberto Fiorenza, Universita' di Napoli "Federico II", Napoli (Naples), Italy
Prof. Patricia J. Y. Wong, Nanyang Technological University, Singapore
Prof. Salvatore A. Marano, Universita degli Studi di Catania, Catania, Italy
Prof. Sung Guen Kim, Kyungpook National University, Daegu, South Korea
Prof. Maria Alessandra Ragusa, Universita di Catania, Catania, Italy
Prof. Gerassimos Barbatis, University of Athens, Athens, Greece
Prof. Jinde Cao, Distinguished Prof., Southeast University, Nanjing 210096, China
Prof. Kailash C. Patidar, University of the Western Cape, 7535 Bellville, South Africa
Prof. Mitsuharu Otani, Waseda University, Japan
Prof. Luigi Rodino, University of Torino, Torino, Italy
Prof. Carlos Lizama, Universidad de Santiago de Chile, Santiago, Chile
Prof. Jinhu Lu, Chinese Academy of Sciences, Beijing, China
Prof. Narcisa C. Apreutesei, Technical University of Iasi, Iasi, Romania
Prof. Sining Zheng, Dalian University of Technology, Dalian, China
Prof. Daoyi Xu, Sichuan University, Chengdu, China
Prof. Zili Wu, Xi'an Jiaotong-Liverpool University, Suzhou, Jiangsu, China
Prof. Wei-Shih Du, National Kaohsiung Normal University, Kaohsiung City, Taiwan
Prof. Khalil Ezzinbi, Universite Cadi Ayyad, Marrakesh, Morocco
Prof. Youyu Wang, Tianjin University of Finance and Economics, Tianjin, China
Prof. Satit Saejung, Khon Kaen University, Thailand
Prof. Chun-Gang Zhu, Dalian University of Technology, Dalian, China
Prof. Mohamed Kamal Aouf, Mansoura University, Mansoura City, Egypt
Prof. Yansheng Liu, Shandong Normal University, Jinan, Shandong, China
Prof. Naseer Shahzad, King Abdulaziz University, Jeddah, Saudi Arabia
Prof. Janusz Brzdek, Pedagogical University of Cracow, Poland
Prof. Mohammad T. Darvishi, Razi University, Kermanshah, Iran
Prof. Ahmed El-Sayed, Alexandria University, Alexandria, Egypt
Table of Contents
Keynote Lecture 3: Multidimensional Optimization Methods with Fewer Steps Than the 16
Dimension: A Case of "Insider Trading" in Chemical Physics
Paul G. Mezey
Robust Adaptive Control with Disturbance Rejection for Symmetric Hyperbolic Systems of 48
Partial Differential Equations
Mark J. Balas, Susan A. Frost
Analytical Solution for Some MHD Problems on a Flow of Conducting Liquid in the Initial 61
Part of a Channel in the Case of Rotational Symmetry
Elena Ligere, Ilona Dzenite
ISBN: 978-1-61804-240-8 11
Advances in Applied and Pure Mathematics
Unary Operators 91
József Dombi
Flanged Wide Reinforced Concrete Beam Subjected to Fire - Numerical Investigations 132
A. Puskás, A. Chira
A New Open Source Project for Modeling and Simulation of Complex Dynamical Systems 138
A. A. Isakov, Y. B. Senichenkov
Analysis of Physical Health Index and Children Obesity or Overweight in Western China 148
Jingya Bai, Ye He, Xiangjun Hai, Yutang Wang, Jinquan He, Shen Li
New Tuning Method of the Wavelet Function for Inertial Sensor Signals Denoising 153
Ioana-Raluca Edu, Felix-Constantin Adochiei, Radu Obreja, Constantin Rotaru, Teodor
Lucian Grigorie
ISBN: 978-1-61804-240-8 12
Advances in Applied and Pure Mathematics
Gaussian Mixture Models Approach for Multiple Fault Detection - DAMADICS Benchmark 167
Erika Torres, Edwin Villareal
A Comparative Study on Principal Component Analysis and Factor Analysis for the 187
Formation of Association Rule in Data Mining Domain
Dharmpal Singh, J. Pal Choudhary, Malika De
Noise Studies in Measurements and Estimates of Stepwise Changes in Genome DNA 212
Chromosomal Structures
Jorge Munoz-Minjares, Yuriy S. Shmaliy, Jesus Cabal-Aragon
ISBN: 978-1-61804-240-8 13
Advances in Applied and Pure Mathematics
Keynote Lecture 1
Abstract: We show how real and complex Fourier transforms are extended to W.R. Hamilton's
algebra of quaternions and to W.K. Clifford's geometric algebras. This was initially motivated by
applications in nuclear magnetic resonance and electric engineering. Followed by an ever wider
range of applications in color image and signal processing. Clifford's geometric algebras are
complete algebras, algebraically encoding a vector space and all its subspace elements,
including Grassmannians (a vector space and all its subspaces of given dimension k).
Applications include electromagnetism, and the processing of images, color images, vector field
and climate data. Further developments of Clifford Fourier Transforms include operator
exponential representations, and extensions to wider classes of integral transforms, like Clifford
algebra versions of linear canonical transforms and wavelets.
ISBN: 978-1-61804-240-8 14
Advances in Applied and Pure Mathematics
Keynote Lecture 2
Robust Adaptive Control of Linear Infinite Dimensional Symmetric Hyperbolic Systems with
Application to Quantum Information Systems
Brief Biography of the Speaker: Mark Balas is presently distinguished faculty in Aerospace
Engineering at Embry-Riddle Aeronautical University. He was the Guthrie Nicholson Professor of
Electrical Engineering and Head of the Electrical and Computer Engineering Department at the
University of Wyoming. He has the following technical degrees: PhD in Mathematics, MS
Electrical Engineering, MA Mathematics, and BS Electrical Engineering. He has held various
positions in industry, academia, and government. Among his careers, he has been a university
professor for over 35 years with RPI, MIT, University of Colorado-Boulder, and University of
Wyoming, and has mentored 42 doctoral students. He has over 300 publications in archive
journals, refereed conference proceedings and technical book chapters. He has been visiting
faculty with the Institute for Quantum Information and the Control and Dynamics Division at
the California Institute of Technology, the US Air Force Research Laboratory-Kirtland AFB, the
NASA-Jet Propulsion Laboratory, the NASA Ames Research Center, and was the Associate
Director of the University of Wyoming Wind Energy Research Center and adjunct faculty with
the School of Energy Resources. He is a life fellow of the AIAA and a life fellow of the IEEE.
ISBN: 978-1-61804-240-8 15
Advances in Applied and Pure Mathematics
Keynote Lecture 3
Multidimensional Optimization Methods with Fewer Steps Than the Dimension: A Case of
"Insider Trading" in Chemical Physics
Abstract: "Insider trading" in commerce takes advantage of information that is not commonly
available, and a somewhat similar advantage plays a role in some specific, very high-
dimensional optimization problems of chemical physics, in particular, molecular quantum
mechanics. Using a specific application of the Variational Theorem for the expectation value of
molecular Hamiltonians, an optimization problem of thousands of unknowns does often
converge in fewer than hundred steps. The search for optimum, however, is typically starting
from highly specific initial choices for the values of these unknowns, where the conditions
imposed by physics, not formally included in the optimization algorithms, are taken into
account in an implicit way. This rapid convergence also provides compatible choices for "hybrid
optimization strategies", such as those applied in macromolecular quantum chemistry [1]. The
efficiency of these approaches, although highly specific for the given problems, nevertheless,
provides motivation for a similar, implicit use of side conditions for a better choice of
approximate initial values of the unknowns to be determined.
[1]. P.G. Mezey, "On the Inherited "Purity" of Certain Extrapolated Density Matrices", Computational and
Theoretical Chemistry, 1003, 130-133 (2013).
ISBN: 978-1-61804-240-8 16
Advances in Applied and Pure Mathematics
Keynote Lecture 4
Abstract: Designing of new version of Rand Model Designing under the name RMD 7 is coming
to an end. It will be possible using dynamic objects, dynamic connections (bonds), and arrays of
objects in the new version. These types are used for Simulation Modeling, and Agent Based
Modeling. The first trial version will be available at year-end.
The tools developed by MvStudium_Group are considered by authors as universal tools for
automation modeling and simulation of complex dynamical systems. We are feeling strongly
that at least nitty-gritty real system is multi-component, hierarchical, and event-driven system.
Modeling of such systems requires using object-oriented technologies, expressive graphical
languages and various mathematical models for event-driven systems. The last versions of
Model Vision Modeling Language are intended for multi-component models with variable
structure and event-driven behavior.
Brief Biography of the Speaker: PhD degree in Numerical Analysis from St. Petersburg State
University (1984).
Dr. Sci. degree (Computer Science) from St. Petersburg Polytechnic University (2005).
Author of 125 scientific publications-conference papers, articles, monographs and textbooks.
A board member of National Simulation Society - NSS (http://simulation.su/en.html), and
Federation of European Simulation Societies- EuroSim (http://www.eurosim.info/).
A member of Scientific Editorial Board of "Simulation Notes Europe" Journal (http://www.sne-
journal.org/), and "Computer Tools in Education" Journal(http://ipo.spb.ru/journal/).
Chairman and Chief-Editor of COMOD 2001-2014 conferences (https://dcn.icc.spbstu.ru/).
ISBN: 978-1-61804-240-8 17
Advances in Applied and Pure Mathematics
ISBN: 978-1-61804-240-8 18
Advances in Applied and Pure Mathematics
Abstract—We show how real and complex Fourier transforms II. C LIFFORD ’ S GEOMETRIC ALGEBRA
are extended to W.R. Hamiltons algebra of quaternions and to
W.K. Clifford’s geometric algebras. This was initially motivated by
Definition 1 (Clifford’s geometric algebra [15], [36]). Let
applications in nuclear magnetic resonance and electric engineering. {e1 , e2 , . . . , ep , ep+1 , . . ., en }, with n = p + q, e2k = εk ,
Followed by an ever wider range of applications in color image and εk = +1 for k = 1, . . . , p, εk = −1 for k = p + 1, . . . , n, be
signal processing. Cliffords geometric algebras are complete algebras, an orthonormal base of the inner product vector space Rp,q
algebraically encoding a vector space and all its subspace elements. with a geometric product according to the multiplication rules
Applications include electromagnetism, and the processing of images,
color images, vector field and climate data. Further developments of ek el + el ek = 2εk δk,l , k, l = 1, . . . n, (1)
Clifford Fourier Transforms include operator exponential represen-
tations, and extensions to wider classes of integral transforms, like where δk,l is the Kronecker symbol with δk,l = 1 for k = l,
Clifford algebra versions of linear canonical transforms and wavelets. and δk,l = 0 for k 6= l. This non-commutative product and the
additional axiom of associativity generate the 2n -dimensional
Keywords—Fourier transform, Clifford algebra, geometric algebra, Clifford geometric algebra Cl(p, q) = Cl(Rp,q ) = Clp,q =
quaternions, signal processing, linear canonical transform Gp,q = Rp,q over R. The set {eA : A ⊆ {1, . . . , n}}
with eA = eh1 eh2 . . . ehk , 1 ≤ h1 < . . . < hk ≤ n,
e∅ = 1, forms a graded (blade) basis of Cl(p, q). The grades
I. I NTRODUCTION k range from 0 for scalars, 1 for vectors, 2 for bivectors, s
for s-vectors, up to n for pseudoscalars. The vector space
We begin by introducing Clifford Fourier transforms, in- Rp,q is included in Cl(p, q) as the subset of 1-vectors. The
cluding the important class of quaternion Fourier transforms general elements of Cl(p, q) are real linear combinations
mainly along the lines of [31] and [5], adding further detail, of basis blades eA , called Clifford numbers, multivectors or
emphasize and new developments. hypercomplex numbers.
There is the alternative operator exponential Clifford Fourier In general hAik denotes the grade k part of A ∈ Cl(p, q).
transform (CFT) approach, mainly pursued by the Clifford The parts of grade 0 and k + s, respectively, of the geometric
Analysis Group at the university of Ghent (Belgium) [5]. product of a k-vector Ak ∈ Cl(p, q) with an s-vector Bs ∈
New work in this direction closely related to the roots of −1 Cl(p, q)
approach explained below is in preparation [11].
We mainly provide an overview of research based on the Ak ∗ Bs := hAk Bs i0 , Ak ∧ Bs := hAk Bs ik+s , (2)
holistic investigation [28] of real geometric square roots of are called scalar product and outer product, respectively.
−1 in Clifford algebras Cl(p, q) over real vector spaces Rp,q . For Euclidean vector spaces (n = p) we use Rn = Rn,0
These algebras include real and complex numbers, quater- and Cl(n) = Cl(n, 0). Every k-vector B that can be written
nions, Pauli- and Dirac algebra, space time algebra, spinor as the outer product B = b1 ∧ b2 ∧ . . . ∧ bk of k vectors
algebra, Lie algebras, conformal geometric algebra and many b1 , b2 , . . . , bk ∈ Rp,q is called a simple k-vector or blade.
more. The resulting CFTs are therefore perfectly tailored to Multivectors M ∈ Cl(p, q) have k-vector parts (0 ≤ k ≤
work on functions valued√ in these algebras. In general the n): scalar part Sc(M ) = hM i = hM i0 = M0 ∈ R, vector
continuous manifolds of −1 in Cl(p, q) consist of several part hM i1 ∈ R p,q
bi-vector part hM i2 , . . . , and pseudoscalar
conjugacy classes and their connected components. Simple Vn , p,q
part hM in ∈ R
examples are shown in Fig. 1. X
A CFT analyzes scalar, vector and multivector signals in M= MA eA = hM i + hM i1 + hM i2 + . . . + hM in . (3)
terms of sine and cosine waves with multivector coefficients. A
Basically, the imaginary unit i ∈ C in the √ transformation The principal reverse of M ∈ Cl(p, q) defined as
kernel ei φ = cos φ + i sin φ is replaced by a −1 in Cl(p, q). n
X k(k−1)
This produces a host of CFTs, an incomplete brief overview M
f= (−1) 2 hM ik , (4)
is sketched in Fig.√ 2, see also the historical overview in [5]. k=0
Additionally the −1 in Cl(p, q) allow to construct further often replaces complex conjugation and quaternion conjuga-
types of integral transformations, notably Clifford wavelets tion. Taking the reverse is equivalent to reversing the order
[21], [37]. of products of basis vectors in the basis blades eA . The
operation M means to change in the basis decomposition
E. Hitzer is with the Department of Material Science, International Christian of M the sign of every vector of negative square eA =
University, Mitaka, Tokyo, 181-8585 Japan e-mail: hitzer@icu.ac.jp.
thanks εh1 eh1 εh2 eh2 . . . εhk ehk , 1 ≤ h1 < . . . < hk ≤ n. Reversion,
Manuscript received May 31, 2014; revised ... M , and principal reversion are all involutions.
ISBN: 978-1-61804-240-8 19
Advances in Applied and Pure Mathematics
For M, N ∈ Cl(p, q) we get M ∗ N e = P MA NA . Two For A = M(2d, R), the centralizer (set of all elements
A
multivectors M, N ∈ Cl(p, q) are orthogonal if and only if in Cl(p, q) commuting with f ) and the conjugacy class of a
M ∗N e = 0. The modulus |M | of a multivector M ∈ Cl(p, q) square root f of −1 both have R-dimension 2d2 with two
is defined as connected components. For the simplest case d = 1 we have
X the algebra Cl(2, 0) isomorphic to M(2, R).
|M |2 = M ∗ Mf= MA2 . (5)
For A = M(2d, R2 ) = M(2d, R) × M(2d, R), the square
A
roots of (−1, −1) are pairs of two square roots of −1 in
A. Multivector signal functions M(2d, R). They constitute a unique conjugacy class with four
connected components, each of dimension 4d2 . Regarding the
A multivector valued function f : Rp,q → Cl(p, q), has 2n
four connected components, the group of inner automorphisms
blade components (fA : Rp,q → R)
X Inn(A) induces the permutations of the Klein group, whereas
f (x) = fA (x)eA . (6) the quotient group Aut(A)/Inn(A) is isomorphic to the group
A of isometries of a Euclidean square in 2D. The simplest
We define the inner product of two functions f, g : Rp,q → example with d = 1 is Cl(2, 1) isomorphic to M (2, R2 ) =
Cl(p, q) by M(2, R) × M(2, R).
Z For A = M(d, H), the submanifold of the square roots f
(f, g) = g dn x
f (x)g(x) of −1 is a single connected conjugacy class of R-dimension
Rp,q
Z 2d2 equal to the R-dimension of the centralizer of every f .
The easiest example is H itself for d = 1.
X
= eA efB fA (x)gB (x) dn x, (7)
A,B Rp,q For A = M(d, H2 ) = M(d, H) × M(d, H), the square
roots of (−1, −1) are pairs of two square roots (f, f 0 ) of −1
with the symmetric scalar part
Z in M(d, H) and constitute a unique connected conjugacy class
XZ
hf, gi = g dn x =
f (x) ∗ g(x) fA (x)gA (x) dn x, of R-dimension 4d2 . The group Aut(A) has two connected
Rp,q A Rp,q components: the neutral component Inn(A) connected to
(8) the identity and the second component containing the swap
automorphism (f, f 0 ) 7→ (f 0 , f ). The simplest case for d = 1
and the L2 (Rp,q ; Cl(p, q))-norm is H2 isomorphic to Cl(0, 3).
Z XZ
2 2 n For A = M(2d, C), the square roots of −1 are in bijection
kf k = h(f, f )i = |f (x)| d x = fA2 (x) dn x,
Rp,q Rp,q to the idempotents [2]. First, the ordinary square roots of −1
A
(9) (with k = 0) constitute a conjugacy class of R-dimension
2 p,q p,q 4d2 of a single connected component which is invariant under
L (R ; Cl(p, q)) = {f : R → Cl(p, q) | kf k < ∞}.
Aut(A). Second, there are 2d conjugacy classes of exceptional
(10)
square roots of −1, each composed of a single connected
component, characterized by the equality Spec(f ) = k/d (the
B. Square roots of −1 in Clifford algebras
pseudoscalar coefficient) with ±k ∈ {1, 2, . . . , d}, and their
Every Clifford algebra Cl(p, q), s8 = (p − q) mod 8, is R-dimensions are 4(d2 − k 2 ). The group Aut(A) includes
isomorphic to one of the following (square) matrix algebras1 conjugation of the pseudoscalar ω 7→ −ω which maps the
M(2d, R), M(d, H), M(2d, R2 ), M(d, H2 ) or M(2d, C). conjugacy class associated with k to the class associated with
The first argument of M is the dimension, the second the −k. The simplest case for d = 1 is the Pauli matrix algebra
associated ring2 R for s8 = 0, 2, R2 for s8 = 1, C for isomorphic to the geometric algebra Cl(3, 0) of 3D Euclidean
s8 = 3, 7, H for s8 = 4, 6, and H2 for s8 = 5. For even space R3 , and to complex biquaternions [42].
n: d = 2(n−2)/2 , for odd n: d = 2(n−3)/2 .
It has been shown [27], [28] that Sc(f ) = 0 for every
square root of −1 in every matrix algebra A isomorphic to C. Quaternions
Cl(p, q). One can distinguish ordinary square roots of −1,
Quaternions are a special Clifford algebra, because the
and exceptional ones. All square roots of −1 in Cl(p, q) can
algebra of quaternions H is isomorphic to Cl(0, 2), and to
be computed using the package CLIFFORD for Maple [1],
the even grade subalgebra of the Clifford algebra of three-
[3], [29], [38].
dimensional Euclidean space Cl+ (3, 0). But quaternions were
In all cases the ordinary square roots f of −1 constitute a
initially known independently of Clifford algebras and have
unique conjugacy class of dimension dim(A)/2, which has as
their own specific notation, which we briefly introduce here.
many connected components as the group G(A) of invertible
Gauss, Rodrigues and Hamilton’s four-dimensional (4D)
elements in A. Furthermore, we have Spec(f ) = 0 (zero
quaternion algebra H is defined over R with three imaginary
pseudoscalar part) if the associated ring is R2 , H2 , or C. The
units:
exceptional square roots of −1 only exist if A ∼ = M(2d, C).
1 Compare chapter 16 on matrix representations and periodicity of 8, as
well as Table 1 on p. 217 of [36]. ij = −ji = k, jk = −kj = i, ki = −ik = j,
2 Associated ring means, that the matrix elements are from the respective
2 2 2
ring R, R2 , C, H or H2 . i = j = k = ijk = −1. (11)
ISBN: 978-1-61804-240-8 20
Advances in Applied and Pure Mathematics
Every quaternion can be written explicitly as III. I NVENTORY OF C LIFFORD F OURIER T RANSFORMS
and has a quaternion conjugate (equivalent3 to Clifford con- Recently a rigorous effort was made in [8] to design a
jugation in Cl+ (3, 0) and Cl(0, 2)) general geometric Fourier transform, that incorporates most
of the previously known CFTs with the help of very general
q = qr − qi i − qj j − qk k, pq = q p, (13) sets of left and right kernel factor products
Z
which leaves the scalar part qr unchanged. This leads to the 0
FGF T {h}(ω) = L(x, ω)h(x)R(x, ω)dn x,
norm of q ∈ H Rp 0 ,q 0
Y
p q L(x, ω) = e−s(x,ω) , (22)
|q| = qq = qr2 + qi2 + qj2 + qk2 , |pq| = |p||q|. (14)
s∈FL
1
The part V (q) = q − qr = − q) = qi i + qj j + qk k is
2 (q with p0 + q 0 = n0 , FL = {s1 (x, ω), . . . , sL (x, ω)} a set of
0 0 0 0
called a pure quaternion, and it squares to the negative number mappings Rp√ ,q
× Rp ,q → I p,q into the manifold of real
−(qi2 + qj2 + qk2 ). Every unit quaternion (i.e. |q| = 1) can be multiples of −1 in Cl(p, q). R(x, ω) is defined similarly,
0 0
written as: and h : Rp ,q → Cl(p, q) is the multivector signal function.
q
q = qr + qi i + qj j + qk k = qr + qi2 + qj2 + qk2 µ(q)
= cos α + µ(q) sin α = eα µ(q) , (15) B. CFT due to Sommen and Buelow
This clearly subsumes the CFT due to Sommen and Buelow
where [7]
q Z n
Y
cos α = qr , sin α = qi2 + qj2 + qk2 , FSB {h}(ω) = h(x) e−2πxk ωk ek dn x, (23)
Rn k=1
V (q) qi i + qj j + qk k 2
µ(q) = = q , and µ(q) = −1.
|q| qi2 + qj2 + qk2 where x, ω ∈ Rn with components xk , ωk , and {e1 , . . . ek } is
an orthonormal basis of R0,n , h : Rn → Cl(0, n).
(16)
The inverse of a non-zero quaternion is
C. Color image CFT
−1 q q
q = 2 = . (17) It is further possible [16] to only pick strictly mutually com-
|q| qq √
muting sets of −1 in Cl(p, q), e.g. e1 e2 , e3 e4 ∈ Cl(4, 0)
The scalar part of a quaternion is defined as and construct CFTs with therefore commuting kernel factors
1 in analogy to (23). Also contained in (22) is the color image
Sc(q) = qr = (q + q), (18) CFT of [40]
2
Z
with symmetries FCI {h}(ω) =
1 1
e 2 ω·xI4 B e 2 ω·xB h(x)
R2
Sc(pq) = Sc(qp) = pr qr − pi qi − pj qj − pk qk , 1 1
e− 2 ω·xB e− 2 ω·xI4 B d2 x, (24)
Sc(q) = Sc(q), ∀p, q ∈ H, (19)
where B ∈ Cl(4, 0) is a bivector and I4 B ∈ Cl(4, 0) its
and linearity dual complementary bivector. It is especially useful for the
Sc(αp + βq) = αSc(p) + βSc(q) = αpr + βqr , introduction of efficient non-marginal generalized color image
Fourier descriptors.
∀p, q ∈ H, α, β ∈ R. (20)
The scalar part and the quaternion conjugate allow the defini-
D. Two-sided CFT
tion of the R4 inner product4 of two quaternions p, q as
The main type of CFT, which we will review here is the
Sc(pq) = pr qr + pi qi + pj qj + pk qk ∈ R. (21) general two sided CFT [24] with only one kernel factor on
Definition 2 (Orthogonality of quaternions). Two quaternions each side
p, q ∈ H are orthogonal p ⊥ q, if and only if the inner product
Z
0
F f,g {h}(ω) = e−f u(x,ω) h(x)e−gv(x,ω) dn x, (25)
Sc(pq) = 0. Rp0 ,q0
√ 0 0 0 0
3 This may be important in generalisations of the QFT, such as to a with f, g two −1 in Cl(p, q), u, v : Rp ,q × Rp ,q → R
space-time Fourier transform in [19], or a general two-sided Clifford Fourier p0 ,q 0 p,q
and often R = R . In the following we will discuss a
transform in [24].
4 Note that we do not use the notation p · q, which is unconventional for family of transforms, which belong to this class of CFTs, see
full quaternions. the lower half of Fig. 2.
ISBN: 978-1-61804-240-8 21
Advances in Applied and Pure Mathematics
√
E. Quaternion Fourier Transform (QFT) −1. The QFMT can be generalized straightforward to a
One of the nowadays most widely applied CFTs is the Clifford Fourier Mellin transform applied to signals h : R2 →
quaternion Fourier transform (QFT) [19], [26] Cl(p, q), p + q = 2 [23], with f, g ∈ Cl(p, q), p + q = 2.
Z
F f,g {h}(ω) = e−f x1 ω1 h(x)e−gx2 ω2 d2 x, (26)
R2
H. Volume-time CFT and spacetime CFT
which also has variants were one of the left or right kernel
factors is dropped, or both are placed together at the right or The spacetime algebra Cl(3, 1) of Minkowski space with or-
left side. It was first described by Ernst, et al, [14, pp. 307- thonormal vector basis {et , e1 , e2 , e3 }, −e2t = e21 = e22 = e33 ,
308] (with f = i, g = j) for spectral analysis in two- has three blades et , i3 , ist of time vector, unit space volume
dimensional nuclear magnetic resonance, suggesting to use 3-vector and unit hyperspace volume 4-vector, which are
the QFT as a method to independently adjust phase angles isomorphic to Hamilton’s three quaternion units
with respect to two frequency variables in two-dimensional
spectroscopy. Later Ell [12] independently formulated and e2t = −1, i3 = e1 e2 e3 = e∗t = et i−1 2
3 , i3 = −1,
explored the QFT for the analysis of linear time-invariant ist = et i3 , i2st = −1. (30)
systems of PDEs. The QFT was further applied by Buelow,
et al [6] for image, video and texture analysis, by Sangwine The Cl(3, 1) subalgebra with basis {1, et , i3 , ist } is therefore
et al [43], [5] for color image analysis and analysis of non- isomorphic to quaternions and allows to generalize the two-
stationary improper complex signals, vector image processing, sided QFT to a volume-time Fourier transform
and quaternion polar signal representations. It is possible to Z
split every quaternion-valued signal and its QFT into two
FV T {h}(ω) = e−et ωt h(x)e−~x·~ω d4 x, (31)
quasi-complex components [26], which allow the application R3,1
of complex discretization and fast FT methods. The split can
be generalized to the general CFT (25) [24] in the form with x = tet + ~x ∈ R3,1 , ~x = x1 e1 + x2 e2 + x3 e3 , ω =
1 ~ ∈ R3,1 , ω
ωt et + ω ~ = ω1 e1 + ω2 e2 + ω3 e3 . The split (27)
x± = (x ± f xg), x ∈ Cl(p, q). (27) with f = et , g = i3 = e∗t becomes the spacetime split of
2
special relativity
In the case of quaternions the quaternion coefficient space
R4 is thereby split into two steerable (by the choice of two 1
pure quaternions f, g) orthogonal two-dimensional planes [26]. h± = (1 ± et he∗t ). (32)
2
The geometry of this split appears closely related to the
quaternion geometry of rotations [39]. For colors expressed It is most interesting to observe, that the volume-time Fourier
by quaternions, these two planes become chrominance and transform can indeed be applied to multivector signal functions
luminance when f = g = gray line [13]. valued in the whole spacetime algebra h : R3,1 → Cl(3, 1)
without changing its form [19], [22]
F. Quaternion Fourier Stieltjes transform Z
Georgiev and Morais have modified the QFT to a quaternion FST {h}(ω) = e−et ωt h(x)e−i3 ~x·~ω d4 x. (33)
R3,1
Fourier Stieltjes transform [18].
Z The split (32) applied to spacetime Fourier transform (33)
1 2
FStj (σ , σ ) = e−f x1 ω1 dσ 1 (x1 )dσ 2 (x2 )e−gx2 ω2 , (28) leads to a multivector wavepacket analysis
R2
pseudoscalar −1 has been well studied and applied [20] KA (x1 , ω1 ) = √ ef (a1 x1 −2x1 ω1 +d1 ω1 )/(2b1 ) ,
1
f 2πb1
Z g 1 2 2
FP S {h}(ω) = h(x)e−in x·ω dn x, KA2 (x2 , ω2 ) = √ eg(a2 x2 −2x2 ω1 +d2 ω2 )/(2b2 ) . (37)
g2πb2
Rn
in = e1 e2 . . . en , n = 2, 3(mod 4), (35) The two-sided QLCT of signals h ∈ L1 (R2 , H) can now
generally be defined as
where h : Rn → Cl(n, 0), and {e1 , e2 , . . . , en } is the
orthonormal basis of Rn . Historically the special case of
Z
f,g f g
L (ω) = KA (x1 , ω1 )h(x)KA (x2 , ω2 )d2 x. (38)
(35), n = 3, was already introduced in 1990 [32] for the R2
1 2
b2 = 0, for which the LCT is essentially a chirp multiplication. f 2 = g 2 = −1, is only indicated for the case of the right-sided QLCT,
but with the restriction that f, g should be an orthonormal pair of pure
We now generalize the definitions of [34] using the fol- quaternions, i.e. Sc(f g) = 0. Otherwise [34] always strictly sets f = i
lowing two kernel functions with two pure unit quaternions and g = j.
ISBN: 978-1-61804-240-8 23
Advances in Applied and Pure Mathematics
Fig. 1. Manifolds [28] of square roots f of −1 in Cl(2, 0) (left), Cl(1, 1) (center), and Cl(0, 2) ∼
= H (right). The square roots are f = α+b1 e1 +b2 e2 +βe12 ,
with α, b1 , b2 , β ∈ R, α = 0, and β 2 = b21 e22 + b22 e21 + e21 e22 .
ISBN: 978-1-61804-240-8 24
Advances in Applied and Pure Mathematics
[9] J. Ebling, G. Scheuermann, Clifford convolution and pattern matching [33] K. Kou, J. Morais, Y. Zhang, Generalized prolate spheroidal wave
on vector fields, In Proc. IEEE Vis., 3, IEEE Computer Society, Los functions for offset linear canonical transform in Clifford Analysis, Math.
Alamitos, 2003. 193–200, Meth. Appl. Sci. 2013, 36 pp. 1028-1041. DOI: 10.1002/mma.2657.
[10] J. Ebling, G. Scheuermann, Cliff. Four. transf. on vector fields, IEEE [34] K. Kou, J-Y. Ou, J. Morais, On Uncertainty Principle for Quaternionic
Trans. on Vis. and Comp. Graph., 11(4), (2005), 469–479. Linear Canonical Transform, Abs. and App. Anal., Vol. 2013, IC 72592,
[11] D. Eelbode, E. Hitzer, Operator Exponentials for the Clifford Fourier 14 pp.
Transform on Multivector Fields, in preparation, 18 pages. Preprint: http: [35] Linwang Yuan, et al, Geom. Alg. for Multidim.-Unified Geogr. Inf.
//vixra.org/abs/1403.0310. System, AACA, 23 (2013), 497–518.
[12] T. A. Ell, Quaternionic Fourier Transform for Analysis of Two- [36] P. Lounesto, Clifford Algebras and Spinors, CUP, Cambridge (UK),
dimensional Linear Time-Invariant Partial Differential Systems. in Pro- 2001.
ceedings of the 32nd IEEE Conference on Decision and Control, Decem- [37] B. Mawardi, E. Hitzer, Clifford Algebra Cl(3,0)-valued Wavelet Transf.,
ber 15-17, 2 (1993), 1830–1841. Clifford Wavelet Uncertainty Inequality and Clifford Gabor Wavelets, Int.
[13] T.A. Ell, S.J. Sangwine, Hypercomplex Fourier transforms of color J. of Wavelets, Multiresolution and Inf. Proc., 5(6) (2007), 997–1019.
images, IEEE Trans. Image Process., 16(1) (2007), 22–35. [38] Waterloo Maple Incorporated, Maple, a general purpose computer
[14] R. R. Ernst, et al, Princ. of NMR in One and Two Dim., Int. Ser. of algebra system. Waterloo, http://www.maplesoft.com (
2012). c
Monogr. on Chem., Oxford Univ. Press, 1987. [39] L. Meister, H. Schaeben, A concise quaternion geometry of rotations,
[15] M.I. Falcao, H.R. Malonek, Generalized Exponentials through Appell Mathematical Methods in the Applied Sciences 2005; 28(1): pp. 101–126.
sets in Rn+1 and Bessel functions, AIP Conference Proceedings, Vol. [40] J. Mennesson, et al, Color Obj. Recogn. Based on a Clifford Fourier
936, pp. 738–741 (2007). Transf., in L. Dorst, J. Lasenby, ”Guide to Geom. Algebra in Pract.”,
[16] M. Felsberg, et al, Comm. Hypercomplex Fourier Transf. of Multidim. Springer 2011, 175–191.
Signals, in G. Sommer (ed.), ”Geom. Comp. with Cliff. Algebras”, [41] B. Rosenhahn, G. Sommer Pose estimation of free-form objects, Euro-
Springer 2001, 209–229. pean Conference on Computer Vision, Springer-Verlag, Berlin, 127, pp.
[17] M. Felsberg, Low-Level Img. Proc. with the Struct. Multivec., PhD thesis, 414–427, Prague, 2004, edited by Pajdla, T.; Matas, J.
Univ. of Kiel, Inst. fuer Inf. & Prakt. Math., 2002. [42] S. J. Sangwine, Biquaternion (Complexified Quaternion) Roots of −1.
[18] S. Georgiev, J. Morais, Bochner’s Theorems in the Framework of Adv. Appl. Clifford Algebras 16(1) (2006), 63–68.
Quaternion Analysis in: E. Hitzer, S.J. Sangwine (eds.), ”Quaternion and [43] S. J. Sangwine, Fourier transforms of colour images using quaternion, or
Clifford Fourier Transf. and Wavelets”, TIM 27, Birkhauser, Basel, 2013, hypercomplex, numbers, Electronics Letters, 32(21) (1996), 1979–1980.
85–104. [44] K.B. Wolf, Integral Transforms in Science and Engineering, Chapters
[19] E. Hitzer, Quaternion Fourier Transform on Quaternion Fields and 9&10, New York, Plenum Press, 1979. http://www.fis.unam.mx/∼bwolf/
Generalizations, AACA, 17 (2007), 497–517. integral.html
[20] E. Hitzer, B. Mawardi, Clifford Fourier Transf. on Multivector Fields [45] Y. Yang, K. Kou, Uncertainty principles for hypercomplex signals in
and Unc. Princ. for Dim. n = 2 (mod 4) and n = 3 (mod 4), P. Angles the linear canoncial transform domains, Signal Proc., Vol. 95 (2014), pp.
(ed.), AACA, 18(S3,4), (2008), 715–736. 67–75.
[21] E. Hitzer, Cliff. (Geom.) Alg. Wavel. Transf., in V. Skala, D. Hildenbrand [46] Yuan Linwang, et al, Pattern Forced Geophys. Vec. Field Segm. based
(eds.), Proc. GraVisMa 2009, Plzen, 2009, 94–101. on Clifford FFT, To appear in Computer & Geoscience.
[22] E. Hitzer, Directional Uncertainty Principle for Quaternion Fourier [47] Yu Zhaoyuan, et al, Clifford Algebra for Geophys. Vector Fields, To
Transforms, AACA, 20(2) (2010), 271–284. appear in Nonlinear Process in Geophysics.
[23] E. Hitzer, Clifford Fourier-Mellin transform with two real square roots
of −1 in Cl(p, q), p + q = 2, 9th ICNPAA 2012, AIP Conf. Proc., 1493,
(2012), 480–485.
[24] E. Hitzer, Two-sided Clifford Fourier transf. with two square roots of −1
in Cl(p, q), Advances in Applied Clifford Algebras, 2014, 20 pages, DOI:
10.1007/s00006-014-0441-9. First published in M. Berthier, L. Fuchs, C.
Saint-Jean (eds.) electronic Proceedings of AGACSE 2012, La Rochelle,
France, 2-4 July 2012. Preprint: http://arxiv.org/abs/1306.2092 .
[25] E. Hitzer, The Clifford Fourier transform in real Clifford algebras, in E.
H., K. Tachibana (eds.), ”Session on Geometric Algebra and Applications,
IKM 2012”, Special Issue of Clifford Analysis, Clifford Algebras and
their Applications, Vol. 2, No. 3, pp. 227-240, (2013). First published in
K. Guerlebeck, T. Lahmer and F. Werner (eds.), electronic Proc. of 19th
International Conference on the Application of Computer Science and
Mathematics in Architecture and Civil Engineering, IKM 2012, Weimar,
Germany, 0406 July 2012. Preprint: http://vixra.org/abs/1306.0130 .
[26] E. Hitzer, S. J. Sangwine, The Orthogonal 2D Planes Split of Quater-
nions and Steerable Quaternion Fourier Transf., in: E. Hitzer, S.J. Eckhard Hitzer Eckhard Hitzer is Senior Asso-
Sangwine (eds.), ”Quaternion and Clifford Fourier Transf. and Wavelets”, ciate Professor at the Department of Material Sci-
TIM 27, Birkhauser, Basel, 2013, 15–39. ence at the International Christian University in
[27] E. Hitzer, R. Abłamowicz, Geometric Roots of −1 in Clifford Algebras Mitaka/Tokyo, Japan. His special interests are theo-
Cl(p, q) with p + q ≤ 4. Adv. Appl. Clifford Algebras, 21(1), (2011) retical and applied Clifford geometric algebras and
121–144, DOI: 10.1007/s00006-010-0240-x. Clifford analysis, including applications to crystal
[28] E. Hitzer, J. Helmstetter, and R. Abłamowicz, Square roots of -1 in real symmetry visualization, neural networks, signal and
Clifford algebras, in: E. Hitzer, S.J. Sangwine (eds.), ”Quaternion and image processing. Additionally he is interested in
Clifford Fourier Transf. and Wavelets”, TIM 27, Birkhauser, Basel, 2013, environmental radiation measurements, renewable
123–153. energy and energy efficient buildings.
[29] E. Hitzer, J. Helmstetter, and R. Abłamowicz, Maple work-
sheets created with CLIFFORD for a verification of results in [28],
http://math.tntech.edu/rafal/publications.html (
2012).
c
[30] E. Hitzer, Quaternionic Fourier-Mellin Transf., in T. Sugawa (ed.),
Proc. of ICFIDCAA 2011, Hiroshima, Japan, Tohoku Univ. Press, Sendai
(2013), ii, 123–131.
[31] E. Hitzer, Extending Fourier transformations to Hamiltons quater-
nions and Cliffords geometric algebras, In T. Simos, G. Psihoyios
and C. Tsitouras (eds.), Numerical Analysis and Applied Mathematics
ICNAAM 2013, AIP Conf. Proc. 1558, pp. 529–532 (2013). DOI:
10.1063/1.4825544, Preprint: http://viXra.org/abs/1310.0249
[32] B. Jancewicz. Trivector Fourier transf. and electromag. Field, J. of Math.
Phys., 31(8), (1990), 1847–1852.
ISBN: 978-1-61804-240-8 25
Advances in Applied and Pure Mathematics
Abstract—The method of adaptive multilevel splitting is extended and c ∈ R. We apply adaptive multilevel splitting for the
by estimating the whole distribution function and not only the estimation of the whole distribution function of φ(X) up to
probability over some point. An asymptotic result is proved that is a certain threshold and not only of the probability over that
used to construct confidenc bands inside a closed interval in the tail
of the distribution function. A simulation study is performed giving point.
empirical evidence to the theory. Confidenc bands are constructed
using as an example the infinit norm of a multivariate normal
random vector. A. Algorithm
In this section, we present the algorithm for computing the
CDF of φ(X) via adaptive multilevel splitting using the idea
I. I NTRODUCTION of [7]. The algorithm is given below:
A powerful method for rare event simulation, is multilevel
splitting. Multilevel splitting, introduced in [8] and [11], has Algorithm II.1 (Computing the CDF via adaptive multilevel
become very popular the last decades. The main principle of splitting).
multilevel splitting, is to split the state space into a sequence 1) Input c ∈ R and N ∈ N. Set p = 1−1/N and c0 = −∞.
(1) (1)
of sub-levels that decrease to a rarer set. Then the probability 2) Generate N i.i.d. random elements X1 , . . . , XN ∼
of this rarer set is estimated by using all the subsequent X. Set j := 1.
(j)
conditional probabilities of the sub-levels. 3) Let cj = mini φ(Xi ).
One of the main problems of multilevel splitting is how to 4) If cj ≥ c set cj = c and GOTO 7.
fi the sub-levels in advance. Several adaptive methods were 5) Let
introduced by [3], [4], and [7]. The difference to this article (
(j) (j)
is that we are reporting all the hitting probabilities up to a (j+1) Xi if φ(Xi ) > cj
Xi = (j)
certain threshold i.e. constructing the distribution function. X̃i ∼ L(X|φ(X) > cj ) if φ(Xi ) ≤ cj ,
[7] have shown that several properties for their probability
estimator hold. They have shown that their estimator follows a where the X̃i are independent and also independent of
(j) (j)
certain discrete distribution. Using this distribution, they have {X1 , . . . , XN }.
constructed appropriate confidenc intervals, shown that the 6) Set j := j + 1 and GOTO 3.
estimator is unbiased and have given an explicit form of the 7) The CDF is given by the step function
variance. We extend their results by showing that the estimated j−1
X
distribution function follows a certain stochastic process and F̂ (k) = (1 − pi )I{ci ≤ k < ci+1 }, k ≤ c. (1)
given this process, we construct appropriate confidenc bands. i=0
We will apply adaptive multilevel splitting to construct the The critical part of this algorithm is the updating Step 5.
confidenc bands for the infinit norm of a multivariate normal At each iteration the paths have to be independent and drawn
random vector. The multivariate normal distribution is an from the conditional law. Those particles that were over the
important tool in the statistical community. The most important threshold cause no problem. The ones that were below have
algorithm that is used for the computation of multivariate to be killed and regenerate new particles, keeping the sample
normal probabilities is given in [5]. Another method, that is size at each iteration constant.
designed to exploit the infinit norm is given in [9]. Efficien ways to regenerate the particles have been studied
The article is structured as follows. In Section II, we in [3], [4] and [7] . In [4] adaptive multilevel splitting was
state and prove the main theorem and give the form of the considered for the estimation of very small entrance probabil-
confidenc bands. In Section III, we apply the method to the ities in strong Markov processes using an almost surely finit
multivariate normal distribution and construct the appropriate stopping time. The particles that were below the threshold
confidenc bands. A discussion is contained in Section IV. in Step 5 were killed. For each killed particle, a new particle
from the conditional law was generated by choosing uniformly
II. M ETHODOLOGY a survived particle and extending it from the firs point it
Suppose we want to estimate the hitting probability pc = crossed the threshold until the stopping time. [3] and [7]
P(φ(X) > c), where X is a random element on some considered adaptive multilevel splitting in static distributions.
a probability space (Ω, F, P) with distribution function µ, In this setting the particles could not be extended and some
together with some measurable real function φ(X) : Ω → R transition kernel (e.g. a Metropolis kernel) was constructed to
ISBN: 978-1-61804-240-8 26
Advances in Applied and Pure Mathematics
overcome this difficult . The new transition was accepted if σ = 0.1 σ = 0.3
its corresponding value was over the threshold.
1.00
1.0
Before we move to the next section, let us introduce some
0.9
0.95
notation. We denote the random variable Mc = max{m ∈ N :
acceptance probability
acceptance probability
cm ≤ c}, which depends on the number of particles N. The
0.8
0.90
adaptive multilevel splitting estimator for P(φ(X) > c) = 1 −
0.7
F (c), using the idea of [7], is then given by p̂c = (1−1/N )Mc .
0.85
Since we are interested in estimating the distribution function
0.6
0.80
F , we consider the stochastic process {Mc }c∈IF , where IF
0.5
denotes the support of F.
0.75
0.4
0.70
B. Construction of the confidence bands
We assume that the distribution function F of φ(X) is con- 0 500 1000 1500 0 500 1000 1500
tinuous and in addition that F is strictly increasing over some update step update step
closed interval I = [cmin , cmax ], such that 0 ≤ F (cmin ) < σ = 0.5 σ = 0.7
F (cmax ) < 1. The reason for introducing the interval I, is
1.0
1.0
that we want to construct confidenc bands over I in the tail
of the distribution function F.
0.8
0.8
We denote the survivor function and the integrated hazard
acceptance probability
acceptance probability
function of φ(X) by S(x) = 1−F (x) and Λ(x) = − log S(x),
0.6
respectively. We also defin the set I˜ = {Λ(c) : c ∈ I}, which
0.6
is also a closed interval given by I˜ = [Λ(cmin ), Λ(cmax )].
The next proposition proves that {Mc }c∈IF is a Poisson
0.4
0.4
0.2
Λ(c).
0.2
d
Proposition II.1. {Mc }c∈IF = {M̃Λ(c) }c∈IF , where
{M̃t }t∈R≥0 is a Poisson process of rate N. 0 500 1000 1500 0 500 1000 1500
Proof: The proof is exactly the same as Corollary 1 of update step update step
[7], but with a different conclusion. The random variable Mc
can be written as Fig. 1. Plots of the acceptance probability of the transition kernel for each
step of the algorithm. We use a sample size N = 100, t = 1000 with c = 20
Mc = max{m : cm ≤ c} and for covariance matrix Σ2 .
ISBN: 978-1-61804-240-8 27
Advances in Applied and Pure Mathematics
√
jumps of size 1, then Zt attains jumps of size 1/ N which N=100 N=1000
turns to 0 as N → ∞.
Finally, we need the predictable variation process < . > of S(c) S(c)
0.00015
0.00015
Zt to converge in probability to an increasing function H. In ^
S(c)
^
S(c)
our case we must have H(t) = t. We have < M̃t −N t >= N t
and as < βDt >= β 2 < Dt > for β ∈ R and any stochastic
0.00010
0.00010
process Dt , the third property holds. As the three sufficien
conditions hold, the result follows.
0.00005
0.00005
Using this lemma, we are now able to prove the theorem.
Proof of Theorem II.1: As log p̂c = Mc log(1 − 1/N ),
the process {Ac }c∈I is transformed to
1 4.0 4.1 4.2 4.3 4.4 4.5 4.0 4.1 4.2 4.3 4.4 4.5
Ac = √ (Mc − Mcmin − N (Λ(c) − Λ(cmin )))
N c c
d 1
= √ (M̃Λ(c) − M̃Λ(cmin ) − N (Λ(c) − Λ(cmin ))), c ∈ I. N=5000 N=10000
N
Consider the continuous time transformation c → Λ−1 (t + S(c) S(c)
0.00015
0.00015
Λ(cmin )). We get the transformed process {Ãt }t∈Iˆ, with ^
S(c)
^
S(c)
1
Ãt = √ (M̃t+Λ(cmin ) − M̃Λ(cmin ) − N t),
0.00010
0.00010
N
where Iˆ = [0, Λ(cmax ) − Λ(cmin )]. Since a Poisson process
is a Levy process, the process M̃t+Λ(cmin ) − M̃Λ(cmin ) is also
0.00005
0.00005
a Poisson process of rate N. So we can use Lemma 1 to say
d
{Ãt }t∈Iˆ → {Bt }t∈Iˆ.
As the map sup |.|, where the supremum runs over some
closed interval, is continuous between the set of cádlág func- 4.0 4.1 4.2 4.3 4.4 4.5 4.0 4.1 4.2 4.3 4.4 4.5
tions to the set of real numbers, by the continuous mapping
c c
theorem [2, Theorem 2.7], we get the convergence
Fig. 2. Plots of the survivor function S(c) and the its estimate Ŝ(c) using
d
sup |Ãt | → sup |Bt |. (3) different sample sizes N and c ∈ [4, 4.5] for covariance matrix Σ1 .
t∈Iˆ t∈Iˆ
√
Next, we use the result that {Bt }0≤t≤t̃ / t̃ is the same as and Ŵ (c) = Ŝ(c)/Ŝ(cmin ), where Ŝ(k) = p̂k and Λ̂(k) =
a standard Brownian motion on [0, 1] and also use Slutsky’s − log(p̂k ).
lemma. [7] have proved that p̂c is an unbiased estimator of
−1/N
S(c) for all c and also proved that var(p̂c ) = p2c (pc − Corollary 1. The conditional integrated hazard function Y (c),
1). As var(p̂c ) → 0 as N → ∞, by a standard result, we with c ∈ I, has α - confidence bands given by
d √
get consistency i.e. p̂c → S(c). As the function log(x) is Ŷ (c) hα b
d ±
Y (c) = ± a z , c ∈ I, (4)
continuous for all 0 < x < 1, we get − log(p̂c ) → Λ(c) for zN N N
all c. We have that where zN = −N log(1 − 1/N ) and hα is the α -quantile
p for the distribution function of supt∈[0,1] |B(t)|. Equivalently,
supt∈Iˆ |Bt | supt∈Iˆ |Bt | Λ(cmax ) − Λ(cmin ) we get confidence for the the conditional survivor distribution
√ =p p
b Λ(cmax ) − Λ(cmin ) − log(p̂cmax ) + log(p̂cmin ) W (c) by
d
W ± (c) = exp(−Y ± (c)), c ∈ I. (5)
→ sup |Bt |.
0≤t≤1 Proof: Use Theorem II.1 and solve appropriately .
d
As supc∈I |Ac | = supt∈Iˆ |Ãt |, we get the required result
III. C OMPUTING THE DISTRIBUTION FUNCTION OF THE
supc∈I |Ac | d INFINITY NORM OF A MULTIVARIATE NORMAL RANDOM
√ → sup |Bt |.
b 0≤t≤1 VECTOR
In this section, we apply Algorithm II.1 and construct
Next, we construct the confidenc bands for the conditional appropriate confidenc bands for both the integrated hazard
integrated hazard function Y (c) = Λ(c) − Λ(cmin ) and the function and the distribution function together with their
conditional survivor function W (c) = S(c)/S(cmin ). Their conditional versions. As an example, we use the multivariate
corresponding estimators are given by Ŷ (c) = Λ̂(c) − Λ̂(cmin ) normal distribution.
ISBN: 978-1-61804-240-8 28
Advances in Applied and Pure Mathematics
0.00020
S(c) S(c) W(c) W(c)
5e-06
5e-06
^ ^ ^ ^
S (c ) S( c) W(c) W(c)
0.00015
3e-06
3e-06
0.00010
0.00005
1e-06
1e-06
14.0 14.2 14.4 14.6 14.8 15.0 14.0 14.2 14.4 14.6 14.8 15.0 19.0 19.2 19.4 19.6 19.8 20.0 19.0 19.2 19.4 19.6 19.8 20.0
c c c c
N=5000 N=10000 N=5000 N=10000
0.00020
0.00020
5e-06
5e-06
^ ^ ^ ^
S (c ) S( c) W(c) W(c)
3e-06
3e-06
0.00010
0.00010
1e-06
1e-06
14.0 14.2 14.4 14.6 14.8 15.0 14.0 14.2 14.4 14.6 14.8 15.0 19.0 19.2 19.4 19.6 19.8 20.0 19.0 19.2 19.4 19.6 19.8 20.0
c c c c
Fig. 3. Plots of the survivor function S(c) and the its estimate Ŝ(c) using Fig. 4. Plots of the conditional survivor function W (c) and the its estimate
different sample sizes N and c ∈ [14, 15] for covariance matrix Σ2 . Ŵ (c) using different sample sizes N with cmin = 10 and c ∈ [19, 20] for
covariance matrix Σ2 .
ISBN: 978-1-61804-240-8 29
Advances in Applied and Pure Mathematics
1.0
1.0
6
Y(c) W(c) Λ(c) S(c)
2.5
^
Y(c)
^
W(c) Λ^ (c) ^
S(c)
0.8
0.8
2.0
0.6
0.6
1.5
0.4
0.4
1.0
0.2
0.5
0.2
0.0
0.0
0
14.0 14.5 15.0 15.5 16.0 14.0 14.5 15.0 15.5 16.0 0 2 4 6 8 10 12 0 2 4 6 8 10 12
c c c c
N=1000 N=1000 N=1000 N=1000
1.0
1.0
6
Y(c) W(c) Λ(c)
2.5
S(c)
^
Y(c)
^
W(c) Λ^ (c)
5
^
S(c)
0.8
0.8
2.0
0.6
0.6
1.5
0.4
1.0
0.4
0.2
0.5
0.2
1
0.0
0.0
0
14.0 14.5 15.0 15.5 16.0 14.0 14.5 15.0 15.5 16.0 0 2 4 6 8 10 12 0 2 4 6 8 10 12
c c c c
Fig. 5. Confidenc bands for the conditional integrated hazard function Y (c) Fig. 6. Confidenc bands for the integrated hazard function Λ(c) and the
and the conditional survivor function W (c) with covariance matrix Σ2 . We survivor function S(c) for c ∈ [0, 12] with covariance matrix Σ2 . We have
have different sample sizes N together with cmin = 14 and cmax = 16. used different sample sizes N.
with a choice of the very small sample size N = 100, the confidenc bands get narrower with increasing sample size. In
adaptive multilevel splitting estimator seems to outperform the all figures we notice that the true quantity always lies inside
usual empirical distribution estimator. We also notice that the the confidenc bands. The confidenc bands also mimics the
estimator is most likely to be on one side of the true quantity true quantity.
and especially for lower sample sizes.
We continue by testing the coverage probability of the confi Remark 1. One can apply Algorithm II.1 for the estimation of
dence bands given in (4), i.e. we are estimating the probability the multivariate t distribution. There are different versions of
that the conditional integrated hazard function Y (c) lies inside the t distribution. One such form is given in [12]. [6] describe
the confidenc bands Y ± (c), where c lies in the closed interval this form and in their Section 2.1 construct a crude Monte
I = [cmin , cmax ]. We are using α = 0.95 - quantile of the Carlo estimator. Considering the form of the estimator, it can
distribution function of supt∈[0,1] |B(t)| which is hα ≈ 2.24. be seen, given some transition kernel for the variables, that
Each estimation is based on 1000 replications using different our algorithm can be easily fit to this example.
sample sizes N ∈ {102 , 103 , 104 }. For covariance matrix Σ1 ,
we have used I = [4, 6] and got the coverage probabilities IV. D ISCUSSION
0.946, 0.946 and 0.955 for N = 102 , 103 , 104 respectively.
For the covariance matrix Σ2 , we have taken I = [14, 20] and We have extended the results of [7] by applying adaptive
we got 0.894, 0.916 and 0.942 respectively. In both cases, the multilevel splitting for the estimation of the whole distribution
coverage probability seems to converge to the true value 0.95, function and not only for the probability over some point.
giving empirical evidence to the Theorem II.1. A simulation study was performed, using as an example
In the remaining Figures 5 - 8, we are plotting confidenc the infinit norm of a multivariate normal random vector.
bands for the integrated hazard function and the survivor Confidenc bands were constructed for both the distribution
function or their conditional versions. Each plot is based on function and the integrated hazard function together with their
one run for sample sizes N ∈ {100, 1000}. As expected, the conditional versions. A test for the coverage probability of the
ISBN: 978-1-61804-240-8 30
Advances in Applied and Pure Mathematics
N=100 N=100
0.0030
9.0
Y(c) W(c)
^ ^
Y(c) W(c)
8.5
0.0020
8.0
7.5
0.0010
7.0
6.5
6.0
0.0000
18.0 18.5 19.0 19.5 20.0 18.0 18.5 19.0 19.5 20.0 N=100 N=100
17.5
6e-07
c c Λ(c) S(c)
N=1000 N=1000 Λ^ (c) ^
S(c)
0.0030
16.5
9.0
4e-07
Y(c) W(c)
^ ^
Y(c) W(c)
8.5
0.0020
8.0
15.5
2e-07
7.5
0.0010
7.0
14.5
0e+00
6.5
18.0 18.5 19.0 19.5 20.0 18.0 18.5 19.0 19.5 20.0
6.0
0.0000
c c
18.0 18.5 19.0 19.5 20.0 18.0 18.5 19.0 19.5 20.0 N=1000 N=1000
17.5
6e-07
c c Λ(c) S(c)
Fig. 7. Confidenc bands for the conditional integrated hazard function Y (c) Λ^ (c) ^
S(c)
and the conditional survivor function W (c) with covariance matrix Σ2 . We
16.5
c c
Fig. 8. Confidenc bands for the integrated hazard function Λ(c) and the
survivor function S(c) for c ∈ [18, 20] with covariance matrix Σ2 . We have
used different sample sizes N.
ISBN: 978-1-61804-240-8 31
Advances in Applied and Pure Mathematics
R EFERENCES
[1] Aalen, O., Borgan, Ø., and Gjessing, H.: Survival and event history
analysis: a process point of view. Springer Verlag (2008).
[2] Billingsley, P.: Convergence of probability measures. Wiley New York
(1999).
[3] Cérou, F., Del Moral, P., Furon, T., and Guyader, A.: Rare event
simulation for a static distribution. INRIA-00350762 (2009).
[4] Cérou, F. and Guyader, A.: Adaptive multilevel splitting for rare event
analysis. Stochastic Analysis and Applications, 25(2):417–443 (2007).
[5] Genz, A.: Numerical computation of multivariate normal probabilities.
Journal of Computational and Graphical Statistics, 1:141–149 (1992).
[6] Genz, A. and Bretz, F.: Comparison of methods for the computation of
multivariate t probabilities. Journal of Computational and Graphical
Statistics, 11(4):950–971 (2002).
[7] Guyader, A., Hengartner, N., and Matzner-Lber, E.: Simulation and
estimation of extreme quantiles and extreme probabilities. Preprint
(2011).
[8] Kahn, H. and Harris, T.: Estimation of particle transmission by random
sampling. National Bureau of Standards Applied Mathematics Series,
12:27–30 (1951).
[9] Phinikettos, I. and Gandy, A.: Fast computation of high dimensional
multivariate normal probabilities. Computational Statistics and Data
Analysis (2010).
[10] Rebolledo, R.: Central limit theorems for local martingales. Probability
Theory and Related Fields, 51(3):269–286 (1980).
[11] Rosenbluth, M. and Rosenbluth, A.: Monte Carlo calculation of the
average extension of molecular chains. The Journal of Chemical
Physics, 23(2):356–359 (1955).
[12] Tong, Y.: Multivariate normal distribution (1990).
ISBN: 978-1-61804-240-8 32
Advances in Applied and Pure Mathematics
Abstract—Chung, Graham and Leighton defined the guess- This result might be of independent interest.
ing secrets game in [1]. In this game, player B has to guess a
set of c > 1 secrets that player A has choosen from a set of N The paper is organized as follows. Section II gives an
secrets. To unveil the secrets, player B is allowed to ask a series
of boolean questions. For each question, A can adversarially overview of the coding theory concepts used throughout
select one of the secrets but once his choice is made he must the paper. Section III presents a formal description of
answer truthfully. In this paper we present a solution to the the game of guessing secrets for the case of c = 2
c = 2 guessing secrets problem consisting in an error correcting secrets. In Section IV we show that dual binary Hamming
code equipped with a tracing algorithm that, using the Bahl, codes give a solution to the game of guessing secrets. We
Cocke, Jelinek and Raviv algorithm as its underlying routine,
efficiently recovers the secrets. present a new analysis of the Bahl, Cocke, Jelinek and
Raviv algorithm in Section V. In Section VI, a tracing
Keywords—BCJR algorithm, coding theory, guessing secrets, algorithm that allows to recover the secrets, using the
separating codes.
BCJR algorithm as its underlying routine, is discussed.
Finally, our conclusions are given in Section VII.
I. Introduction
In the original “I’ve got a secret” TV game show [2] a II. Background on coding theory
contestant with a secret was questioned by four panelists. A. Binary (2,2)-separating codes
The questions were directed towards guessing the secret. A In this section we give a description of binary (2,2)-
prize money was given to the contestant if the secret could separating codes.
not be guessed by the panel. In this paper we consider a Let IFn2 be the vector space over IF2 , then C ⊆ IFn2 is
variant of the game, as defined by Chung, Graham and called a code. The field, IF2 is called the code alphabet. A
Leighton [1]. In this variant, called “guessing secrets”, code C is called a linear code if it forms a subspace of
there are two players A and B. Player A draws a subset IFn2 . The number of nonzero coordinates in x is called the
of c ≥ 2 secrets from a set Π of N secrets. Player B asks weight of x and is commonly denoted by w(x). The Ham-
a series of questions in order discover the secrets. We will ming distance d(a, b) between two words a, b ∈ IFnq is the
follow the approach of Alon, Guruswami, Kaufman and number of positions where a and b differ. The minimum
Sudan discussed in [3]. distance d of C, is defined as the smallest distance between
The game of guessing secrets is related to many different two different code words. If the dimension of the subspace
topics in communications and security such as separating is k, and its minimum Hamming distance is d, then we call
systems [4], efficient delivery of Internet content [1] and C an [n,k,d]-code. An error correcting code ⌊ of ⌋ minimum
the construction of schemes for the copyright protection distance d can correct up to can correct d−1 errors.
2
of digital data [3]. As a matter of fact, our results can A (n − k) × n matrix H, is a parity check matrix for the
be used as a tracing algorithm for the fingerprinting code code C, if C is the set of code words c for which Hc =
in [5]. 0, where 0 is the all-zero (n − k) tuple. Each row of the
matrix is called a parity check equation. A code whose code
A. Our contribution words satisfy all the parity check equations of a parity
check matrix is called a parity check code.
We present a solution to the guessing secrets problem For any two words a, b in IFnq we define the set of
consisting in a (2,2)-separating linear block code. We descendants D(a, b) as D(a, b) := {x ∈ IFnq : xi ∈
also design a tracing algorithm that, from the trellis {ai , bi }, 1 ≤ i ≤ n}. For
representation of the code, recovers the secrets using the ∪ a code C, the descendant code
C ∗ is defined as: C ∗ := a∈C,b∈C D(a, b).
Bahl, Cocke, Jelinek and Raviv (BCJR) algorithm [6] as If c ∈ C ∗ is a descendant of a and b, then we call a and
its underlying routine. The algorithm discussed consists of b parents of c.
several iterations of the BCJR algorithm that “corrects” A code C is (2, 2)-separating [4], if for any two disjoint
2 ⌋ + 1 errors, which is one
(in a list decoding flavor) ⌊ d−1 subsets of code words of size two, {a, b} and {c, d}, where
more error than the error correcting bound of the code. {a, b}∩{c, d} = ∅, their respective sets of descendants are
also disjoint, D(a, b) ∩ D(c, d) = ∅.
This work has been supported in part by the Spanish Government
through project Consolider Ingenio 2007 CSD2007-00004 “ARES” Next corollary from [7] gives a sufficient condition for a
and TEC2011-26491 “COPPI”. linear code to be (2,2)-separating.
Marcel Fernandez is with the Department of Telematics Engineer-
ing. Universitat Politècnica de Catalunya. C/ Jordi Girona 1 i 3.
Corollary 1 ([7]): All linear, equidistant codes are (2,2)-
Campus Nord, Mod C3, UPC. 08034 Barcelona. Spain. separating.
ISBN: 978-1-61804-240-8 33
Advances in Applied and Pure Mathematics
precisely it helps to compute the a posteriori probability configuration, when there is a single pair of secrets. And
(APP) functions: a triangle configuration, when there are three possible
1) P (St = m|rn1 ) (associated with each node in the disjoint pairs secrets.
trellis) that indicates the conditional probability of The solution for the c = 2 secrets problem will then
being in state m at time instant t given that the consist, in finding the appropriate star or triangle config-
received bit sequence is rn1 . uration for a given sequence of answers. Also, we require
2) P (St−1 = m′ ; St = m|rn1 ) (associated with each the strategy to be invertible [1], which means that, given
branch in the trellis) that indicates the joint a sequence of answers, there exists an efficient algorithm
probability of being in state m′ at time t − 1 and in capable of recovering the secrets.
state m at time t given that the received bit sequence
is rn1 . A. Explicit construction of the strategy
However it is simpler to obtain the joint probabilities
Following the discussion in [3], we denote the questions
1) The function λt (m) is defined as the joint probability
in an oblivious strategy as a sequence G of n boolean
of being in state m at time instant t and that the
functions gi : {1, . . . , N } → {0, 1}. For a given secret x
received bit sequence (word) is rn1 .
the sequence of answers to the questions gi will then be
C(x) = ⟨g1 (x), g2 (x), . . . , gn (x)⟩.
λt (m) = P (St = m; rn1 )
Without loss of generality we suppose that log2 N is
2) The function σt (m′ , m) is defined as the joint prob- an integer. In this case, using the binary representation
ability of being in state m′ at time instant t − 1, and for {1, . . . , N } we can redefine C as the mapping C :
in state m at time instant t, and that the received {0, 1}log2 N → {0, 1}n . From this point of view C can be
bit sequence is rn1 . seen as an error-correcting code. From now on we will
refer to a given strategy G using its associated code C,
σt (m′ , m) = P (St−1 = m′ ; St = m; rn1 ) and to the sequence of answers to a given secret using its
associated code word.
Note that since P (rn1 ) = λn (0) the APP probabilities
The question now is: which properties an error-
are easily obtained through the following expressions
correcting code must possess in order to solve the guessing
P (St = m; rn1 ) λt (m) secrets problem?. Depending on the sequence of answers,
P (St = m|rn1 ) = = (2) player B needs to recover a triangle or a star configuration.
P (rn1 ) λn (0)
In either case, he can use the following strategy. Use the
N = |Π| secrets as vertices to construct a complete graph
P (St−1 = m′ ; St = m; rn1 )
P (St−1 = m′ ; St = m|rn1 ) = KN . The pair of secrets (s1 , s2 ) can then be seen as an
P (rn1 ) edge of KN . Since we are considering each question as
σt (m′ , m) function gi : {1, . . . , N } → {0, 1}, the answer induces a
= (3)
λn (0) partition Π = gi−1 (0) ∪ gi−1 (1). If the answer of player A
to question gi is a ∈ {0, 1} and the pair of secrets chosen
III. Guessing two secrets with binary answers by A is (s1 , s2 ), we have that (s1 , s2 ) ∩ gi−1 (a) ̸= ∅. Now
In this section we present a formal description of the player B can remove all edges within the subgraph of KN
game of guessing secrets for the case of c = 2 secrets. spanned by gi−1 (1 − a). It follows that from the questions
In the first part of the game, player A draws exactly gi (1 ≤ i ≤ n), that B asks, he must be able to remove
two secrets S = {s1 , s2 }, from a set Π of N secrets. all edges until he is left with a subgraph “that contains no
Then, player B asks a series of boolean questions in pair of disjoint edges” [3].
order discover the secrets. For each question asked, A can We now show how the strategy described in the previous
adversarially choose a secret among the 2 secrets, but once paragraph can be accomplished using a certain code C. Let
the choice is made he must answer truthfully. C(s1 ), C(s2 ), C(s3 ) and C(s4 ) be the sequence of answers
We first note that there’s no way to guarantee that associated with four distinct secrets s1 , s2 , s3 and s4 . Note
player B can learn both secrets, since if all replies are that each sequence will correspond to a code word of
related to just one of the two secrets, then B cannot learn C. The questions that B asks, should have the following
nothing about the other. property: for every two disjoint pairs of secrets, there is
Note also, that B can never assert that a certain secret a question gi that allows to rule out at least one of the
is one of A’s secrets, since A can always take three secrets pairs. This implies that there should exist at least one
{s1 , s2 , s3 } and answer using a majority strategy. In this value i, i ∈ {1, . . . , n}, called the discriminating index for
case, the answer that A provides will be feasible for the which C(s1 )i = C(s2 )i ̸= C(s3 )i = C(s4 )i . A code with
three sets of secrets {s1 , s2 }, {s1 , s3 } and {s2 , s3 }. a discriminating index for every two disjoint pairs of code
Using the above reasoning, we see that for a given words, is called a (2,2)-separating code [4] and was defined
answer we have the following possible configurations for above in Section II. Moreover, such a code gives a strategy
the sets of secrets: A star configuration, when all pairs that solves the c = 2 guessing secrets game. Thus, we have
of secrets share a common element. A degenerated star proved the following lema.
ISBN: 978-1-61804-240-8 35
Advances in Applied and Pure Mathematics
Lemma 1: [3] A (2,2)-separating code solves the c = 2 and let y be a descendant generated by the coalition Γ.
guessing secrets game. Then
d
d(y, w) = d(y, u) = d(y, v) = , (5)
IV. Solution to the Guessing Secrets Game. 2
Dual binary Hamming codes
only if
We begin to discuss our work in this section. d
d(y, u) = d(y, v) = ,
2
Using Lemma 1 and the reasoning above we have the
following lemma whose proof is immediate. where u and v are different, and therefore the Hamming
Lemma 2: Let N be the number of secrets in the set of weight of the descendant y, denoted by w(y), satisfies w(y)
secrets. Without loss of generality suppose that N = 2k mod 2 = 0.
for a given k. Then a [n = 2k − 1, k, d = 2k−1 ] dual binary Note that in this case, in order to recover the secrets,
Hamming code solves the c = 2 guessing secrets game. we exceed the correcting capacity of the code and we will
have to use tailor made decoding algorithms such as the
Remark 1: Note that for a (2,2)-separating [n, k, d] ones discussed below in Section VI.
code solving the c = 2 guessing secrets game, n is the
number of questions, 2k is the number of secrets in the
set. Moreover, a descendant as defined in Section II is the V. The Bahl, Cocke, Jelinek and Raviv
sequence of answers given by player B, and the parents Algorithm
of this descendant are the secrets chosen by player A.
We continue to present our work in this section where
provide a new analysis of the Bahl, Cocke, Jelinek and
We now obtain some results on dual binary Hamming
Raviv algorithm described in Section II-D.
codes related to the guessing secrets problem. For lack of
space we do not provide the proofs.
We first show that considering a dual binary Hamming
A. Computation of the joint probability function σt (m′ , m)
code C, a 2-coalition Γ cannot generate any descendant
that is closer (in the Hamming sense) to a code word w ∈ For our purposes we will need the σt (m′ , m) function.
C − Γ than is to the coalition’s own code words, that is From the definitions in Section II-D, and the analysis in
the Appendix, the σt (m′ , m) function gives information
min{d(x, y)|x ∈ Γ} ≤ (w, y), ∀w ∈ C − Γ. about the transitions (symbols) in the trellis at time t.
According to Remark 1 this means that the code word of This transition information can be obtained by checking
one of the secrets chosen by player A, will as close to the what happens in the trellis before, after and at time t.
sequence of answers given by player B as any other code Intuitively σ gives us the probability of a given symbol in
word. a given position.
Proposition 1: Let C be a dual binary Hamming [n = Therefore, if t represents each position in the code, 1 ≤
2k − 1, k, d = 2k−1 ] code. Let Γ = {u, v} ⊂ C be a t ≤ n, we have that
coalition, and let y be a descendant generated by Γ. Symbol 0
Then, we always have that ∑
σt (m, m′ ) =
d d
d(w, y) ≥ and min{d(x, y)|x ∈ Γ} ≤ , branch(m,m′ )=0
2 2 1 ∑
P (r1 |c1 ) · · · P (rt |ct = 0) · · · P (rn |cn ) (6)
where w is any code word, w ∈ C − Γ. 2k
c:ct =0
From Proposition 1 it follows that the worst situation
is when Symbol 1
d
d(y, w) = d(y, u) = d(y, v) = , (4)
2 ∑
σt (m, m′ ) =
for some w ∈ C − Γ.
branch(m,m′ )=1
Note that the (2, 2)-separability of the dual binary 1 ∑
Hamming codes, determines that there can only exist P (r1 |c1 ) · · · P (rt |ct = 1) · · · P (rn |cn ) (7)
2k
a single code word w with this property. Moreover, for c:ct =1
Symbol 1
P r 0 σ1 (0, 0) 1 − 2ϵ + 6ϵ2 − 2ϵ3 + ϵ4 σ4 (0, 2) + σ4 (11, 3) + σ4 (12, 4) + σ4 (13, 5) =
= =
P r 1 t=1 σ1 (0, 7) (−1 + ϵ)2 (1 + ϵ)2 1
1 |B) + P (r1 |C) + P (r1 |F ) + P (r1 |G))
(P (rn n n n
≥ 1 for 0.3 ≤ ϵ ≤ 1 (10) 8
So the algorithm points us to the correct decision of a We have that code words A, D, E, H point us towards
‘0’ in the 1st position, and the error can be corrected. making a decision in favor of symbol ‘0’ whereas code
words B, C, F, G lead us to decide in favor of symbol
C. Correcting beyond the error correcting bound. Identify- ‘1’. Since we are assuming that the descendant is rt2 =
ing the parents of a descendant (1, 1, 0, 0, 0, 0, 0):
In this section we discuss a key property of the BCJR 1) The closest code words to rt2 = (1, 1, 0, 0, 0, 0, 0) are
algorithm. This property is essential for the results in this A, G, H. A and H lead us to decide in favor of a ‘0’
paper. Intuitively, this property consists of the following while G leads us towards a ‘1’. This intuitively says
fact. Suppose we run the BCJR algorithm using as input a that the symbol in this position should be a ‘0’.
descendant z of a certain coalition {u, v}. In the symbols 2) The other 2 code words that indicate us to make
of the descendant z where the parents u and v agree the a decision for a ‘0’, are D, E. They are at distance
BCJR algorithm returns a higher reliability. 4 of rt2 = (1, 1, 0, 0, 0, 0, 0), which is the minimum
This is better illustrated with an example. Suppose now distance of the code.
that we have the descendant rt2 = (1, 1, 0, 0, 0, 0, 0). Note By taking rt2 = (1, 1, 0, 0, 0, 0, 0) and again using (8), the
that rt2 can only be generated by the coalitions of code σ4 (m′ , m) expressions are:
words {A, G}, {A, H} and {G, H} (see Figure 1). Note Symbol 0
also that in the 6th position all three code words A, G, H σ4 (0, 0) + σ4 (1, 1) + σ4 (6, 6) + σ4 (7, 7) =
have a 0. On the other hand in the 4th position A and H (8) 1 1 ( )
have a 0 and G has a 1. We will see that for the 0 in the 6th = 7
2(1 + ϵ)5 (1 − ϵ)2 + 2(1 + ϵ)3 (1 − ϵ)4
82
position the BCJR algorithm outputs a higher reliability Symbol 1
than for the 0 in the 4th position. In other words we will
σ4 (0, 2) + σ4 (11, 3) + σ4 (12, 4) + σ4 (13, 5) =
prove the following proposition.
(8) 1 1 ( )
Proposition 3: Let Γ = {u, v} be two code words of a = 7
2(1 + ϵ)3 (1 − ϵ)4 + (1 + ϵ)(1 − ϵ)6 + (1 + ϵ)5 (1 − ϵ)2
82
dual binary Hamming code. Let z be a descendant created
by {u, v}. Then:
Therefore,
1) For a star configuration (see Section III), the output
reliabilities of the symbols of z given by the BCJR
d P r 0 σ4 (0, 0) + σ4 (1, 1) + σ4 (6, 6) + σ4 (7, 7)
algorithm will correspond to u if d(u, z) ≤ −1 and =
2 P r 1 t=4 σ4 (0, 2) + σ4 (11, 3) + σ4 (12, 4) + σ4 (13, 5)
to v otherwise.
2) For a degenerated star configuration (see Sec- (1 + ϵ)2
= >1 (11)
tion III), the output reliabilities of the symbols of 1 + ϵ2
z given by the BCJR algorithm will be larger for So the algorithm points us for a decision of a ‘0’ in
symbols in which u and v agree, than in symbols the 4th position. Note that the same reasoning applies to
where they differ. positions 1, 2, 3, 5, 7.
3) In case another code word say w forms a triangle
configuration with u and v (see Section III) then the We now evaluate the 6th position (which is the only
output reliabilities of the symbols of z given by the position where the 3 parents A, G, H have the same symbol
(‘0’)).
BCJR algorithm are largest in the positions where Symbol 0
u, v and w agree.
σ6 (0, 0) + σ6 (1, 1) =
For clarity, we prove Proposition 3 using an example.
1
1 |A) + P (r1 |B) + P (r1 |G) + P (r1 |H))
(P (rn n n n
1) Example. Proof of Proposition 3: Since the code is
8
equidistant d = 4. By by comparing rt2 = (1, 1, 0, 0, 0, 0, 0)
Symbol 1
with each code word we have that there are 3 code words
at exactly distance 2 (A, G, H), 4 code words at distance σ6 (2, 0) + σ6 (3, 1) =
4 (C, D, E, F ) and 1 code word at distance 6 (B) 1
1 |C) + P (r1 |D) + P (r1 |E) + P (r1 |F ))
(P (rn n n n
Again without loss of generality, we turn our attention 8
to the 4th position. We have seen that in this position In this position, we have that code words A, B, G, H point
Symbol 0
us towards making a decision in favor of symbol ‘0’ whereas
σ4 (0, 0) + σ4 (1, 1) + σ4 (6, 6) + σ4 (7, 7) = code words C, D, E, F lead us to decide in favor of symbol
1 ‘1’. Again, since we are assuming that the descendant is
1 |A) + P (r1 |D) + P (r1 |E) + P (r1 |H))
(P (rn n n n
8 rt2 = (1, 1, 0, 0, 0, 0, 0), we note that:
ISBN: 978-1-61804-240-8 38
Advances in Applied and Pure Mathematics
an efficient algorithm to recover the secrets. The explicit 2) The before function αt (m): The function αt (m) =
set of questions is based on a dual binary Hamming code. P (St = m; rt1 ) denotes the joint probability of being in
The recovery of the secrets consists in the decoding of a state m at time instant t and that the received bit sequence
block code beyond its error correction bound. In order to before (up to) time t is rt1 .
perform this decoding efficiently we present a modification Therefore,
of the BCJR algorithm, that passing through the trellis ∑
representing the block code, returns all the code words of αt (m) = αt−1 (m′ ) · γt (m′ , m)
m′
the code within distance d/2 of a given word.
Since we will always assume that at time t = 0 the state
of the trellis is m = 0, that is S0 = 0, then the boundary
Appendix conditions on alpha are
Given a trellis, to compute the joint probability func- α0 (0) = 1 (14)
tions λt (m) and σt (m′ , m) the following auxiliary func-
tions are used: α0 (m) = 0 for m ̸= 0
1) The before function αt (m) 3) The after function βt (m): The after function
βt (m) = P (rnt+1 |St = m) denotes the probability of
αt (m) = P (St = m; rt1 ) receiving the bit sequence rnt+1 after time t conditioned
on being in state m at time t.
that denotes the joint probability of being in state m Therefore,
at time instant t and that the received bit sequence ∑
before (up to) time t is rt1 . βt (m) = βt+1 (m′ ) · γt+1 (m, m′ )
2) The after function βt (m) m′
ISBN: 978-1-61804-240-8 41
Advances in Applied and Pure Mathematics
α0 (0) = 1 (20)
α0 (m) = 0 for m ̸= 0
The value of γt (m′ , m) is readily obtained from (19), and
therefore to obtain αt (m) we need αt−1 (m′ ) so we will be
moving through the trellis from left to right, i.e. in the
forward direction.
In Section VII-A2 we saw that the after function
βt (m) = P (rnt+1 |St = m) can be computed from the trellis
recursively using
∑
βt (m) = βt+1 (m′ ) · γt+1 (m, m′ )
m′
with the boundary conditions
βn (0) = 1 (21)
βn (m) = 0 for m ̸= 0
′
Again, the value of γt (m , m) is readily obtained
from (19), and therefore to obtain βt (m) we need βt+1 (m′ )
so we will be moving through the trellis from right to left,
i.e. in the backward direction.
ISBN: 978-1-61804-240-8 42
Advances in Applied and Pure Mathematics
Abstract—In the problem of high-frequency diffraction by satisfy the Helmholtz equations for the upper (j = 1) and
interface cracks in linear elastic materials we propose a numerical lower (j = 2) half-planes respectively:
method which is based on a separation of the oscillating solution r
and a certain slowly varying function. New effective approximate ∂ 2 wj ∂ 2 wj 2 ρj
factorization of the symbolic function is offered while using of + + kj w = 0, kj = ω , (2.1)
∂x2 ∂y 2 µj
the Wiener-Hopf method for high-frequency asymptotic solution.
This technique described in literature for regular (Fredholm) where µj and ρj designate elastic shear modulus and mass
integral equations is applied here to hyper-singular equations density for respective medium.
arising in diffraction by thin cracks on the boarder between two The boundary conditions correspond to stress-free faces of
different elastic media. The algorithm proposed is efficient for
both high and moderate frequencies.
the crack, and the continuity of the displacement and the stress
on the interface outside the crack. This implies:
Index Terms—diffraction, integral equation, high frequency,
numerical method. ∂w1 ∂w2
y=0: = = 0 , |x| ≤ a ;
∂y ∂y
I. I NTRODUCTION ∂w1 ∂w2
w1 = w2 , µ1 = µ2 , |x| > a . (2.2)
µ ¶ µ ¶
ik1 µ2 γ2 It thus can be seen that the leading term of the kernel’s
= 1− cos θ δ(s − k1 sin θ) − 1 + A2 , (2.7)
γ1 µ1 γ1 expansion for small x is hyper-singular, and kernel K0 (x) has
where the value of A1 in terms of A2 has been used, see the weak (integrable) logarithmic singularity only. A stable
Eq. (2.5). Now, Eqs. (2.3) and (2.7) imply: direct numerical algorithm to solve integral equations with
· such kernels is described in [8].
γ1 − ik1 cos θ
W2 (s, y) = µ1 2π δ(s − k1 sin θ)−
µ1 γ1 + µ2 γ2 III. A SYMPTOTIC A NALYSIS OF THE BASIC I NTEGRAL
¸ (2.8) E QUATION
γ1 Q(s) γ2 y
− e , In the high-frequency regime the numerical treatment of
µ1 γ1 + µ2 γ2
equation (2.9) becomes inefficient, because it is necessary to
and the remaining still unused boundary condition in (2.2), keep a fixed number of nodes per wave length. As a result, this
namely ∂w2 (x, 0)/∂y = 0, |x| ≤ a, by applying the inverse leads to a huge size of the discrete mesh. For this reason, let
Fourier transform to Eq. (2.8), results in the basic integral us construct an asymptotic solution of integral equation (2.9),
equation for the unknown function q(x): as ak1 → ∞. The method we use is allied to the classical
Zak1 ”Edge Waves” technique [4]. Let us represent the solution of
q(ξ)K(x−ξ)dξ =f (x), |x| ≤ ak1 ; (2.9) equation (2.9) as a combination of three functions:
−ak1 q(x) = q1 (ak1 + x) + q2 (ak1 − x) − q0 (x), (3.1)
Z∞ Z∞ satisfying, respectively, the following three equations:
1 1
K(x) = L(s)e−ixs ds = L(s)cos(xs)ds, Z∞
2π π
−∞ 0 q1 (ak1 + ξ)K(x− ξ)dξ = f (x)+
√ √
s2 − 1 s2 − k 2 µ1 2 k2 µ1 ρ2 −ak1
L(s) = √ √ , µ= , k = 22 = ,
2
µ s −1+ s −k 2 2 µ2 k1 µ2 ρ1 Z 1
−ak
"√ √ # + [q2 (ak1 − ξ)−q0 (ξ)]K(x − ξ)dξ, −ak1 < x < ∞, (3.2a)
( s2 −1−icos θ) s2 −k 2 −ixs
f (x) = √ √ e =Ae−ix sin θ , −∞
µ s2 −1 + s2 −k 2 s=sin θ Zak1
p
2 q2 (ak1 − ξ)K(x− ξ) dξ = f (x)+
−2icos θ sin θ−k 2
A =p
sin2 θ−k 2 −iµcos θ −∞
Z∞
written in a dimensionless form.
First of all, let us notice that the denominator of the fraction + [q1 (ak1+ ξ)− q0 (ξ)]K(x− ξ) dξ, − ∞ < x < ak1 , (3.2b)
in function L(s) does not vanish. This follows from the ak1
consideration
√ √of the three possible cases: (i) Both square roots Z∞
s2 − 1 and s2 − k 2 are real-valued. In this case they both q0 (ξ) K(x − ξ) dξ = f (x), −∞ < x < ∞ . (3.2c)
are positive, hence the sum of two positive quantities cannot −∞
possess the zero value; (ii) One of them is real-valued and
The equivalence of equation (2.9) and the system of three
the other one is imaginary. In this case the sum of these two
equations (3.2) is easily proved if one applies the combination
square roots may vanish if and only if they both are trivial
(3.2a)+(3.2b)–(3.2c).
that is impossible; (iii) Both these square roots are imaginary.
The leading asymptotic term of the solution can be con-
This case can be reduced to case (i), since the denominator is
structed by rejecting the residual integrals in the right-hand
the same quantity as in (i) multiplied by −i.
sides of (3.2a) and (3.2b). Under such a treatment, these
It is clear that kernel K(x) is even: K(x) = K(|x|).
two equations contain integral operators only in their left-
Besides, the kernel is smooth: K(|x|) ∈ C1 (0, 2ak1 ] outside
hand sides, becoming the Wiener-Hopf equations on semi-
the origin. In order to estimate its behavior as x → 0, let us
infinite intervals. As soon as they are solved, the correctness
extract explicitly the leading asymptotic term of function L(s)
of the hypothesis, that the rejected right-hand-side tails are
at infinity: L(s) = |s|/(µ + 1) + O(1/|s|), s → ∞. Therefore,
asymptotically small, can be checked by substituting the found
Z∞ solutions into those tail integrals. Physically, this means that
1
K(x) = s cos(xs) ds + K0 (x) = the reciprocal wave influence of the edges to each other is
π(µ + 1)
0 asymptotically small, in the first approximation.
1 It should be noted that the third equation (3.2c) is a simple
=− + K0 (x), (2.10) convolution integral equation on the infinite axis, and its
π(µ + 1) x2
solution is easily obtained by the Fourier transform (f (x) =⇒
Z∞ " √ 2 √ #
F (s)):
1 s − 1 s2 − k 2 s
K0 (x) = √ √ − cos(xs) ds = Z∞ Z∞
π µ s2 −1 + s2 −k 2 µ+1 1 F (s) −ixs δ(s−sin θ) −ixs
0 q0 (x) = e ds = a e ds =
2π L(s) L(s)
= O(ln |x|), x → 0. −∞ −∞
ISBN: 978-1-61804-240-8 44
Advances in Applied and Pure Mathematics
less or greater than the unit value, then the precision of the The second term of this asymptotic estimate (not writ-
approximation becomes considerably better. By allowing the ten here for the sake of brevity) shows that the difference
+
materials’ parameters to differ maximum by one order only: q1,2 (ak1 ± x) − q0 (x) not only tends to zero as x → ±∞ but
1/10 ≤ µ1 /µ2 , ρ1 /ρ2 ≤ 1, the worst case is µ1 /µ2 = 1, also is integrable at infinity. This guarantees the right-hand-
ρ1 /ρ2 = 1/10 with ε = 2% attained for B = 0.77. In side tails in Eq. (3.2) to be asymptotically small as ak1 → ∞,
practice, staying far from the extremal values of the physical that justifies the basic hypothesis permitting rejection of the
parameters, the maximum relative error is always less than tails.
1–2%. Finally, we note that similar problems have been studied in
With the introduced approximation (4.1) the efficient fac- [11] where a polynomial approximation form is applied to the
torization of symbolic function L(s) is taken in the form factorization problem, and in [12] where a numerical treatment
· √ √ ¸ of factorization is performed. The principal distinctive feature
(B + 1) s + 1 s + k
L(s) = √ √ √ × of the factorization proposed in the present work is that
µ + 1 (B s + 1 + s + k) + it catches all qualitative properties of the basic branching
· √ √ ¸ complex-valued symbolic function, in the way permitting the
(B + 1) s − 1 s − k
× √ √ √ = L+ (s)L− (s). (4.2) explicit-form solution of the posed problem. It also provides
µ + 1 (B s − 1 + s − k) −
any desired level of precision, by obvious combination of the
Omitting some transformations, one comes to the following proposed factorization (4.1) with the ideas presented in [11].
expression of Fourier transforms Q+
1,2 (s)
A(µ + 1)e± iak1 sin θ V. THE ESSENCE OF THE PROPOSED NUMERICAL METHOD
Q+
1,2 (s) = × (4.3)
(B + 1)2 (± sin θ − s) The basic idea of the proposed method is to seek the basic
µ ¶µ ¶ unknown function q(x) as a product of the high-frequency
1 B 1 B
× √ +√ √ +√ . asymptotic representation (3.1) and a certain slowly varying
1 ∓ sin θ k ∓ sin θ s+1 s+k
function G(x), namely
The Fourier inversion of this function may be performed
by passing to inverse Laplace transform, with the change q(x) = [q1 (ak1 + x) + q2 (ak1 − x) − q0 (x)] G(x),
is = −p, s = ip, where p is the Laplace parameter. Expression
(4.3) contains elementary functions with tabulated Laplace ak1 → ∞. (5.1)
inversions [10]: Physically, for extremely high frequencies function G must
1 eπi/4 tend to an identically unit value. For moderately high fre-
√ = √ quencies the new unknown function G(x) is a certain slowly
(α − s) s + β (p + iα) p − iβ
0 h i varying one, playing the role of modulating amplitude for
ie−iαx p
⇐= √ Erf e−πi/4 (α + β)x0 , (4.4) the rapidly oscillating component. One thus may take a small
α+β number of nodes, to find this function from the main integral
where Erf(z) is the probability integral. Since x0 = ak1 ± x, equation, uniformly for all high frequencies. The substitution
the inversion of (4.3) gives of (5.1) into Eqs. (2.9) rewrites it in the form
Z ak1
A(µ + 1)i 1
q (1,2) (ak1 ± x) = × [q1 (ak1 + ξ) + q2 (ak1 − ξ) − q0 (ξ)]G(ξ) dξ×
(B + 1)2 2π −ak1
µ ¶ Z ∞
1 B
× √ +√ e∓ix sin θ × × L(s)eis(ξ−x) ds = f (x), |x| ≤ ak1 . (5.2)
1 ∓ sin θ k ∓ sin θ −∞
h p i
Erf e−πi/4 (1 ± sin θ)(ak1 ± x) It should be noted that functions q(1,2) (x) in (4.5) can be
× √ + rewritten in terms of Fresnel integrals S2 (x), C2 (x) of real-
1 ± sin θ
valued arguments:
h p i
B Erf e−πi/4 (k ± sin θ)(ak1 ± x) q(1,2) (x) = D1,2 e∓ix sin θ [C2 (a1,2 x) + iS2 (a1,2 x)]+
+ √ . (4.5)
k ± sin θ
+E1,2 e∓ix sin θ [C2 (b1,2 x) + iS2 (b1,2 x)];
For large argument the probability integral tends to 1, then 3π √
B1,2 e−i 4 2
one can see that D1,2 = √ ; a1,2 = ks ± sin θ;
a1,2
+ A(µ + 1)i ∓ix sin θ 1
q1,2 (ak1 ± x) ∼ e (√ + 3π √
(B + 1)2 1 ∓ sin θ b1,2 e− 4 2
µ ¶ E1,2 = p ; b1,2 = 1 ± sin θ. (5.3)
B 1 B b1,2
+√ )× √ +√ =
k ∓ sin θ 1 ± sin θ k ± sin θ In discretization of Eq. (5.2) let us choose N equal sub-
Ae−ix sin θ intervals over full interval (−a, a), where N is the same for
= = 2 e−ix sin θ = q0 (x), x → ±∞ . (4.6)
L(sin θ) all large values of parameter ak1 . Then, assuming that the
ISBN: 978-1-61804-240-8 46
Advances in Applied and Pure Mathematics
Abstract— Given a linear continuous-time infinite- Our focus will be on applying our results to Symmetric
dimensional plant on a Hilbert space and disturbances of known Hyperbolic Systems of partial differential equations. Such
and unknown waveform, we show that there exists a stabilizing systems, originated by K.O. Friedrichs and P. D. Lax,
direct model reference adaptive control law with certain describe many physical phenomena such as wave behavior,
disturbance rejection and robustness properties. The closed electromagnetic fields, and the theory of relativistic quantum
loop system is shown to be exponentially convergent to a fields; for example, see [15]-[18]. To illustrate the utility of
neighborhood with radius proportional to bounds on the size of the adaptive control law, we apply the results to control of
the disturbance. The plant is described by a closed densely symmetric hyperbolic systems with coercive boundary
defined linear operator that generates a continuous semigroup conditions. Other closely related work on compensators for
of bounded operators on the Hilbert space of states. infinite dimensions from a different viewpoint can be found
Symmetric Hyperbolic Systems of partial differential
in [21].
equations describe many physical phenomena such as wave
behavior, electromagnetic fields, and quantum fields. To
II. ROBUSTNESS OF THE ERROR SYSTEM
illustrate the utility of the adaptive control law, we apply the
results to control of symmetric hyperbolic systems with coercive We begin by considering the definition of Strict
boundary conditions. Dissipativity for infinite-dimensional systems and the general
form of the “adaptive error system” to later prove stability.
Keywords: infinite dimensional systems, partial differential The main theorem of this section will be utilized in the
equations, adaptive control. following section to assess the stability of the adaptive
controller with disturbance rejection for linear diffusion
I. INTRODUCTION
systems.
Many control systems are inherently infinite dimensional
Noting that there can be some ambiguity in the literature
when they are described by partial differential equations.
with the definition of strictly dissipative systems, we modify
Currently there is renewed interest in the control of these
the suggestion of Wen in [8] for finite dimensional systems
kinds of systems especially in flexible aerospace structures
and expand it to include infinite dimensional systems.
and the quantum control field [1]-[2]. It is especially of
interest to control these systems adaptively via finite- Definition 1: The triple (Ac, B, C) is said to be Strictly
dimensional controllers. In our work [3]-[6] we have Dissipative if Ac is a densely defined ,closed operator on
accomplished direct model reference adaptive control and
disturbance rejection with very low order adaptive gain laws D( Ac ) X a complex Hilbert space with inner product
for MIMO finite dimensional systems. When systems are
subjected to an unknown internal delay, these systems are ( x, y) and corresponding norm x ( x, x) and
also infinite dimensional in nature. The adaptive control generates a C 0 semigroup of bounded operators U (t ) ,
theory can be modified to handle this situation [7]. However,
this approach does not handle the situation when partial and ( B, C ) are bounded finite rank input/output operators
with rank M where B : R X and C : X R . In
m m
differential equations describe the open loop system.
This paper considers the effect of infinite dimensionality addition there exist symmetric positive bounded operator P
( Px, x) pmax x , i.e. P is
2 2
on the adaptive control approach of [4]-[6]. We will show on X such that pmin x
that the adaptively controlled system is globally stable, but
the adaptive error is no longer guaranteed to approach the bounded and coercive, and
origin. However, exponential convergence to a neighborhood
1
Re( PAc e, e) 2 [( PAc e, e) (e, PAc e)]
can be achieved as a result of the control design. We will
prove a robustness result for the adaptive control which
extends the results of [4].
e
2
(1)
M.J. Balas is with the Aerospace Engineering Department, Embry- PB C *
Riddle Aeronautical University , Daytona Beach, FL 32119
(balsam@erau.edu).
where 0 and e D( Ac ) .
S.A. Frost is with the Intelligent Systems Division, NASA Ames
Research Center, Moffett Field, CA 94035(susan.frost@nasa.gov).
ISBN: 978-1-61804-240-8 48
Advances in Applied and Pure Mathematics
We also say that ( A, B, C ) is Almost Strictly then the gain matrix, G(t), is bounded, and the state, e(t)
at
Dissipative (ASD) when there exists a G*
mxm
such that exponentially with rate e approaches the ball of radius
The following theorem shows that convergence to a First we note that if see [10] Theo 8.8 p 151. Consider
neighborhood with radius determined by the supremum norm the positive definite function,
of is possible for a specific type of adaptive error system.
In the following, we denote tr(M M ) as
M 1 T
1
2
1
V ( Pe, e) tr G 1G T
2
2 (3)
the trace norm of a matrix M where 0 .
where G(t ) G(t ) G and P satisfies (1). Taking the
Theorem 2: Consider the coupled system of differential time derivative of (3) (we assume this can be done in X )
equations and substituting (2) into the result yields
e 1
t Ac e B G (t ) G z
V [( PAc e, e) (e, PAc e)] ( PBw, e)
2
G
tr G 1G T ( Pe, v)
y
e Ce (2)
G (t ) ey z aG (t ) where w Gz . Invoking the equalities in the definition of
T
G
is a Hilbert space with V Re( PAc e, e) e y , w
e1 e2 a tr G 1G T
inner product
1
, (e1 , e2 ) tr G1 G2 ,
1 2
G G
tr (e y z T G T ) ( Pe, v )
e
1
G e tr(G G )
2 1 2 ey , w
norm and where G (t ) is
e a tr (G G ) 1G T
2
the mxm adaptive gain matrix and is any positive definite
constant matrix, each of appropriate dimension. Assume the
( Pe, v)
following:
e a tr G 1G T
2
i.) ( A, B, C ) is ASD with Ac A BG*C
ii.) there exists M G 0 such that a tr G 1G T ( Pe, v)
tr(G*G*T ) M G 2 1
iii.) there exists M 0 such that p 2 ( Pe, e)
min
sup (t ) M 1 T
2a tr G G
t 0 1
2
iv.) there exists 0 such that a , where
pmax
a tr G 1G T ( Pe, v )
pmax is defined in Definition 1
2aV a tr G 1G T ( Pe, v)
v.) the positive definite matrix satisfies
2 Now, using the Cauchy-Schwartz Inequality
M
tr( ) ,
1
aM G
tr G 1G T G* 2
G 2
ISBN: 978-1-61804-240-8 49
Advances in Applied and Pure Mathematics
a G* pmax M
2
1 1 2 2
2 ( Pe, e) G 2
2 2
e(t )
e at
V (0)1/2
pmin
1
1
V2 (6)
pmax M
Therefore,
a pmin
1 e at
V 2aV
1
a G* pmax M
2
2 Taking the limit superior of (6), we have
1 M
2
V
tr ABC trCAB ,
pmax
Now, using the identity lim e(t ) R* (7)
a pmin
1 1
G*
2
tr G * 1 (G * )T tr (G * )T G * 1
2
2 End of Proof.
1
2 In order to accomplish some degree of disturbance
tr (G* )T G* (G* )T G* tr( )
1 1 2
rejection in a MRAC system, we shall make use of a
definition, given in [7], for the persistent disturbance:
tr G (G ) tr( )
* * T
2 1 2
persistent if it satisfies the disturbance generator equations:
M
MG
M u D (t ) z D (t ) u D (t ) z D (t )
aM G a or
zD (t ) Fz D (t ) z D (t ) LD (t )
which implies
V (t )1/2 V (0)1/2 e at
a
1 e at
(5) input and plant output m-vectors respectively, u D is a
ISBN: 978-1-61804-240-8 50
Advances in Applied and Pure Mathematics
disturbance with known basis functions D . We assume the Since B, C are finite rank operators, so is BGe*C .
columns of are linear combinations of the columns of B Therefore, Ac A BGe*C with D( Ac ) D(A) generates
(denoted Span( ) Span(B)). This can be relaxed a bit by
using ideal trajectories, but we will leave that to another a C0 semigroup U c (t ) because A does; see [9] Theo. 2.1
time. p. 497. Using equations (10), we have
The above system must have output regulation to a G G G*
neighborhood:
G
y t
N (0, R)
(9)
Since the plant is subjected to unknown bounded signals, we y T aG
cannot expect better regulation than (9). The adaptive
0
controller will have the form: where e 0. By Theorem 1, the corollary
0 D
u Ge y GD D
follows for and i sufficiently small.
Ge yy e aGe
T
(10) End of Proof.
GD y D D aG D
T
Corollary 1 provides a control law that is robust with
Using Theorem 1, we have the following corollary respect to persistent disturbances and unknown bounded
at
about the corresponding direct adaptive control strategy: disturbances, and, exponentially with rate e , produces:
Corollary 1: Assume the following:
*
i.) There exists a gain, Ge such that the triple lim y (t )
1 pmax BM
v.
a pmin
( AC A BG C, B, C ) is SD, i.e.
*
e
IV. SYMMETRIC HYPERBOLIC SYSTEMS
( A, B, C ) is ASD,
ii.) A is a densely defined ,closed operator on We will illustrate the above robust adaptive controller on
the following m input, m output Symmetric Hyperbolic
D( A) X and generates a C 0 semigroup of Problem:
bounded operators U (t ) ,
iii.) Span( ) Span(B) x
t Ax B(u uD ) v
Then the output y (t ) exponentially approaches a (c1 , x)
neighborhood with radius proportional to the magnitude of (c , x )
the disturbance, , for sufficiently small and i . 2 (11)
y Cx (c3 , x)
Furthermore, each adaptive gain matrix is bounded.
...
Proof:Since Span()Span(B), there exists a transformation (cm , x)
GD* such that L BGD* 0
Let, G Ge GD , G* Ge* GD* , and with inner product (v, w) (vT w)dz and is a
G G G* . Then u G Ge* y GD* D G bounded open set with smooth boundary, and where
T
y b1
where, . The error differential equation becomes b
D 2
x B b3 : m X linear; bi D( A) ,
t ( A BGe C ) x ( BGD L) D
* *
Ac 0 ...
bm
AC x Bw
x(0) x0 D( A) X L2N () , and
w G
bounded signal C : X m linear;ci D( A) .
For this application we will assume the disturbances are
step functions. Note that the disturbance functions can be any
ISBN: 978-1-61804-240-8 51
Advances in Applied and Pure Mathematics
ISBN: 978-1-61804-240-8 52
Advances in Applied and Pure Mathematics
systems using m actuator and m sensors and adaptive output
Ps x Pu x
2 2
feedback.We showed that under some limitations on operator
spectrum that we can accomplish robust adaptive control.
This allows the possibility of rather simple direct adaptive
x
2
control which also mitigates persistent disturbances for a
by choosing ge* . large class of applications in wave behavior, electromagnetic
fields, and some quantum fields.
Re( Ac x, x) x and C B* implies
2
Therefore,
that there exists APPENDIX
G g 0 such that
*
e
*
e
Proof of Theorem 2.
( Ac A BG C, B, C )
*
e is strictly dissipative with We will assume the operator A is closed. If not we can
work with the closure of A. It is easy to see that A is skew
PI . self-adjoint, i.e. ( Af , g ) ( f , Ag ) for all
Here is a simple first order symmetric hyperbolic system f , g D( A) . Therefore the spectrum of A must lie on the
example to illustrate some of the above:
imaginery axis and any complex (A) if it has nonzero
1 0 0
xt xz x u uD real part.
1 0 0 0 1 With coercive boundary conditions, it can be shown that
A1 A0 B for all x D( A) , we have
y 0 1 x 2
x
N
K 2 Ax x
2 2
C
i 1 zi
where 0 is small.
If we use and so D(A) H 1 .
G g* 0
*
We want to show that this resolvent operator is
min( g* , ) q1 q2
2 2
(sequentially) compact:
0 Take a bounded sequence
x
2
{xk }k 1 X and define hk R(1) xk D( A) .
q
x 1 .
2
hk
.
where N
K 2 Ahk hk
2 2
q2 Then
i 1 zi
Then ( Ac A BGe C, B, C ) is strictly dissipative with
*
hk xk .
In Theorem 1 we proved a robustness result for adaptive 2
hk xk
2 2
control under the hypothesis of almost strict dissipativity for Ahk
infinite dimensional systems. This idea is an extension of the
concept of m-accretivity for infinite dimensional systems; Therefore
see [9] pp278-280. In Cor 1, we showed that adaptive
regulation to a neighborhood was possible with an adaptive
controller modified with a leakage term. This controller
could also mitigate persistent disturbances. The results in
ISBN: 978-1-61804-240-8 53
Advances in Applied and Pure Mathematics
h
[13] M.Balas and R. Fuentes, “A Non-Orthogonal Projection Approach to
2
hk K2 xk hk
2 2 Characterization of Almost Positive Real Systems with an Application
k to Adaptive Control”, Proc of American Control Conference, 2004.
[14] P. Antsaklis and A. Michel, A Linear Systems Primer,Birkhauser,
K 2 hk xk
2
(1 K 2 ) hk
2 2007.
[15] L. Bers, F. John, and M. Schecter, Partial Differential Equations, J.
REFERENCES
[1] A. Pazy,Semigroups of Linear Operators and Applications to partial
Differential Equations,Springer 1983.
[2] D. D’Alessandro, Introduction to Quantum Control and Dynamics,
Chapman &Hall, 2008.
[3] Balas, M., R. S. Erwin, and R. Fuentes, “Adaptive control of
persistent disturbances for aerospace structures”, AIAA GNC,
Denver, 2000.
[4] R. Fuentes and M. Balas, "Direct Adaptive Rejection of Persistent
Disturbances", Journal of Mathematical Analysis and Applications,
Vol 251, pp 28-39, 2000
[5] Fuentes, R and M. Balas, “Disturbance accommodation for a class of
tracking control systems”, AIAA GNC, Denver, Colorado, 2000.
[6] Fuentes, R. and M. Balas, “Robust Model Reference Adaptive
Control with Disturbance Rejection”, Proc. ACC, 2002.
[7] M. Balas, S. Gajendar, and L. Robertson, “Adaptive Tracking Control
of Linear Systems with Unknown Delays and Persistent Disturbances
(or Who You Callin’ Retarded?)”, Proceedings of the AIAA
Guidance, Navigation and Control Conference, Chicago, IL,Aug
2009.
[8] Wen, J., “Time domain and frequency domain conditions for strict
positive realness”, IEEE Trans Automat. Contr., vol. 33, no. 10,
pp.988-992, 1988.
[9] T. Kato, Perturbation Theory for Linear Operators ,corrected 2nd
edition, Springer, 1980.
ISBN: 978-1-61804-240-8 54
Advances in Applied and Pure Mathematics
Keywords— control volume, crown fire, fire spread, forest fire, mathematical model, numerical method.
ISBN: 978-1-61804-240-8 55
Advances in Applied and Pure Mathematics
dT ∂ d
ρc p = (− ρc p v ′j T ′) + q 5 R5 − α v (T − Ts ) + Here and above is the symbol of the total
dt ∂ x j (3) dt
(substantial) derivative; αv is the coefficient of phase
− (q T− − q T+ ) / h;
exchange; ρ - density of gas – dispersed phase, t is time;
dcα ∂ vi - the velocity components; T, TS, - temperatures of gas
ρ = (− ρ v′j cα′ ) + R5α − Qcα + and solid phases, UR - density of radiation energy, k -
dt ∂ xj (4)
coefficient of radiation attenuation, P - pressure; cp –
+ ( Jα− − Jα+ ) / h, α = 1,5; constant pressure specific heat of the gas phase, cpi, ρi, ϕi
– specific heat, density and volume of fraction of
∂ c ∂ U R
− k (cU R − 4σTS4 ) + (q R− − q R+ ) / h = 0; (5) condensed phase (1 – dry organic substance, 2 –
∂ x j 3k ∂ x j
moisture, 3 – condensed pyrolysis products, 4 – mineral
part of forest fuel, 5 – gas phase), Ri – the mass rates of
4
∂ TS chemical reactions, qi – thermal effects of chemical
∑ρ c
i =1
i pi ϕi
∂t
= q 3 R3 − q 2 R2 + k (cU R − 4σTS4 ) (6) reactions; kg , kS - radiation absorption coefficients for
gas and condensed phases; Te - the ambient temperature;
+ α V (T − TS );
cα - mass concentrations of α - component of gas -
5
cα 5 5 dispersed medium, index α=1,2,3, where 1 corresponds
∑
α =1
c α = 1, Pe = ρRT ∑
α =1 M α
, g = ( 0, 0, g ), ∑
i =1
ϕ i = 1. to the density of oxygen, 2 - to carbon monoxide CO, 3 -
to carbon dioxide and inert components of air, 4 - to
The system of equations (1)–(6) must be solved taking particles of black, 5 - to particles of smoke; R – universal
into account the initial and boundary conditions gas constant; Mα , MC, and M molecular mass of α -
components of the gas phase, carbon and air mixture; g is
=t 0=
: v1 0,=
v2 0,=
v3 0,=
T Te ,=
cα cα e ,= ϕ1 ϕie ; (7)
Ts Te ,= the gravity acceleration; cd is an empirical coefficient of
the resistance of the vegetation, s is the specific surface
∂v
x1 =
− x1e : v1 ==
Ve , v2 0, 3 = 0, T =
Te , cα =
cα e , of the forest fuel in the given forest stratum. In system of
∂ x1 equations (1)-(6) are introduced the next designations:
(8)
c ∂ UR m = ρv3 , τ i = − ρ vi′ v3′ , J α = − ρ v3′ cα′ , J T = − ρ v3′ T ′
− + cU R / 2 =
0;
3k ∂ x1
Upper indexes “+” and “-” designate values of functions
∂v ∂ v2 ∂ v3 ∂ cα at x3=h and x3=0 correspondingly. It is assumed that heat
=x1 x1e =
: 1 0,= =
0, 0,= 0,
∂ x1 ∂ x1 ∂ x1 ∂ x1 (9)
and mass exchange of fire front and boundary layer of
atmosphere are governed by Newton law and written
∂T c ∂ UR c
= 0, + = UR 0 ; using the formulas
∂ x1 3k ∂ x1 2 (qT− − qT+ ) / h = −α (T − Te ) / h,
∂v ∂ v2 ∂ v3 ∂ cα ( J α− − J α+ ) / h = −α (c − cαe ) / hc p .
=x2 x20=
: 1 0,= 0,= =
0, 0,
∂ x2 ∂ x2 ∂ x2 ∂ x2 (10)
To define source terms which characterize inflow
∂T c ∂ UR c
=−
0, + UR =
0; (outflow of mass) in a volume unit of the gas-dispersed
∂ x2 3k ∂ x2 2 phase, the following formulae were used for the rate of
formulation of the gas-dispersed mixture m , outflow of
∂v ∂ v2 ∂ v3 ∂ cα
=x2 x2 e =
: 1 0,= =
0, 0,= 0, oxygen R51 , changing carbon monoxide R52
∂ x2 ∂ x2 ∂ x2 ∂ x2 (11) M M1
∂T c ∂ UR c Q = (1 − α c ) R1 + R2 + c R3 , R51 = − R3 − R5 ,
= 0, + =UR 0 . M1 2M 2
∂ x2 3k ∂ x2 2 R52 = ν g (1 − α c ) R1 − R5 , R53 = 0.
Here νg – mass fraction of gas combustible products of
pyrolysis, α4 and α5 – empirical constants. Reaction rates
ISBN: 978-1-61804-240-8 56
Advances in Applied and Pure Mathematics
of these various contributions (pyrolysis, evaporation, physical processes. In the first stage, the hydrodynamic
combustion of coke and volatile combustible products of pattern of flow and distribution of scalar functions was
pyrolysis) are approximated by Arrhenius laws whose calculated. The system of ordinary differential equations
parameters (pre-exponential constant ki and activation of chemical kinetics obtained as a result of splitting was
energy Ei) are evaluated using data for mathematical then integrated. A discrete analog was obtained by means
models [1]. of the control volume method using the SIMPLE like
algorithm [15]. The accuracy of the program was
E E checked by the method of inserted analytical solutions.
R1= k1 ρ1ϕ1 exp − 1 , R2= k2 ρ 2ϕ 2Ts−0.5 exp − 2 , Analytical expressions for the unknown functions were
RTs RTs
0.25
substituted in (1)–(6) and the closure of the equations
E cM c2 M −2.25 E were calculated. This was then treated as the source in
R3 k3 ρϕ 3 sσ c1 exp − 3 =
= , R5 k5 M 2 1 T exp − 5 .
RTs M1 M2 RT each equation. Next, with the aid of the algorithm
described above, the values of the functions used were
The initial values for volume of fractions of condensed inferred with an accuracy of not less than 1%. The effect
phases are determined using the expressions: of the dimensions of the control volumes on the solution
d (1 −ν z ) Wd α cϕ1e ρ1 was studied by diminishing them. The time step was
=ϕ1e = , ϕ 2e = , ϕ 3e selected automatically.
ρ1 ρ2 ρ3
Fields of temperature, velocity, component mass
where d -bulk density for surface layer, νz – coefficient fractions, and volume fractions of phases were obtained
of ashes of forest fuel, W – forest fuel moisture content. numerically. The distribution of basic functions shows
It is supposed that the optical properties of a medium are that the process of crown forest fire initiation goes
independent of radiation wavelength (the assumption that through the next stages. The first stage is related to
the medium is “grey”), and the so-called diffusion increasing maximum temperature in the fire source. At
approximation for radiation flux density were used for a this process stage the fire source a thermal wind is
mathematical description of radiation transport during formed a zone of heated forest fire pyrolysis products
forest fires. To close the system (1)–(6), the components which are mixed with air, float up and penetrate into the
of the tensor of turbulent stresses, and the turbulent heat crowns of trees. As a result, forest fuels in the tree
and mass fluxes are determined using the local- crowns are heated, moisture evaporates and gaseous and
equilibrium model of turbulence [1]. The system of dispersed pyrolysis products are generated. Ignition of
equations (1)–(6) contains terms associated with gaseous pyrolysis products of the ground cover occurs at
turbulent diffusion, thermal conduction, and convection, the next stage, and that of gaseous pyrolysis products in
and needs to be closed. The components of the tensor of the forest canopy occurs at the last stage. As a result of
turbulent stresses ρ vi′v ′j , as well as the turbulent fluxes heating of forest fuel elements of crown, moisture
of heat and mass ρ v′j c pT ′ , ρ v′j cα′ are written in terms of evaporates, and pyrolysis occurs accompanied by the
release of gaseous products, which then ignite and burn
the gradients of the average flow properties using the away in the forest canopy. At the moment of ignition the
formulas gas combustible products of pyrolysis burns away, and
∂v ∂ vj 2 the concentration of oxygen is rapidly reduced. The
− ρ vi′ v ′j = µ t i + − Kδ i j ,
∂ x ∂ xi 3 temperatures of both phases reach a maximum value at
j
the point of ignition. The ignition processes is of a gas -
∂T ∂ cα
− ρ v j c p T =′ λ t , − ρ v j cα′= ρ Dt , phase nature. Note also that the transfer of energy from
∂ xj ∂ xj the fire source takes place due to radiation; the value of
radiation heat flux density is small compared to that of
λt = µ t c p / Prt , ρDt = µ t / Sct , µ t = cµ ρK 2 / ε , the convective heat flux. At Ve ≠ 0, the wind field in the
where µt, λt, Dt are the coefficients of turbulent forest canopy interacts with the gas-jet obstacle that
forms from the forest fire source and from the ignited
viscosity, thermal conductivity, and diffusion,
forest canopy and burn away in the forest canopy.
respectively; Prt, Sct are the turbulent Prandtl and
Figures 2 - 5 present the distribution of temperature
Schmidt numbers, which were assumed to be equal to 1.
In dimensional form, the coefficient of dynamic
T (T = T / Te , Te = 300 K ) (1- 2., 2 – 2.6, 3 – 3, 4 – 3.5,
turbulent viscosity is determined using local equilibrium 5 – 4.) for gas phase, concentrations of oxygen c1 (1 –
model of turbulence [1]. The thermodynamic, 0.1, 2 – 0.5, 3 – 0.6, 4 – 0.7, 5 – 0.8, 6 – 0.9) and volatile
thermophysical and structural characteristics correspond
to the forest fuels in the canopy of a different (for combustible products of pyrolysis c 2 (1 – 1., 2- 0.1, 3 –
example pine) type of forest. The system of equations 0.05, 4 – 0.01) ( cα = сα / c1e , с1e = 0.23 ) and
(1)–(6) must be solved taking into account the initial and temperature of condensed phase
boundary conditions.
TS (TS = TS / Te , Te = 300 K ) (1- 2., 2 – 2.6, 3 – 3, 4 –
3. NUMERICAL METHOD AND RESULTS 3.5, 5 – 4.) for wind velocity Ve= 10 m/s at h=10 m: 1)
The boundary-value problem (1)–(6) is solved t=3 sec., 2) t=10 sec, 3) t=18 sec., 4) t= 24 sec. The
numerically using the method of splitting according to boundary-value problem is solved numerically using the
ISBN: 978-1-61804-240-8 57
Advances in Applied and Pure Mathematics
method of splitting according to physical processes. Fig.2. Field of isotherms of the forest fire spread (gas phase).
Fields of temperature, velocity, component mass
fractions, and volume fractions of phases were obtained
numerically. As a result of heating of forest fuel
elements of crown, moisture evaporates, and pyrolysis
occurs accompanied by the release of gaseous products,
which then ignite and burn away in the forest canopy. At
the moment of ignition the gas combustible products of
pyrolysis burns away, and the concentration of oxygen is
rapidly reduced. The temperatures of both phases reach a
maximum value at the point of ignition. The ignition
processes is of a gas - phase nature. Note also that the
Fig.3. The distribution of oxygen c1
transfer of energy from the fire source takes place due to Mathematical model and the result of the calculation
radiation; the value of radiation heat flux density is small give an opportunity to consider forest fire spread for
compared to that of the convective heat flux. At Ve ≠ 0, different wind velocity. Figures 6 (a, b, c, d) present the
the wind field in the forest canopy interacts with the gas- distribution of temperature for gas phase, concentration
jet obstacle that forms from the forest fire source and of oxygen and volatile combustible products of pyrolysis
from the ignited forest canopy and burn away in the
forest canopy. The isotherms and lines of equal levels of
c 2 concentrations and temperature of condensed phase
gas phase components concentrations were deformed and for wind velocity Ve= 5 m/s at h=10 m: 1) t=3 sec., 2)
moved in the forest canopy by the action of wind. It is t=10 sec, 3) t=18 sec., 4) t= 20 sec., 5) t=31 sec., 6) t=
concluded that the forest fire begins to spread. 40 sec.
Mathematical model and the result of the calculation
give an opportunity to consider forest fire spread for
different wind velocity, canopy bulk densities and
moisture forest fuel. The results obtained in this papers
show the decrease of the wind induces a decrease of the
rate of fire spread.
One of the objectives of this paper could be to
develop modeling means to reduce forest fire hazard in
forest or near towns. In this paper it presents numerical
results to study forest fire propagation through fire
breaks and around glades. The results of numerical
calculation present the forest fire front movement using
distributions of temperature at different instants of time
for various sizes of firebreaks. The fire break is situated Fig.4 The distribution of c2
in the middle of domain. In the first case the fire could
not spread through this fire break. If the fire break
reduces to 4 meters the fire continue to spread but the
isotherms of forest fire is decreased after overcoming of
fire break. The dependence of critical fire break value for
different wind velocities is presented in paper. Of course
the size of safe distance depends not only of wind
velocity, but type and quality of forest combustible
materials, its moisture, height of trees and others
conditions. This model allows studying an influence all
these main factors. The isotherms and lines of equal
levels are moved in the forest canopy and deformed by
the action of wind. Similarly, the fields of component Fig.5 Field of isotherms of the forest fire spread (solid phase).
concentrations are deformed. It is concluded that the
The results reported in Fig. 6 show the decrease of the
forest fire begins to spread.
wind induces a decrease of the rate of fire spread. One of
the objectives of this paper could be to develop modeling
means to reduce forest fire hazard in forest or near
towns. In this paper it presents numerical results to study
forest fire propagation through firebreak. This problem
was considered by Zverev [16] in one dimensional
mathematical model approach.
ISBN: 978-1-61804-240-8 58
Advances in Applied and Pure Mathematics
Fig.6 Fields of isotherms of gas (a) and solid phase (d), isolines
of oxygen(b) and gas products of pyrolysis(c).
Figures 7 and 8 (Figure 8 b is a continuation of Figure 8 Fig.8 Field of isotherms. Firebreak equals 4 m.
a) present the forest fire front movement using
distributions of temperature at different instants of time
for two sizes of firebreaks (4.5 and 4 meters). The fire
break is situated in the middle of domain (x1 =100 m). In
the first case the fire could not spread through this fire
break.
ISBN: 978-1-61804-240-8 59
Advances in Applied and Pure Mathematics
REFERENCES
[1] Grishin A.M., Mathematical Modeling Forest Fire and New
Methods Fighting Them, Tomsk: Publishing House of Tomsk
University (Russia), 1997.
[2] Van Wagner C.E., Conditions for the start and spread of crown
fire, Canadian Journal of Forest Research 7, 1977, pp. 23-34.
[3] Alexander V.E., Crown fire thresholds in exotic pine plantations
of Australasia, PhD diss., Australian National University, 1998.
[4] Van Wagner C.E., Prediction of crown fire behavior in conifer
stands, In Proceedings of 10th conference on fire and forest
meteorology. Ottawa, Ontario. (Eds D. C. MacIver, H. Auld and
R. Whitewood), 1989.
ISBN: 978-1-61804-240-8 60
Advances in Applied and Pure Mathematics
ISBN: 978-1-61804-240-8 61
Advances in Applied and Pure Mathematics
R in the channel later side, through which a conducting fluid with the following boundary conditions
flows into the channel with the constant velocity V0 ez (see
1, 0 r R
Fig.1). z 1 : Vr 0 , V z (6)
~
z 0, r R
h
z 1: Vr 0 , Vz 0 , (7)
D
0 r : Vr 0 , P / r 0 , (8)
~
r ~
where L1 f L0 f f / r 2 , R R/h .
~ ~
~
B e=B0ez R -h R
~
V V0 ez III. PROBLEM SOLVING
z h
Due to the rotational symmetry of the problem with respect
Fig. 1. The geometry of the flow. to r, the Hankel transform (see [1]) is used for the problem
solving. The Hankel transform of order 1 with respect to r is
The case of transverse magnetic field is considered, i.e.,
applied to the functions Vr and P / r , but the Hankel
when the external magnetic field B e B 0 e z is parallel to the transform of order 0 is applied to the functions Vz and P / z :
~z axis. It is also assumed that the channel walls z h are
non-conducting and induced streams do not flow through the
~ Vˆr ( , z ) Vr J1 ( r )rdr , Vˆz ( , z ) Vz J 0 ( r )r dr ,
hole { ~z h , 0 ~
r R }. 0 0
On introducing the dimensionless variables, when the half-
width of channel h is used as a length scale, the magnitude of Pˆ ( , z ) PJ0 ( r )r dr , (9)
the velocity of fluid in the entrance region V0 - as a velocity 0
for this problem (see [3],[4]), the problem takes the form
Vˆz 22 Ha2 Vˆz 4 Vˆz 0 .
( 4)
(15)
P
L1Vr Ha2Vr 0 , (3)
r The general solution of (15) has the form
P
L0V z 0 , (4) Vˆz C1 sinhk1 z C2 sinhk 2 z C3 coshk1 z C4 coshk 2 z , (16)
z
Vz 1
(r Vr ) 0 . (5) where k1 2 2 , k 2 2 2 , (17)
z r r
Ha / 2 , C 1 - C 4 are arbitrary constants.
ISBN: 978-1-61804-240-8 62
Advances in Applied and Pure Mathematics
In order to reduce the number of the constants C1-C4 in (16) As a result, one obtains
and simplify the problem solving, the problem is divided into
two sub-problems: an odd and even problem with respect to z, k cosh k1 sinhk 2 z k 2 cosh k 2 sinhk1 z
Vˆz 1 , (23)
on considering a plane channel with two holes in its lateral 1
sides z h in the region 0 r R .
In the odd problem with respect to z the fluid with where k1 , k 2 are given by (17),
velocities V0 e z / 2 flows into the channel through the both 1 k2 coshk2 sinhk1 k1 coshk1 sinhk 2 . (24)
holes at z h . The geometry of the flow for this problem is ( ) RJ1 (R) / 2 (25)
presented in Fig. 2.
In the even problem with respect to z the fluid with velocity The function V̂r is determined from (12):
V0 ez / 2 flows into the channel through the hole at ~z h
and flows out with the same velocity through the hole at cosh k1 cosh k 2 z cosh k 2 cosh k1 z
~ Vˆr (26)
z h . The geometry of the flow is presented in Fig. 3. 1
Then the solution of the general problem is equal to
the sum of solutions of the odd end even problems. Functions P̂ and dPˆ / dz are determined from (10)-(11):
A. Solution of the Odd Problem
k 2 cosh k 2 cosh k1 z k1 cosh k1 cosh k 2 z
The odd problem with respect to z is the problem on an Pˆ Ha 2 (27)
inflow of fluid into the channel through both holes at z h . 1
d Pˆ coshk 2 sinhk1 z coshk1 sinhk 2 z
~ Ha (28)
z ~
V
1
V0 ez dz 1
z h 2
h
Then on using the inverse complex Hankel transform,
the solution of the problem (3)–(5) with boundary conditions
0
(18)-(19) is obtained in the form of the convergent improper
~
r integrals:
h
~ ~
R R
coshk1 coshk 2 z cosh k 2 coshk1 z
~ ~ 1
V0 ez Vr J 1 (r )d , (29)
B e=B0ez V
1
z h 2 0
k1 coshk1 sinhk 2 z k 2 coshk 2 sinhk1 z
Fig. 2. The odd problem with respect to z. Vz J 0 (r ) d (30)
0
1
The dimensionless boundary conditions for the problem are
P k cosh k 2 cosh k1 z k1 cosh k1 cosh k 2 z
Ha2 2 J 1 (r ) d
r 1
1 / 2, 0 r R 0
z 1: Vr 0 , Vz (18) (31)
0, r R
P coshk 2 sinhk1 z coshk1 sinhk 2 z 2
r : Vx 0 , P / x 0 . (19) Ha J 0 (r )d (32)
z 0
1
On applying the Hankel transforms (9) to the boundary
conditions (18)-(19), one gets B. Solution of the Even Problem
The geometry of the flow for the even problem with
z 1: Vˆr 0 , Vˆz R J1 (R) /(2 ) (20) respect to z is shown in Fig. 3.
~
z
For the odd problem, V̂z is the odd function with ~ 1
V V0 ez
z h 2
respect to z and, therefore, C3 C4 0 in (16), i.e.
h
z 1 : Vˆz R J1 (R) /(2 ) and dVˆz / dz 0 . (22) Fig. 3. The even problem with respect to z.
ISBN: 978-1-61804-240-8 63
Advances in Applied and Pure Mathematics
The dimensionless boundary conditions for this P sinhk1 cosh k 2 z sinhk 2 cosh k1 z
problem are Ha 2 J 0 (r )d (44)
z 0
2
1 / 2, 0 r R
z 1: Vr 0 , Vz (33) where k1 , k 2 are given by (17) and by (25) .
0, r R
r : Vx 0 , P / x 0 . (34) IV. NUMERICAL RESULTS
On applying the Hankel transform (9) to boundary A. Numerical Results for the Odd Problem
conditions (32)-(33), one gets On the base of obtained solution, the velocity field is
studied numerically. Results of calculations of the
z 1: Vˆr 0 , Vˆz RJ1 (R) /(2 ) (35) velocity radial component Vr for the odd problem at
the Hartmann numbers Ha=10 and Ha=50 are presented
For the even problem, the function V̂z is even with graphically in Fig. 4. The component Vr is an odd
respect to z, therefore, C1 C2 0 in (16), i.e., function with respect to z. It can be seen from Fig.4
that Vr has the M-shaped profiles only near the
Vˆz (, z) C3 coshk1 z C4 coshk2 z . (36) entrance hole ( 1 r 1.1 at На=10 and 1 r 1.1
На=50). Even at a small distance from the entrance, the
In order to determine C3 and C 4 , the boundary profiles of Vr take the shape peculiar to the Hartmann
condition (35) and (12) are used, as a result, one gets flow in a plane channel in transverse magnetic field.
The magnitude of the velocity is inversely proportional to
k sinhk 2 cosh k1 z k1 sinhk1 cosh k 2 z
Vˆz 2 (37) the distance from the hole.
2
Vr Ha=10
r=1.01
where k1 , k 2 are given by (17), is given by (25) and 0.25 r=1.1
0.15 r=2
The function V̂r is determined from (12):
0.1 r=3
r=4
r=6
sinhk1 sinhk 2 z sinhk 2 sinhk1 z 0.05
Vˆr (38)
2 0.2 0.4 0.6 0.8 1
z
Vr Ha=50
0.3
The functions P̂ and dPˆ / dz are determined from (10)- 0.25
r=1.01
r=1.1
(11):
0.2
r=1.5
On applying the inverse complex Hankel transform to Fig. 4. Profiles of the velocity radial component Vr for the odd
(37)-(40), the solution to the problem (3)–(5) with boundary problem at R=1.
conditions (33), (34) is obtained in the form of convergent
improper integrals: B. Numerical Results for the Even Problem
Fig. 5 presents the results of calculation of Vr by means of
sinhk 2 sinhk1 z sinhk1 sinhk 2 z
Vr 1 (r )d , (41) formula (41). One can see that Vr differs from zero only near
0
2 the entrance region. Additionally, in Fig. 5 Vr is positive for
k2 sinhk2 cosh k1z k1 sinhk1 cosh k2 z some values of r, e.g., for 0 r 1 and Ha=10. However,
Vz J 0 (r )d (42)
2 since the fluid flows out through the hole at z 1 , the r-
0
component of the velocity must be negative for 0 r 1 at
P k sinhk1 sinhk 2 z k 2 sinhk 2 sinhk1 z
Ha 1 J 1 (r ) d Ha=0. It means that in the transverse magnetic field in the
r 0
2 region 0 z 1 there exists an opposite flow, which occurs due
(43) to vortices generated in the channel (see Fig. 6). The vector
ISBN: 978-1-61804-240-8 64
Advances in Applied and Pure Mathematics
field of the velocity for Ha=10 is shown in Fig. 6. One can see that as in the previous case, the profiles of the
velocity component Vr differ from the Hartmann flow profiles
Vr Ha=10 only near the entrance region. The magnitude of the velocity is
r=1.01
0.1 inversely proportional to the distance from the hole.
r=1.1 0.05
Vr
Ha=50
r=2
z r=1.01 0.35
-1 -0.5 0.5 1
r=1.5 0.3
-0.05
r=1.1 0.25
r=2 0.15
Vr 0.1
0.1 Ha=50
0.05
r=1.01 z
0.05 -1 -0.5 0.5 1
r=1.1
z Fig. 8. Profiles of Vr for the general problem at R=1 and Ha=50.
-1 -0.5 0.5 1
-0.05
V. REMARKS TO THE SOLUTION OF MHD PROBLEM ON AN
INFLOW OF CONDUCTING FLUID INTO A CIRCULAR CHANNEL
-0.1
THROUGH THE CHANNEL’S LATERAL SIDE
Fig. 5. Profiles of Vr for the even problem at R=1. A. Formulation of the Problem
A circular channel is located in the region
z D ={ 0 ~r R, 0 ~ 2 , ~z }. There is a split
1
in the channel lateral surface in the region
~ ~
{~r R, d ~z d }, through which a conducting fluid flows
0.5
~
into the channel with the constant velocity V V0 er (see
r Fig.9). The case of longitudinal magnetic field is considered,
0.8 1.2 1.4 1.6 1.8 2
i.e., when the external magnetic field B e B0 ez is parallel to
-0.5
the ~ z axis.
-1 ~
r
Fig. 6. Velocity field for the even problem at R=1 and Ha=10.
R ~
z
0
C. Numerical Results for the General Problem
The solution of the general problem is equal to the sum of ~ ~
d d
solutions to the odd and even problems with respect to z. ~
~
B B0 ez V V0 er
Results of calculation of Vr for the general problem at the r R
Hartmann numbers Ha=10 and Ha=50 are presented in Fig. 7
and Fig. 8. Fig. 9. The geometry of the flow in the circular channel.
Vr
Ha=10
It is supposed that walls ~ r R are nonconducting and
0.35
induced streams do not flow through the split ~
r=1.01
r=1.1
r R,
0.3 ~ ~ ~ ~
d z d in the region R r . In this problem
0.25
r=1.5 0.2 E 0 (see [3]). The dimensionless variables are introduced
r=2 0.15 similarly as it was done for the first problem, but the chanel
r=3 0.1
radius R is used as the length scale.
r=4 0.05
The problem is described by the system of equations
-1 -0.5 0.5 1
z (3)-(5) with the boundary conditions:
ISBN: 978-1-61804-240-8 65
Advances in Applied and Pure Mathematics
~
d d /R ,
where A const. The functions V (r ) and
2
V zs (r , ) F s V z (r , z ) (r , z ) sin z dz ,
new new
dP / dz are the velocity of the flow and the pressure gradient V z
0
in the channel at the sufficient distance from the entrance
region, and which satisfy the following equation (see [3], [4]):
P c (r , ) F c P new (r , z ) 2
P
new
(r , z ) cos z dz .
0
dP 1 d dV
(r ) 0 (47)
dz r d r dr It is also used that
V (r ) 0 . 2z 1
with the boundary condition: r 1: Fs F c . (56)
1 z 2 2
2
1 z
B. Solution of the Problem.
On solving the problem, the symmetry of the problem with As a result, the following system of ordinary differential
respect to z is used, i.e., the velocity component Vr (r , z ) and equations for the unknown functions Vrc , Vzs , P c is obtained:
pressure P(r , z ) are even functions with respect to z, but the
component Vz (r , z ) is the odd function with respect to z. It dPc
( L1r Ha2 )Vrc 0 , (57)
means that the functions Vr (r , z ) and Vz (r , z ) satisfy dr
2
additional boundary conditions: P c L0rVzs V (r ) F c ( ) 0 , (58)
z 0, 0 r 1 : Vz 0 , Vr / z 0. (48)
Vzs
1 d
r dr
2
rVrc V (r ) F c ( ) 0 ,
(59)
The problem can be solved by the Fourier cosine and
Fourier sine transforms, but since Vz and P / z do not d2 1 d 1
were L0r 2 , L1 r L0 r 2 ,
tends to zero at z , the new functions for the velocity dr 2
r dr r
and pressure gradient are to be introduced before using these 1
transforms: un F c ( ) F c 2
.
1 z
2
V new V arctan(z ) V (r ) e z (49)
The boundary conditions are
P new
P 2
arctan(z) A (50) 2 sin L
z z r=1: Vrc (r , ) Vz 0
s
, (60)
As a result, the problem has the form: s
On eliminating the functions V z , P c from system (57)-
P new
( L1 Ha2 )Vr 0 , (51) c
(59), the equation for Vr is obtained in the form
r
P new 2 2z d ~ ~
L0Vz V (r ) 0, ( Lr L0r )Vrc 22 L1rVrc 2 (2 Ha2 )Vrc 0.
new
(52)
1 z 2 2
(61)
z dr
V z new 1 2 1 ~ 1 d d2 1 d ~ 1
(rVr ) V (r ) 0. (53) where L0r , Lr 2 , L1r Lr 2 .
z r r 1 z2 r dr dr r dr r
Differential equation (61) completely coincides with
c
Boundary conditions are differential equation for Vr obtained in [2], therefore, the
ISBN: 978-1-61804-240-8 66
Advances in Applied and Pure Mathematics
~
L1r I1 (k r ) k 2 I1 (k r ) un Lr I 0 (k r ) k 2 I 0 (k r ) .
Then
dPc ~
c1 ( ) I 1 (k1 r ) c~2 ( ) I 1 (k 2 r ) , (66)
dr
2 F c ( ) A
P c iHa c2 k 2 I 0 (k 2 r ) c1k1 I 0 (k1r ) . (67)
c ( ) I (k r ) c ( ) I (k r )coszd ,
2
Vr (68)
0
1 1 1 2 1 2
sin z
c ( )k I (k1 r ) c 2 ( )k 2 I 0 (k 2 r )
2
Vz d , (69)
1 1 0
0
P 2 ~
c ( ) I1 (k1r ) c~2 ( ) I1 (k2 r )coszd ,
0 1
(70)
r
P
c ( )k 2 I 0 (k 2 r ) c1 ( )k1 I 0 (k1 r )sin z d ,
2
iHa
z
2
0
(71)
where c~1 ( ) Ha (i Ha) c1 ( ),
c~ ( ) Ha (i Ha) c ( )
2 2
REFERENCES
[1] Antimirov M.Ya., Kolyshkin A.A., Vaillancourt R. Applied Integral
Transforms.- Rhole Island USA: American Mathematical Society, 1993.
[2] Antimirov M., Ligere E. “Analytical solution for
magnetohydrodynamical problems at flow of conducting fluid in the
initial part of round and plane channels”, Magnetohydrodynamics.
vol.36, no.3.,pp. 241-250, 2000.
ISBN: 978-1-61804-240-8 67
Advances in Applied and Pure Mathematics
ISBN: 978-1-61804-240-8 68
Advances in Applied and Pure Mathematics
Definition 1.1. ([14]) Let be a set of the complex plane f A S with supp f F .
and let S be a compact subset. An algebra A S of - A subspace Y of X is said to be invariant
valued functions defined on is called S -normal if for any with respect to an A S -spectral function
finite open S -covering
n
G S G i i 1 of , there are the U : A S B X if U f Y Y , for any f A S .
We recall several important properties of an A -
functions, f S , f i A S 1 i n such that:
spectral function U ([8]), because we want to obtain similar
1) f S 0, 1 , f i 0, 1 properties for an A S -spectral function:
1 i n ; 1. U has the single-valued extension property,
2) supp f S G S , supp f i G i where is the identical function f ;
1 i n ; 2. U
U f x supp f , for any f A and
n
x X ;
3) f S f i 1 on
i 1
3.If U
x supp f ,then
where the support of f A S is defined as: U f x 0 , for any f A and x X ;
supp f ; f 0 . 4.
2) AS is S -normal; 6. U is decomposable.
3) for any f A S and any supp f , the
T B X be an A S -scalar operator
Theorem 1.4. Let
function
f and let U be an A S -spectral function for T . Then we
, for \ have:
f
0, supp U T S and T supp U S .
for
ISBN: 978-1-61804-240-8 69
Advances in Applied and Pure Mathematics
ISBN: 978-1-61804-240-8 70
Advances in Applied and Pure Mathematics
U U f x supp f U x , generated by U f Y , where Y X and A 0 is the set of all
therefore according to Proposition 1.1.2, [8], it follows that functions in AS with compact support, then we have:
U f x 0.
f A0
Uf X X .
The property (5) can be obtained by using (4), as in
the proof of Proposition 3.1.17, [8] and will be omitted.
The proof of (6) is presented in [14], Theorem 3.3. Definition 1.8. Let U be an A S -spectral function. For any
Lemma 1.6. Let U be an A S -spectral function. If G1 is an open set G G S we denote
b) For f A S having the property that f 1 on Theorem 1.9. Let U be an A S -spectral function. Then
a neighborhood of supp U , it results that U f I .
X U F XU F x X ; U x F , for
Proof. We have F S , F S .
supp 1 f 1 supp f 2 G 2
Proof. If U x F , for F S , with F S , let us
supp 1 f 1 supp U
hence
considerG G S an open set with G F S . Then by
0 U 1 f 1 U 1 U f 1 S -normality of A S there is a function f AS such that
therefore 1, for in a neighborhood of F
U f 1 U 1 I and U f 2 0 . f
0, for in a neighborhood of \ G
Moreover, for f A S with the property that and therefore supp f G , whence
f 1 on a neighborhood of supp U we have supp 1 f U x supp 1 f F .
Uf I
According to Theorem 1.5, U 1 f x 0 , hence
because it can be chosen in this case: f 1 f , f 2 g , with
x U f x X U G .
supp g supp U , hence Ug 0 and
G G S being an arbitrary open set with G F , F S ,
accordingly
we have
U f g U 1 I U f U g , whence U f I .
x X U F , i.e. XU F X U F .
Remark 1.7. From Lemma 1.6, it results that if f A S and Conversely, let us show that
X U F X U F , for any F S , F S .
f 1 in a neighborhood of supp U , then U f I . If
we denote by
U f Y the linear subspace of X
f A0
x X U F X U G , for any open set
Let
ISBN: 978-1-61804-240-8 71
Advances in Applied and Pure Mathematics
Corollary 1.10. If U is an A S -spectral function, then for The behaviour of A -scalar and A S -scalar
operators with respect to spectral equivalence is completely
any F S with F S , X U F is a maximal
different. IfT1 B X is A -scalar (respectively, A S -
spectral space for U . scalar) and T 2 B X is spectral equivalent to T1 , then
Proof. It results easily from the previous theorem.
T 2 is not A -scalar (respectively, A S -scalar), in general; in
Theorem 1.11. Let T1, T 2 B X . If T1 is S - this situation, we still know that T 2 is decomposable
decomposable (in particular, decomposable) and T1, T2 are (respectively, S -decomposable) and then T 2 is said to be
spectral equivalent, then T 2 is also S -decomposable (in A -decomposable (respectively, A S -decomposable). If in
particular, decomposable) and addition T commutes with one of its A -spectral
X T1 F X T 2 F , (respectively, A S -spectral) functions U , i.e.
for any F closed, F S (when S , for any T U f U f T , for any f A (respectively, for any
F closed). f A S ), then T is said to be A -spectral
If T1 and T 2 are decomposable, then T1 is spectral
(respectively, A S -spectral).
equivalent to T 2 if and only if their spectral spaces
X T1 F and X T2 F are equal, i.e. Definition 2.1. An operator T B X is called A S -
X T1 F X T 2 F , for any F closed ([8], 2.2.1, decomposable if there is an A S -spectral function U such
2.2.2). that T is spectral equivalent to U .
If T1 and T 2 are S -decomposable and spectral equivalent,
then their spectral spaces are equal, i.e.
ISBN: 978-1-61804-240-8 72
Advances in Applied and Pure Mathematics
Then the assertion (I) implies the assertion (II), and for case Theorem 2.5. For an operator T B X we consider the
S , the assertions (I) and (II) are equivalent.
following four assertions:
Proof. Let us suppose that there is an A S -spectral function
I T is A S -decomposable and commutes with one of its
U such that T and U are spectral equivalent. Since U
A S -spectral functions (i.e. T is A S -spectral);
is S -decomposable (Theorem 1.5), then, according to
Theorem 1.11, it results that T is S -decomposable and we II II1 T is S -decomposable;
have
II2 There is an A S -spectral function U
XT F XU F (1)
commuting with T , i.e. U f T T U f , for any
for any F closed, F S . But X T F is invariant f AS ;
II3 X T F X U F , for any F
to
T and T X T F F (Theorem 2.1.3, [6]),
whence it follows (by (1)) that closed, F S ;
T XU F XU F III III1 There is an A S -spectral function U
and commuting with T ;
T XU F F .
III2 T | X U F F , for any F
In case S , if the assertion (II) is fulfilled, according to closed, F S ;
Theorem 2.2.6, [8], we deduce that T is decomposable and IV T S Q , where S is an A S -scalar operator and
that the equality (1) holds for any closed set F . Then T Q is a quasinilpotent operator commuting with an A S -
is spectral equivalent to U (Theorem 2.2.2, [8]) and
spectral function of S (not to be confused the compact subset
therefore (I) is verified. S with the operator S from the equality T S Q , S
being the scalar part of T and Q the radical part of T ).
Remark 2.3. If T B X is A S -decomposable and U is
one of its A S -spectral functions, then:
Then the assertions I and IV , respectively II and
1) T is S -decomposable; III are equivalent, I implies II , respectively III ,
2) X T F X U F , for any F and finally IV implies II .
closed, F S ; Proof. I II , III . Assuming (I) fulfilled, we prove
3) If V is another A S -spectral function of T, that the assertions (II) and (III) are verified. If T is A S -
then U and V are spectral equivalent (in particular, V decomposable and commutes with one of its A S -spectral
is spectral equivalent to T );
functions U , then U is spectral-equivalent to T .
ISBN: 978-1-61804-240-8 73
Advances in Applied and Pure Mathematics
ISBN: 978-1-61804-240-8 74
Advances in Applied and Pure Mathematics
DECOMPOSABLE AND
A S -SPECTRAL OPERATORS.
n
n
In this section we study the behaviour of these three 1 n k k
k A1 A2 B1 B 2
nk
classes of operators with respect to direct sums, to restrictions
k 0
and quotients with regard to an invariant subspace and to n
n
continuous algebraic homomorphisms, respectively.
1 n k k
k
A1 A2
k
B1nk B 2nk
Lemma 3.1. k 0
1 Let Ai , B i B X i such that Ai is quasinilpotent n
n
equivalent to B i , i 1, 2 . Then A1 A 2 is quasinilpotent
1 n k
k nk
A1 B1
k
A2k B 2n k
k 0
equivalent to B1 B 2 ;
ISBN: 978-1-61804-240-8 75
Advances in Applied and Pure Mathematics
f 1 ).
k 0
n The mapping
nk n k nk
1 A Y
k
B Y U 1
U 2 : AS B X1 X 2 , f U 1f U 2f ,
k 0 is evidently an A S -spectral function for T1 T2 , because
n
A Y B Y we have:
n hence T1 T2 is A S -scalar.
n k nk
h A B h
n
1 nk
A B
k
If T1 and T 2 are A S -decomposable, then there are
k 0 two A S -spectral functions U 1 and U 2 such that T1 is
n
n U 1 , respectively T 2 is spectral
1 n k
k
k
h A h B
nk
spectral equivalent to
ISBN: 978-1-61804-240-8 76
Advances in Applied and Pure Mathematics
ISBN: 978-1-61804-240-8 77
Advances in Applied and Pure Mathematics
IV. CONCLUSIONS
REFERENCES
[1] ALBRECHT, E.J., ESCHMEIER, J., Analytic functional models and
local spectral theory, Proc. London Math. Soc., 75, 323-348, 1997.
[2] APOSTOL, C., Spectral decompositions and functional calculus, Rev.
Roum. Math. Putes et Appl., 13, 1481-1528 1968.
[3] BACALU, I., On restrictions and quotients of decomposable operators,
Rev. Roum. Math. Pures et Appl., 18, 809-813, 1973.
[4] BACALU, I., S decomposable operators in Banach spaces, Rev.
Roum. Math. Pures et Appl., 20, 1101-1107, 1975.
[5] BACALU, I., Some properties of decomposable operators, Rev. Roum.
Math. Pures et Appl., 21, 177-194, 1976.
[6] BACALU, I., Descompuneri spectrale reziduale (Residually spectral
decompositions), St. Cerc. Mat. I (1980), II (1980), III (1981).
[7] BACALU, I., S Spectral Decompositions, Ed. Politehnica Press,
Bucharest, 2008.
[8] COLOJOARĂ, I., FOIAŞ, C., Theory of generalized spectral operators,
Gordon Breach, Science Publ., New York-London-Paris, 1968.
[9] DOWSON, H.R., Restrictions of spectral operators, Proc. London
Math. Soc., 15, 437-457, 1965.
[10] DUNFORD, N., SCHWARTZ, J.T., Linear operators, Interscience
Publishers, New York, I (1958), II (1963), III (1971).
[11] LAURSEN, K.B., NEUMANN, M.M., An Introduction to Local
Spectral Theory, London Math. Soc. Monographs New Series, Oxford
Univ. Press., New-York, 2000.
[12] VASILESCU, F.H., Residually decomposable operators in Banach
spaces, Tôhoku Math. Journ., 21, 509-522, 1969.
[13] VASILESCU F.H., Analytic Functional Calculus and Spectral
Decompositions, D. Reidel Publishing Co., Dordrecht, Editura
Academiei, Bucharest, 1982.
[14] ZAMFIR, M., BACALU, I., A S -scalar operators, U.P.B. Sci.
Bull., Series A, 74, 89-98, 2012.
ISBN: 978-1-61804-240-8 78
Advances in Applied and Pure Mathematics
ISBN: 978-1-61804-240-8 79
Advances in Applied and Pure Mathematics
2 MATHEMATICAL MODELLING
where
2.1 Governing equations γ − b
λ (γ ) = µ ∞ + ∆µ exp − 1 + exp
a γ ,
We assumed the blood flow to be laminar and
incompressible and therefore the governing Navier-Stokes
γ − d
equations for such flows are given by
n(γ ) = n∞ − ∆n exp − 1 + exp
c γ
∇ ⋅V = 0 (1)
∂V
ρ + V . ∇V = −∇. τ − ∇p and .
∂t (2)
These models have been created by various researchers
Walburn and Schneck [7], Cho and Kensey [8], Fung [9],
where V is the three-dimensional velocity vector,
p pressure, ρ density and τ the shear stress term.
Ballyk et al. [10]and Johnston et al. [11], by fitting the input
parameters in these models to experimental data for blood
We considered four different non-Newtonian blood flow viscosity measured at certain shear rates.
models and compared the results obtained with that from the
simple Newtonian model in this study. The effects of these 2.2 Geometry
models on the flow field and the wall shear stress in the
vicinity of the aneurysm are examined. These models are given The diameters of the left and right common carotid artery
below [6]. show a significant amount of variation. A study of 17 healthy
subjects [12] produced diameters in the range of 0.52 to 0.75
cm with an average diameter of 0.64 cm. In this study, two
Blood Models different model of the abdominal aortic aneurysm are used to
investigate blood flow in the initial stages of its development.
1. Newtonian model The carotid artery before and after the dilation is idealized as a
straight rigid tube without any branching arteries.
µ = 0.00345 Pa ⋅ s (3) The flow geometry then consists of straight rigid tube of
diameter d and is divided into three regions, the inlet, the
2. Power Law Model dilation and the outlet region. The lengths of these regions are
4d, 4d and 18d, respectively. The radius of the undeformed
µ = µ 0 (γ ) n −1 (4)
inlet and outlet is . Two different values of the
diameter were used to model the carotid artery, namely,
where =0.01467 and . cm and cm.
The radius of the diseased region [12] is given by
3. Casson Model
a 2 + (b 2)2 a 2 + (b 2)2
r = r0 + a − + − (b 2 − x )2 0 ≤ x ≤ b (8)
[ τ y + η γ ] 2 2a 2a
µ=
γ
(5)
where x is the distance from the start of the aneurysm, a is the
where and =
degree of dilation, and b is the overall length of the diseased
with = 0.0012 Pa·s and . area as shown in Figure 1. Each model had an aspect ratio
which is typical of fusiform aneurysms.
4. Carreau Model
[
µ = µ ∞ + (µ 0 − µ ∞ )1 + (λγ )2 ]
( n −1) / 2
(6)
where Pa·s, Pa·s,
s and .
Fig. 1. Aneurysm Geometry
5. Generalized Power Law Model
Three different degrees of dilation, 25%, 40% and 55% were
n −1 used in this study.
µ = λ γ
(7)
ISBN: 978-1-61804-240-8 80
Advances in Applied and Pure Mathematics
ISBN: 978-1-61804-240-8 81
Advances in Applied and Pure Mathematics
ISBN: 978-1-61804-240-8 82
Advances in Applied and Pure Mathematics
ISBN: 978-1-61804-240-8 83
Advances in Applied and Pure Mathematics
Model
4. CONCLUSIONS Careau
References
ISBN: 978-1-61804-240-8 84
Advances in Applied and Pure Mathematics
ISBN: 978-1-61804-240-8 85
Advances in Applied and Pure Mathematics
have received considerable attention in the literature [1], [2], II. P ROBLEM DESCRIPTION
[3], [4], [5], [6], [7]. This has been motivated by an application A permutation of the n-tuple x0 = [0, 1, ..., n − 1] ∈ Nn is a
to powerline communications when M-ary Frequency-Shift codeword of length n and the set of all codewords of length
Keying (FSK) modulation is used [3], [8], [9], [10], [11]. In n is denoted by Ωn . From an algebraic point of view the set
this application permutations are used to ensure that power Ωn is the single orbit of the symmetric group of permutations
output remains as constant as possible while combatting Sn , i.e.
impulsive noise permanent, narrow band noise from electrical Ωn = {x ∈ Nn |x = gx0 , g ∈ Sn }
equipment or magnetic fields, as well as the more common
Any subset Γ of Ωn is a permutation code of length n. The
white Gaussian noise.
main problem can now be stated as:
A permutation code is a set of permutations in the symmet-
ric group Sn of all permutations on n elements. The codewords Definition 1. Given a code length n and a Hamming distance
are the permutations and the code length is n. The error- d, the maximum permutation code problem (MPCP) consists
correcting capability of a permutation code is related to the of the determination of a code Γ ⊆ Ωn with minimum distance
minimum Hamming distance of the code. The Hamming dis- d and the maximum possible number of codewords.
tance δ between two codewords is the number of elements that Example 1. The problem (6, 5) is to determine a maximal
differ in the two permutations. Alternatively, two permutations code of length n = 6 with minimum distance d = 5. As reported
σ1 and σ2 are at distance δ if σ1 σ2−1 has exactly n − δ fixed in [14], [3], [12] the optimal solution of this problem is
points. The minimum distance d is then the minimum δ taken M(6, 5) = 18. One of the many possible optimal (6, 5) codes
over all pairs of distinct permutations. Such a code is then is shown in Figure 1.
denoted an (n, d) permutation code.
Redundancy in an encoding is minimized if the number of III. A BRANCH AND BOUND ALGORITHM
codewords is as large as possible. Thus if M(n, d) denotes
the maximum number of codewords in an (n, d) permutation Branch and bound approaches work by building a search-
code it is important to determine M(n, d), or if this is not tree that covers all possible assignments of permutations to
possible to find good lower and upper bounds. The most solutions, but with most of the branches of the tree pruned by
complete contributions to lower bounds can be found is in [3], inferring information from lower and upper bounds. The main
[12]. Recently, some improvements based on similar search elements of the algorithm proposed here are described in this
techniques have been presented in [13], while in [14] a study section.
R. Montemanni and J. Barta are with the Dalle Molle Institute for A. Structure of the search-tree node
Artificial Intelligence, University of Applied Sciences of Southern Switzer-
land, Galleria 2, 6928 Manno, Switzerland. Emails: {roberto.montemanni, From now on, the set of nodes of the search tree will be
janos.barta}@supsi.ch. denoted S, and the subtree of the search-tree rooted at node t
D.H. Smith is with the Division of Mathematics and Statistics, Uni-
versity of South Wales, Pontypridd, CF37 1DL, Wales, UK. Email: will be denoted as SubT (t). Each node t of the search-tree is
derek.smith@southwales.ac.uk. then identified by the following elements:
ISBN: 978-1-61804-240-8 86
Advances in Applied and Pure Mathematics
• in(t): a list of permutations that are forced in the solutions The branch and bound algorithm stops when the set S
associated with the search-tree nodes of SubT (t); is empty (all the search-tree nodes have been processed or
• f eas(t): a list of permutations that are feasible according labelled as dominated).
to the list of forced permutations in(t), and to reduction
rules (see Section III-D2); C. Selection of the node to expand
• lb(t): a lower bound for the number of permutations in the
Nodes are expanded in the same order they have previously
optimal solutions associated with the search-tree nodes in
been created. This strategy allows the best exploitation of the
SubT (t). The calculation of the lower bound is described
reduction and pruning rules described in Section III-D.
in Section III-F.
• ub(t): an upper bound for the number of permutations
in the optimal solutions associated with the search-tree D. Reduction and pruning rules
nodes in SubT (t). The calculation of the upper bound is This section discusses some rules useful to prune dominated
described in Section III-E. search-tree nodes and to reduce the size of f eas(t) while
From these four items all the information required by domi- generating a new search-tree node t. These results are based
nance rules and pruning can be derived. on the concept of isomorphism for graphs [16].
1) Pruning rule: Two graphs G and H are said to be
isomorphic if a bijection between the vertex sets of G and
B. Initialization and branching strategy H f : V (G) → V (H) exists, such that any two vertices u and
v of G are adjacent in G if and only if f (u) and f (v) are
Initial lower and upper bounds BestLB and BestUB are adjacent in H.
provided as input to the algorithm (they can be 0 and +∞,
respectively). These initial values will be updated during the Definition 2. The graph induced by search-tree node t is
execution of the algorithm in case improved values are ob- defined as GtI = {VtI , EtI }, with VtI = V \in(t) and EtI =
{i, j}|i, j ∈ VtI , δ (i, j) ≥ d) .
tained. A permutation p is selected and the root r of the search-
tree is the node initialized with in(r) = {p}, lb(r) = BestLB, Remark 1. The graph induced by search-tree node t is
ub(r) = BestUB and f eas(r) = {i ∈ Ωn : δ (i, p) ≥ d}. Initially, considered instead of the subgraph of G with vertices set
r will be the only node contained in S (S := {r}), the set of f eas(t) because f eas(t) might have already have benefited
the nodes to be examined, referred to as open nodes in the from reduction rules in previous iterations, and therefore
remainder of the paper. Due to the symmetry of the problem, isomorphisms could be more difficult to identify.
the first permutation included in in(r) can be chosen arbitrarily.
The set of closed nodes C is initialized as empty (C := 0). / This Definition 3. If the graph GtI induced by search-tree node t
set will contain nodes already examined by the algorithm, and is isomorphic to the graph GIu induced by another search-tree
will be used by pruning and reduction techniques described in node u, it will be said (for short) that node t is isomorphic to
Section III-D. node u and written t ∼= u.
At each iteration of the branch and bound algorithm, an The following result allows one of two isomorphic nodes
open node t from the set S is expanded (see Section III-C to be pruned from the branch and bound tree.
for more details about the strategy used to select node t, and
the rationale behind it), which means that node t is expanded Theorem 1. If a new search-tree node t is such that t ∼= u with
by decomposing it into the associated subproblems. One new u ∈ S ∪C, |in(t)| = |in(u)| then the node t can be classified as
search-tree node t p is created for each permutation p of f eas(t) dominated and moved to set C (S = S\{t} and C = C ∪ {t}).
in such a way that in(t p ) = in(t)∪{p} and the new set f eas(t p ) Proof: The search-tree subtree associated with node t
is determined, also taking into account the reduction and will provide an optimal solution with the same number of
pruning rules described in Section III-D. Sets S and C are permutations of that of node u, that has been already expanded
finally updated: S = S\{t}, C = C ∪ {t}. For each new node (u ∈ C), or is scheduled to be expanded (u ∈ S).
t p the values of lb(t p ) and ub(t p ) are calculated, as described 2) Reduction rule: During the branching of a search-
in Sections III-F and III-E respectively. In the case that the tree node t, all potential new search-tree nodes obtained by
pruning test is positive (see Section III-D) the new node t p is expanding the set f eas(t) with each possible permutation will
pruned and C := C ∪ {t p }, otherwise the set S of open node is be considered.
incremented: S := S ∪ {t p }.
Proposition 1. While creating a new search-tree node u
In case mint∈S ub(t) ≤ BestUB, the global upper bound
obtained by adding pu ∈ f eas(t) into in(u), permutation
of the residual open problems has been improved, and the
pk ∈ f eas(u), with k ∼= v, v ∈ S ∪ C, |in(k)| = |in(v)| (nodes
updating BestUB := mint∈S ub(t) can take place. Also the
already expanded at the same level) can be taken out of
value of BestLB can be updated in case a new incumbent
f eas(u): f eas(u) = f eas(u)\{pk }
heuristic solution is found (in general BestLB := maxt∈S lb(t)).
All the open nodes u of the search tree are examined and Proof: The best possible solution including permutations
pruned in case ub(u) ≤ BestLB, since no improving solution pk and the permutations of in(u) is equivalent to that of
can exist in the search-tree node rooted in u. In such a case the problem v, that has been already expanded (u ∈ C), or
S := S\{u} and S := S ∪ {u}. is scheduled to be expanded (u ∈ S). Therefore all solutions
ISBN: 978-1-61804-240-8 87
Advances in Applied and Pure Mathematics
including permutation pk are not of interest while solving the Combining the n partitions together, gives the global upper
problem associated with f eas(u). bound provided by (2).
Remark 2. When applying the reduction rule described in Remark 3. The result of Proposition 2 can be seen as a
Proposition 1 it is necessary to expand search-tree nodes refinement of a previous result originally presented in [14].
according to the basic strategy described in Section III-C, in
order to avoid situations where the permutations isomorphic to Remark 4. As the index k of the fixed component in the
that associated with node ui are taken out from f eas(vi ) as a codewords can be varied, there are n different partitions that
result of Proposition 1 at level i of the search-tree, but then the can be generated. The algorithm computes for each partition
permutations associated with a descendant v j of vi are taken an upper bound and finally chooses the lowest one.
out of a descendant u j of ui at a level j, with j > i, of the
tree. Such a situation would clearly lead to infeasible solutions F. Lower bounds
since some regions of the search space are left unvisited by
the search-tree. The methods used to provide lower bounds for the optimal
solution cost of the problem associated with the search-tree
In the implementation described here the routines of Nauty node t are based on MCP algorithms [21]. In detail, when
2.5 described in [17] are used to identify graph isomorphisms. examining node t, it is possible to associate a graph Gt =
{Vt , Et } such that the set of vertices Vt = f eas(t) (where each
E. Upper bound vertex is associated with a permutation) and the set of edges
Et = {{i, j}|i, j ∈ Vt , δ (i, j) ≥ d)}. Notice that this graph is a
The set of codewords Ωn can be split into n subsets
subgraph of that induced by node t (see Section III-D). The
W0 , ...,Wn−1 , in such a way that for a fixed value k ∈ {0, ..., n −
problem is then equivalent to solving a MCP on graph Gt .
1} the subset Wi is defined as Wi = {x ∈ Ωn |x(k) = i}. In
This transformation usually has the side-effect that possible
other words, the subset Wi contains all codewords with the k-th
structure and information coming from group theory (and the
component having the value i. Since the partition is obtained
possibility to exploit these) is lost. On the other hand, the
by fixing the value of one component of the codewords, it is
very efficient methods developed for the MCP over the years
clear that the sets Wi are isomorphic to Ωn−1 . Furthermore,
can be used directly. In the remainder of this section it will be
as the sets Wi form a partition of Ωn it is well-known that an
shown how to modify a general-purpose method to the current
upper bound of M(n, d) can be obtained by adding the upper
context, in order to insert permutation codes related concepts
bounds on the subsets Wi :
into its logic.
Theorem 2 (Deza and Vanstone [18]). The original MCP method works as follows, according to
the description provided in [21]. The algorithm considers the
M(n, d) ≤ n · M(n − 1, d) (1)
vertices of Vt in a given order {v1 , v2 , ..., v|Vt | }. Initially, the
The partitioning procedure described in Theorem 2 can be method finds the largest clique C1 that contains the vertex v1 .
carried out on any subset of Ωn . At each search-tree node Then it finds C2 , the largest clique in Gt \{v1 } that contains
t the algorithm generates a partition T0 , ..., Tn−1 of the set v2 and so on. Applying heuristics and pruning techniques,
f eas(t), such that Ti = {x ∈ f eas(t)|x(k) = i} and a partition the search space can, however, be reduced dramatically. The
Q0 , ..., Qn−1 of the set in(t), such that Qi = {x ∈ in(t)|x(k) = notion of the depth is crucial for the algorithm. Suppose
i}. For each subset Ti an upper bound UB(Ti ) is calculated vertex v1 is under investigation. At depth 2, all vertices
using the Maximum Clique Problem (MCP) solver proposed adjacent to v1 are considered. At depth 3, all vertices (that are
in [19] (see also [20]), which is run for 10 seconds on every already in depth 2) adjacent to the first vertex in depth 2 are
subproblem. The choice of this solver is motivated by its considered and so on. When there are no more vertices left at
ability of proving optimality extremely fast on small/medium a certain depth, a clique has been identified, and backtracking
problems. From preliminary experiments the solver described is activated. Pruning rules based on the number of vertices left
in [19] is significantly faster than the one presented in [21] at each level are triggered to have early backtracking. These
(see Section III-F) on the instances treated. rules are very quick to check but very effective. When the
The new upper bound for the search tree node t can be entire search-tree created by the method has been visited (or
expressed as follows. declared dominated) the computation is stopped and the largest
clique has been retrieved.
Proposition 2. Some context-dependent modifications of the general algo-
n−1 rithm previously described were implemented. The modifica-
∑ |Qi | + min{UB(Ti ); M(n − 1, d) − |Qi |} (2) tions to the original method are introduced to anticipate the
i=0 pruning in the branch and bound framework running internally
is a valid upper bound for the search-subtree rooted at the in the MCP algorithm. In detail, during the execution a 2-
search-tree node t. dimension array PosVal is kept in memory, defined as follows.
Proof: In case it can be shown that in a partition {Ti } Definition 4. PosVal[i][ j] contains at any time during the
of f eas(t) the maximum clique has size smaller than M(n − execution of the MCP algorithm [21] the current number of
1, d) − |Qi | a tighter upper bound for that partition is available. permutations with value j in position i that are still feasible
ISBN: 978-1-61804-240-8 88
Advances in Applied and Pure Mathematics
according to the partial assignment under investigation, or are The modified version of the algorithm originally proposed
part of such a partial assignment. in [21] described previously is executed each time a new non-
dominated search-tree node (of the external branch and bound
Notice that the structure is dynamically updated during the
method) is generated, and it is run for a maximum of 720
execution of the algorithm, depending on the partial solution
seconds.
currently under investigation. The pruning is based on the
following theoretical results. IV. E XPERIMENTAL RESULTS
Proposition 3. When The approach discussed in Section III can be applied to
n subproblems to prevent ths size of the search tree increasing
∑ min{M(n − 1, d), PosVal[i][ j]} ≤ BestLB (3) too much for large instances. Results on subproblems can
j=1
then be propagated to full problems in order to obtain new
for some i, 1 ≤ i ≤ n the partial solution under investigation theoretical results, as will be shown later in this section. All
will not lead to any improving result, it can be pruned and it the experiments presented have been carried out on a computer
is possible to backtrack. equipped with a 2.3GHz AMD Quad Opteron processor (only
one core has been used at a time), while the algorithm has
Proof: It is known ([18]) that if a position i of the code-
been coded in ANSI C.
words of a code (n, d) is fixed to a value j then there cannot
In the study presented here two subproblems of (7, 5) are
be more that M(n − 1, d) permutations with this property.
considered, where a position of the permutations is restricted
This is valid for each value of i and j. During the execution
to values from a given set F ⊂ {0, 1, ..., n − 1}. The largest
of the MCP algorithm PosVal[i][ j] can become smaller than
possible code fulfilling the requirements with all permutations
M(n−1, d) for some i or j. In such a case the best upper bound
with the given position strictly from F is sought. Notice that
for the corresponding subproblem is no longer M(n − 1, d),
either the position fixed or the values of F are not important
and the global upper bound is updated according to (3) by
due to the symmetry of the problem. The cardinality of F is the
summing over all possible values j for position i. When
only important factor. Therefore subproblems can be defined
the global upper bound is not greater than the cost of the
as follows.
best solution currently available (BestLB) there is no hope of
finding an incumbent improved solution, and backtracking can Definition 5. Refer to the problem (n, d)||F| as the subproblem
be activated. of (n, d) where a position of the code is restricted to values
from set F.
Proposition 4. When
n Such a problem with |F| = 1 is equivalent to the problem
∑ min{M(n − 1, d), PosVal[i][ j]} ≤ BestLB (6, 5) since the largest possible set with a common symbol
i=1 in the first position is sought. According to [18] this amounts
for some j, 1 ≤ j ≤ n the partial solution under investigation to looking for the largest possible code of length n − 1. More
will not lead to any improving result, it can be pruned and it interesting are the cases when 2 ≤ |F| ≤ n − 1. Notice that
is possible to backtrack. for these problems a trivial upper bound, coming from the
generalization of that described in (1), is the following one:
Proof: The proof is based on the same principles of that
of Proposition 3 but now the sum is over all possible positions M(n, d)||F| ≤ |F| · M(n − 1, d) (5)
i for value j (by columns of PosVal instead of by rows). For |F| = 2 equation (5) returns an upper bound of 36. The
Remark 5. The results of Propositions 3 and 4 can be seen branch and bound described in Section III has been able to
as the adaptation of the upper bound of Proposition 2 to the retrieve a solution with 36 permutations, matching the upper
context of the MCP algorithm, where it is preferable to have bound. This result is not as obvious as it could appear, since
a less precise but much faster upper bound. traditional methods based on maximum clique solvers (e.g.
[19], [21]) are not able to retrieve such a solution in a week
In the context of the current permutation codes algorithm,
of computation, while the method described here was able to
the main target is to have the MCP algorithm complete the
close the problem in 15 382 seconds. The optimal solution
computation in order to prove optimality for the problem under
retrieved is presented in Figure 2. It might turn out to be
investigation in the given time available (see Section IV). For
useful for future studies by other researcher, since inspecting
this reason, to have pruning as early as possible, at each
such a solution might bring new insights about the general
iteration one of the permutations with value j in position i
characteristics of solutions.
is expanded such that For |F| = 3 the branch and bound described in Section III
PosVal[i][ j] = min PosVal[k][l] (4) has been used to obtain a new upper bound, which improves
1≤k≤n,1≤l≤n the one given by (5). The algorithm itself is not able to find
The strategy described in (4) allows the anticipation of the significantly large heuristic solutions (and consequently lower
application of Propositions 3 and 4 as much as possible. bounds), but if it is run with a hypothetical initial lower bound
Different strategies might however be used in order to make of BestLB = 53, the algorithm is able to prove in 922 450
it more likely to have heuristic solutions instead of early seconds that no solution with 54 permutations exists, leading
backtracking, if the algorithm is used in a different perspective. to the following new result.
ISBN: 978-1-61804-240-8 89
Advances in Applied and Pure Mathematics
Unary operators
József Dombi
(Invited Paper)
Abstract—Modal operators play an important role in fuzzy (ν, ν0 ) can be specified such that the curve must intersect. This
theory, and in recent years researchers have devoted more effort tells us something about how strong the negation operator is.
on this topic. In our study, we will construct modal operators.
We give a common form of all types of unary operators. We η(ν) = ν0 (2)
will characterise them using differential equations, Bayesian
inference, and a functional equation. We will show that a special If η(x) has a fix point ν∗ , we use the notation ην∗ (x) and
class of the kappa function is related to the sigmoid function, it if the decision value is ν, then we use the notation ην (x). If
and can be characterised by odds. η(x) is employed without a suffix, then the parameter has no
Keywords—negation, modalities, hedges, sharpness operator, Pli- importance in the proofs. Later on we will characterise the
ant system negation operator in terms of the ν∗ , ν0 and ν parameters.
ISBN: 978-1-61804-240-8 91
Advances in Applied and Pure Mathematics
Theorem 4: The negation operators and we make use of two types of this operator; one that is
strict, and one that is less strict. We will show that with these
f (ν)
ην (x) = f −1 f (ν0 ) , two negation operators we can define the modal hedges.
f (x)
2
f (ν∗ ) Modal logic, which is an area of mathematical logic,
ην∗ (x) = f −1 can be viewed as a logical system obtained by adding logical
f (x)
symbols and inference rules.
have the following properties:
a. They are continuous.
b. They are strictly monotonous decreasing. We will construct linguistic modal hedges called necessity
c. The correspondence principle is valid: and possibility hedges. The construction is based on the fact
that modal operators can be realised by combining two kinds
of negation operators.
ην (0) = 1, ην (1) = 0, ην∗ (0) = 1, ην∗ (1) = 0
d) The involutive property is valid: In intuitionistic logic, another kind of negation operator also
has to be taken into account. Here ∼x means the negated
ην (ην (x)) = x, ην∗ (ην∗ (x)) = x. value of x. ∼1 x and ∼2 x are two negation operators.
e) The neutral value property holds:
In modal logic, ∼1 x means "x" is impossible. In other
ην (ν) = ν0 , ην∗ (ν∗ ) = ν∗ . words, ∼1 is a stronger negation than not "x", i.e. ∼2 x.
Because ∼1 x in modal logic, it means "x is impossible".
Proof: Using the representation form of the negation We can write
operator, this is trivial.
impossible x = necessity(not x)
In fuzzy theory, we utilise two types of negation operator.
These are √
Yager: ηm (x) = m 1 − xm (6) ∼1 x := impossible x
∼2 x := not x
and
1−x
Hamacher, Sugeno: ηa (x) = (7)
1 + ax ∼1 x = ∼2 x
We can express the parameters of the negation operator in We will show that both operators belong to the same class
terms of its neutral values n(ν∗ ) = ν∗ . So we have of unary operators, and also show that because they have
p ln(2) a common form in the Pliant system, we will denote both
ν∗ = η(ν∗ ) = m
1 − ν∗m and m = − of them by τν (x). Depending on the ν value, we get the
ln(ν∗ ) necessity hedge or the possibility hedge.
Then the Yager negation operator has the form
− lnln2
ν∗
As we mentioned above, in modal logic we have two more
ln 2
ην∗ (x) = 1 − x− lnν∗ (8) operators than in the classical logic case, namely necessity and
possibility; and in modal logic there are two basic identities.
In a similar way, for the Hamacher negation operator, These are
1 1 ∼1 x = impossible(x) = necessity(not(x)) = ∼2 x
ην (x) = 1−ν0 1−ν x , ην∗ (x) = (10)
1+ ν0 ν 1−x 1+ ( 1−ν∗ 2 x
ν∗ ) 1−x
(9) ♦x = possible(x) = not(impossible(x)) =∼2 (∼1 x) (11)
This form of the negation operator can be found in [6]. In our context, we model impossible(x) with a stricter
Definition 5: A negation ην1 (x) is stricter than ην2 (x), if negation operator than not(x). Eq.(11) serves as a definition
ν1 < ν2 . of the possibility operator.
III. M ODALITIES INDUCED BY TWO NEGATION If in Eq.(10) we replace x by ∼2 x and using the fact that
OPERATORS ∼2 x is involutive, we get
The linguistic hedge “very” always expresses a tight
interval, whereas ”more or less” expresses a looser interval x =∼1 (∼2 x),
(less tight). In this sense, "very" corresponds to the necessity and with Eq.(11), we have
operator and "‘more or less"’ the possibility operator.
With this starting point, the necessity and possibility operators ♦x =∼2 (∼1 x).
used in fuzzy logic are based on an extension of modal logic
to the continuous case. We begin with the negation operator Definition 6: The general form of the modal operators is
ISBN: 978-1-61804-240-8 92
Advances in Applied and Pure Mathematics
1
where P (H) = x, ν = P (E|H) and τν (x) is modal
1+ P (E|H̄)
τν1 ,ν2 (x) = ην1 (ην2 (x)) , operator in the Pliant system. The ratio of two likelihood
where ν1 and ν2 are neutral values. If ν1 < ν2 , then τν1 ,ν2 (x) functions is called the likelihood ratio. So
is a necessity operator and if ν2 < ν1 , then τν1 ,ν2 (x) is a P (E|H)
possibility operator. ∧E =
From the above definition, we get P (E|H̄)
and we get
−1 f (x)
τν1 ,ν2 (x) = f f (ν1 )
f (ν2 ) , 1 1
ν= or P (H|E) = 1−P (H)
1 + ∧E 1+ 1
This can be rewritten as ∧E P (H)
ISBN: 978-1-61804-240-8 93
Advances in Applied and Pure Mathematics
ISBN: 978-1-61804-240-8 94
Advances in Applied and Pure Mathematics
ISBN: 978-1-61804-240-8 95
Advances in Applied and Pure Mathematics
1
Then (24) and (25) have the same form, i.e. C1 = C2 . IX. C ONCLUSIONS
In this study, we provided a general and theoretical basis
Proof: See [19]. for modalities. Here, we defined necessity and possibility
Theorem 15: If (22) (or (23)) is true, then κ(x) has the form operators, then we defined them via using a generator function
1 of the Pliant operators. We gave a theoretical basis for hedges
κ(x) = 1−x
and the main result is a unified formula for unary operators.
1+C x
The subroutine can be downloaded from the following website:
Proof: Using (19), (20) and (24), we get http://www.inf.u-szeged.hu/~dombi/.
λ
κ(x) 1 x
= . (26) ACKNOWLEDGEMENTS
1 − κ(x) C 1−x
This work was partially supported by the European Union
and the European Social Fund through project FuturICT.hu
(grant no.: TÁMOP-4.2.2.C-11/1/KONV-2012-0013).
A. Extension of κλν (x) to [a, b] interval
We can extend K on the [a, b] interval. K : [a, b] → [0, 1] R EFERENCES
Here, κ(x) is defined on [0, 1]. We can extend it to the [a, b]
interval by applying a linear transformation. That is, [1] P. Cintula, E. Klement, R. Mesiar, and M. Navara, “Fuzzy logics with
an additional involutive negation,” Fuzzy Sets and Systems, vol. 161,
x−a pp. 390–411, 2010.
x := [2] P. Hájek, Metamathematics of fuzzy logic. Kluwer Academic Publish-
b−a ers, Dordrecht, 1998.
and [3] S. Gottwald and P. Hájek, “Triangular norm-based mathematical fuzzy
logics,” Logical Algebraic, Analytic and Probabilistic Aspects of Trian-
xν − a gular Norms, pp. 275–299, 2005.
ν :=
b−a [4] E. Trillas, “Sobre functiones de negacion en la teoria de conjuntas
difusos,” Stochastica, vol. 3, pp. 47–60, 1979.
[5] J. Dombi., “De Morgan systems with an infinitely many negations in
Then we get the strict monotone operator case,” Information Sciences, vol. 181, pp.
1440–1453, 2011.
(λ) 1 [6] ——, “Towards a general class of operators for fuzzy systems,” IEEE
Ka,b (x) = λ Transactions on Fuzzy Systems, vol. 16, pp. 477–484, 2008.
1−ν0 xν − a b − x
1+ ν0 [7] L. A. Zadeh, “A fuzzy-set - theoretic interpretation of linguistic hedges.”
b − xν x − a Journal of Cybernetics, vol. 2(3), pp. 4–34, 1972.
(λ) [8] ——, “Fuzzy logic = computing with words,” IEEE Transactions on
Ka,b (x) has the following properties: Fuzzy Systems, vol. 4, pp. 103–111, 1996.
[9] ——, “From computing with numbers to computing with words-From
(λ) manipulation of measurements to manipulation of perceptions.” IEEE
Ka,b (xν ) = ν0 , Transactions on Circuits and Systems-I: Fundamental Theory and
(λ) Applications, vol. 45(1), pp. 105–119, 1999.
Ka,b (a) = 0 and Ka,b (b) = 1
[10] M. D. Cock and E. E. Kerre., “Fuzzy modifiers based on fuzzy
Ka,b has the same form as that in [20], where we showed relations.” Information Sciences, vol. 160, pp. 173–199, 2004.
that this is the general form for a certain class of membership [11] V. N. Huynh, T. B. Ho, and Y. Nakamori., “A parametric representation
function. of linguistic hedges in Zadeh’s fuzzy logic.” International Journal of
Approximate Reasoning, vol. 30, pp. 203–223, 2002.
[12] S. H. Rubin, “Computing with words.” IEEE Systems, Man and Cyber-
We can extend K such that K ∗ : [a, b] → [A, B]. This netics, Part B, vol. 29(4), pp. 518–524, 1999.
K ∗ (x) function can be written in the following implicit form: [13] I. B. Turksen., “A foundation for CWW: Meta-linguistic axioms.” IEEE
Fuzzy Information, vol. Processing NAFIPS’04, pp. 395–400, 2004.
λ λ
K ∗ (x) − K ∗ (a) K ∗ (b) − K ∗ (xν ) [14] L. A. Zadeh., “The concept of a linguistic variable and its application
x−a b − xν
= to approximate reasoning, part 1.” Information Sciences, vol. 8, pp.
K ∗ (b) − K ∗ (x) K ∗ (xν ) − K ∗ (a) b−x xν − a 199–249, 1975.
(27) [15] ——, “The concept of a linguistic variable and its application to
Now let approximate reasoning, part 2.” Information Sciences, vol. 8, pp. 301–
λ λ 357, 1975.
K ∗ (xν ) − K ∗ (a)
x−a b − xν
X := (28) [16] ——, “The concept of a linguistic variable and its application to
b−x xν − a K ∗ (b) − K ∗ (xν ) approximate reasoning, part 3.” Information Sciences, vol. 9, pp. 43–80,
1975.
Then we get the explicit form of K(x) [17] B. D. Liu, C. Y. Chen, and J. Y. Tsao., “Design of adaptive fuzzy logic
XK ∗ (b) + K ∗ (a) controller based on linguistic-hedge concepts and genetic algorithms.”
K ∗ (x) = (29) IEEE Systems, Man and Cybernetics, Part B, vol. 31(1), pp. 32–53,
1+X 2001.
ISBN: 978-1-61804-240-8 96
Advances in Applied and Pure Mathematics
ISBN: 978-1-61804-240-8 97
Advances in Applied and Pure Mathematics
ISBN: 978-1-61804-240-8 98
Advances in Applied and Pure Mathematics
{
The boundary conditions (5) suggest that Eqs. (1) to (4) have
∂v ∂v ∂p ∂ ∂u ∂u a solution of the form
u +v =− + ∇2 v + We 2
∂x ∂y ∂y ∂x ∂x ∂y u= x F ′( y) , v = − F ( y ), T = x θ ( y) (9)
where prime denotes differentiation with respect to y .
u ∂ + v ∂ ∂v + ∂u + 2 ∂v ∂v
∂x ∂y ∂x ∂y ∂x ∂y
Substituting (9) into Eqs. (2) to (4) and eliminating the
pressure p by cross differentiation of Eqs. (2) and (3) it
∂ ∂ 2v ∂v + 2 ∂u ∂v + ∂u
2
results in, after one integration of the resulting equation, the
+ +
∂y ∂x ∂y
2u 4
∂y ∂y ∂x ∂y following ordinary differential equations
∂ 2v ∂ ∂v 2 ∂v ∂u 2 F ′′′ + F F ′′ − F ′ 2 − We ( F F iv − 2 F ′ F ′′′ + F ′′ 2 )
+2v + λ + +
∂y 2
1
∂y ∂y ∂x ∂y − MF ′ ± λ θ + C1 =
0
(3) (10)
∂T ∂T 1 1
u +v = ∇ 2T (4) θ ' ' + F θ ' − F 'θ = 0 (11)
∂x ∂ y Pr Pr
and the boundary conditions in dimensionless form are where C1 is a constant of integration. The boundary
conditions (5) become
=v 0,= u 0,= T T=w ( x) x at= y 0 F=(0) 0, F= '(0) 0, F = '(∞) 1
(12)
ue = x, ve = − y, T = 0, p = pe as y → ∞ θ=(0) 1, θ (= ∞) 0
Taking the limit y → ∞ in Eq. (10) and using the boundary
(5)
conditions F ' (∞) = 1 and θ (∞) = 0 , we get C1 = 1 .
wherePr is the Prandtl number, We is the Weissenberg
Further, from an analysis of the boundary layer equation (10)
number, M is the Hartmann number, λ1 is a viscoelastic it results in that F ( y ) behaves as
parameter and λ is the constant mixed convection parameter. F ( y ) = y + A as y→∞ (13)
where A = A(We) is a constant accounts for the boundary
It should be mentioned that λ > 0 (heated plate) corresponds
to assisting flow and λ < 0 (cooled plate) corresponds to
layer displacement
opposing flow, while λ = 0 corresponds to forced convection Employing (9), the dimensionless skin friction and the local
flow, respectively. heat transfer given by equations (7) and (8) can now be
written as
Using Eqs. (2) and (3), and boundary conditions (5), the non-
τ w = x F ''(0) (14)
dimensional pressure p = pe of the inviscid or far flow can
be expressed as
qw = − xθ ′ (0) (15)
1 where the values of F ′′ (0) and θ ′ (0) can be calculated
p = p e = − ( x 2 + y 2 ) + Const. (6) from equations (10) to (11) with the boundary conditions (12)
2
for various values of the parameters We, λ , M and Pr .
ISBN: 978-1-61804-240-8 99
Advances in Applied and Pure Mathematics
Abstract—In this paper we describe the structure of a workflow logic operator (represented by ti ≺). An input/output logic
as a graph whose vertices represent tasks and the arcs are operator can be the logical AND (•), the OR (⊗), or the XOR
associated to workflow transitions. To each task an input/output -exclusive-or - (⊕). The set A = {at , au , a1 , a2 , . . . , am } is a
logic operator is associated and this logic operator can be the
logical AND (•), the OR (⊗), or the XOR -exclusive-or - (⊕). finite nonempty set of arcs representing workflow transitions.
Furthermore, we associate a Boolean term to each transition The transition at is the tuple (t, t1 ) and transition au is
present in the workflow. the tuple (tn , u), where the symbols t and u represent
The main contribution of this paper is the analysis of a abstract tasks which indicate the entry and ending point of
workflow through its tasks which allows us to describe the the workflow, respectively. Every transition ai , i ∈ {1, . . . , n}
dynamism of the workflow in a very simple way.
Finally, we describe the logical termination of workflows and corresponds to a tuple of the form (tk , tl ), where tk , tl ∈ T.
we present conditions under which this property is valid. We use the symbol 0 to reference the label of a transition,
i.e., a0i references transition ai , ai ∈ A. The elements a0i are
Index Terms—Graphs, Propositional Logic, Workflows, Busi-
ness Processes. called Boolean terms and form the set A0 .
Given ti ∈ T, the incoming transitions for task ti are the
tuples of the form (tl , ti ), tl ∈ T, and the outgoing transitions
I. I NTRODUCTION are the tuples of the form (ti , tl ), tl ∈ T.
Workflow is an abstraction of a business process that
A consists on the execution of a set of tasks to complete
a process (for example, hiring process, loan application, sales
The incoming/outgoing condition of task ti is the
Boolean expression a0k1 ϕ . . . ϕa0kl , where ϕ ∈ {•, ⊗, ⊕},
a0k1 , . . . , a0kl ∈ A0 and ak1 , . . . , akl are the incoming/outgoing
order processing, etc.). Tasks represent unities of work to be transitions of task ti . The terms a0k1 , . . . , a0kl are connected
executed that can be processed by a combination of resources, with the logic operator ti , ti ≺, respectively. If task ti
such as a computer program, an external system, or human has only one incoming/outgoing transition we assume that the
activity. In the literature several papers are devoted to the condition does not have logic operator.
study of workflows, see for example [1], [2], [3], [4], [8], [9], An Event-Action (EA) model for task ti is an implication of
[10], [11], [12], [13], [14], [15], [16], [17] and the references the form ti : fE fC , where fE and fC are the incoming and
therein. In general, these approaches contemplate the use of outgoing conditions of task ti , respectively. An EA model has
Petri Nets, State and Activity Charts, Event-Condition-Action the behavior with two distinct modes: when fE is evaluated to
rules, Temporal Logic and Markov Chains. true, fC is also evaluated to true; when fE is evaluated to
In this paper we present a formalism to describe and analyse f alse, fC is always f alse. And the EA model ti : fE fC
the structure of workflows based on the concept of graph and is true if both fE , fC are true, otherwise it is false. We say
Propositional Logic. An important highlight of this paper is the that the EA model ti : fE fC is positive if its Boolean
emphasis on the tasks present in the workflow, which allows value is true, otherwise it is said to be negative.
us to identify easily the dynamism present in the workflow. We denote by M the set of all EA models present in W G.
Finally, we describe the logical termination in a very intuitive Task ti is said to be executed if the EA model ti : fE fC
form and we present conditions under which this property is is positive. In this case, task ti has attributed the Boolean value
valid. true.
Remark 1: Given an expression whose Boolean value is true
II. W ORKFLOW S TRUCTURE (respectively, false), we simply can represent this fact by 1,
In this Section we provide a complete description of the (respectively, 0).
structure of workflows. We start by introducing the formal
concept of a workflow. This workflow structure can be also Remark 2: Given an EA model ti : fE fC , if fE is false,
found in [5], [6], [7]. Notice this type of graphs has and then task ti disables all its outgoing transitions. Consequently
input/output logic operator associated with each vertex. fC is also false.
Definition 2.1: [5], [6], [7] A workflow is a tri-logic Notice the workflow starts its execution by enabling transi-
acylic directed graph W G = (T, A, A0 , M ), where T = tion at , i.e., by asserting a0t to be true. In other words, the
{t1 , t2 , . . . , tn } is a finite nonempty set of vertices representing workflow starts its execution by executing task t1 .
workflow tasks. Each task ti (i.e., a vertex) has attributed Notice that a0i is true if transition ai is enabled, otherwise ai
an input logic operator (represented by ti ) and an output is false. Transitions can be enabled by a user or by an external
event. If the EA model ti : fE fC is negative, then both
Glória Cravo is with the Center of Exact Sciences and Engineering, Univer- fE , fC are false. In this case, all the transitions of fC are
sity of Madeira, 9020-105 Funchal, Madeira, Portugal, e-mail: gcravo@uma.pt disabled.
(see http://www.uma.pt).
Fig. 1. Example of a workflow. allow to verify if a workflow will eventually finish, according
to the initial specifications.
Definition 2.3: Let W G = (T, A, A0 , M ) be a workflow.
Example 1: In Figure 1 we present a workflow W G =
We say that W G logically terminates if task tn is executed
(T, A, A0 , M ), where T = {t1 , t2 , . . . , t9 }, A = {at ,
whenever task t1 is executed.
au , a1 , a2 , . . . , a11 }, A0 = {a0t , a0u , a01 , a02 , . . . , a011 }, M =
{t1 : a0t a01 • a02 , t2 : a01 a03 ⊕ a04 , t3 : a02 a08 , In the following result we establish a necessary and suffi-
0 0 0 0 0 0 0 0
t4 : a3 a5 ⊕ a6 , t5 : a4 a7 , t6 : a5 a9 , t7 : a6 a010 , cient condition for the logical termination.
0 0 0 0 0 0 0
t8 : a7 ⊕ a9 ⊕ a10 a11 , t9 : a8 • a11 au }.
The output logic operator of task t2 (t2 ≺) is a XOR (⊕), Theorem 2.4: Let W G = (T, A, A0 , M ) be a workflow.
while the input logic operator of task t9 ( t9 ) is an AND Then W G logically terminates if and only if a0u is true
(•). whenever a0t is true.
The incoming transition for task t2 is a1 = (t1 , t2 ) and its Proof 2: Let us assume that W G logically terminates, i.e.,
outgoing transitions are a3 = (t2 , t4 ) and a4 = (t2 , t5 ). Hence task tn is executed whenever task t1 is executed. This means
the incoming condition for task t2 is a01 , while its outgoing that the EA model tn : fEn a0u is positive whenever the
0
condition is a03 ⊕ a04 . EA model t1 : at fC1 is positive. Bearing in mind that
W G starts its execution by executing task t1 , then the EA
Task t2 is executed if the EA model t2 : a01 a03 ⊕ a04 is
0 model t1 : a0t fC1 is positive. Hence the EA model tn :
positive, i.e., if a1 is true and only one of the Boolean terms
fEn a0u is also positive. Consequently, a0t , fC1 , fEn , a0u
a03 , a04 is true.
are true. Thus, a0u is true whenever a0t is true.
Proposition 2.2: Let W G = (T, A, A0 , M ) be a workflow. Conversely, let us assume that a0u is true whenever a0t is
Let al = (ti , tj ) ∈ A, ti , tj ∈ T. If a0l is true, then ti is true. Let us assume that task t1 is executed. This means that
necessarily executed. the EA model t1 : a0t fC1 is positive. Bearing in mind
Proof 1: Let us assume that a0l is true. Let ti : fEi fCi be that a0u is true, according to the behavior of the EA models,
the EA model associated to task ti . If task ti is not executed, necessarily fEn is true. Hence the EA model tn : fEn a0u
then the EA model ti : fEi fCi is negative. Since the is positive, which means that task tn is executed. So we can
EA model is negative, all outgoing transitions of task ti are conclude that task tn is executed whenever task t1 is executed,
disabled, in particular al is disabled, i.e., a0l is false, wich is which means that W G logically terminates.
a contradiction. Hence task ti is executed.
Example 2: It is not hard to check that in the workflow from
Remark 3: The condition of Proposition 2.2 is not sufficient. Figure 1, a0u is true whenever a0t is true. Thus, the workflow
For example in the workflow from Figure 1, if task t2 is logically terminates.
executed, then the EA model t2 : a01 a03 ⊕ a04 is positive. Next we address our study on the dynamism present in
0 0 0
For a1 = true, a3 = true, a4 = f alse, a4 = (t2 , t5 ), t2 is a workflow. Obviously the dynamism is associated to the
executed, but a04 is false. sequencial execution of its tasks. In the workflow from Figure
Remark 4: Let us consider the Boolean term a0l where al = 1 the execution of task t1 implies the execution of both tasks
(ti , tj ) ∈ A, ti , tj ∈ T. If a0l is true, task tj is not necessarily t2 , t3 ; the execution of task t2 implies the execution of only
executed. For example, in the workflow from Figure 2, let us one of the tasks t4 , t5 ; the execution of task t4 implies the
assume that a0t = true, a01 = true, a02 = f alse, a03 = true, execution of only one of the tasks t6 , t7 ; the execution of only
a04 = true, a05 = true, a06 = true, a07 = true, a08 = true, one of the tasks t5 , t6 , t7 implies the execution of task t8 .
a0u = f alse. Hence, for this assignment the EA model t7 : Finally, the execution of both tasks t3 , t8 implies the execution
a06 ⊕ a08 a0u is negative, which means that task t7 is not of taks t9 . Hence, we can state the execution of task t1 implies
executed. Nevertheless, a8 = (t6 , t7 ) and a08 is true. the execution of t2 • t3 ; the execution of task t2 implies the
execution of t4 ⊕ t5 ; the execution of task t4 implies the
Next we introduce the concept of logical termination. This execution of t6 ⊕ t7 ; the execution of t5 ⊕ t6 ⊕ t7 implies
is a very important structural property, since its analysis will the execution of task t8 ; the execution of t3 • t8 implies the
ISBN: 978-1-61804-240-8 104
Advances
JOURNAL OF LATEX CLASS FILES, VOL. 6, NO. 1, JANUARY 2007 in Applied and Pure Mathematics 3
execution of taks t9 . Notice that when we consider t2 • t3 , the true; if the incoming task is false, the outgoing task is false.
operator • is the output logic operator of task t1 ,while when In other words, if tIi ,→ tOi is a compound task model, then
we consider t5 ⊕ t6 ⊕ t7 , ⊕ is the input logic operator of task tIi is executed if and only if tOi is executed.
t8 .
Notice that in a compound task model tIi ,→ tOi , at least
These remarks led us to introduce the following concept. one of the tasks tIi , tOi is compound.
Definition 2.5: Let W G = (T, A, A0 , M ) be a workflow. Example 4: In the workflow from Figure 1, the set of its task
The compound tasks of W G are the elements of the following models is: T M = {t1 ,→ t2 • t3 , t2 ,→ t4 ⊕ t5 , t4 ,→ t6 ⊕ t7 ,
form: ti1 ϕti2 ϕ . . . ϕtik , ti1 , ti2 , . . . tik ∈ T, ϕ ∈ {•, ⊗, ⊕}. t5 ⊕ t6 ⊕ t7 ,→ t8 , t3 • t8 ,→ t9 }.
The set of all compound tasks of W G is denoted by T 0 , i.e.:
From now on, we use the symbol ←→ with the following
T 0 = {ti1 ϕti2 ϕ . . . ϕtik : ti1 , ti2 , . . . tik ∈ T, ϕ ∈ {•, ⊗, ⊕}}. meaning: X ←→ Y means that the compound statements X
Example 3: In the workflow from Figure 1, T 0 = {t2 • t3 , and Y are logically equivalent.
t4 ⊕ t5 , t6 ⊕ t7 , t5 ⊕ t6 ⊕ t7 , t3 • t8 }. According to simple rules of Logic and taking into account
the behavior of the task models, we can infer the following
Remark 5: Since every task ti has associated a Boolean rules that allow us to identify new task models present in the
value, according to its execution, it is also natural to attribute workflow.
a Boolean value to the compound tasks of W G. The natural
attribution is the following. Given any compound task of W G, Remark 6: Let W G = (T, A, A0 , M ) be a workflow.
ti1 ϕti2 ϕ . . . ϕtik , ϕ ∈ {•, ⊗, ⊕}: (a) If both tasks models tIi ,→ tOi and tIj ,→ tOj belong
If ϕ = •, then the Boolean value of ti1 ϕti2 ϕ . . . ϕtik is 1 to T M and tOi ←→ tIj then the model tIi ,→ tOj still holds
if and only if the Boolean value of all tasks ti1 , ti2 , . . . , tik is in W G.
equal to 1; (b) If both task models tIi ,→ tOi and tIj ,→ tOj belong
If ϕ = ⊗, then the Boolean value of ti1 ϕti2 ϕ . . . ϕtik is 1 if to T M , where tOi ←→ tL ϕtIj , ϕ ∈ {•, ⊗, ⊕} then the task
and only if there exists at least one of the tasks ti1 , ti2 , . . . , tik model tIi ,→ tL ϕtOj still holds in W G.
whose Boolean value is equal to 1; (c) If both task models tIi ,→ tOi and tOj ,→ tIj belong
If ϕ = ⊕, then the Boolean value of ti1 ϕti2 ϕ . . . ϕtik is 1 to T M , where tOi ←→ tL ϕtIj , ϕ ∈ {•, ⊗, ⊕} then the task
if and only if there exists only one of the tasks ti1 , ti2 , . . . , tik model tIi ,→ tL ϕtOj still holds in W G.
with Boolean value equal to 1. (d) If both task models tIi ,→ tOi and tIj ,→ tOj belong
Naturally, we can state that a compound task to T M , where tIi ←→ tL ϕtIj , ϕ ∈ {•, ⊗, ⊕} then the task
ti1 ϕti2 ϕ . . . ϕtik is executed if and only if its Boolean model tL ϕtOj ,→ tOi still holds in W G.
value is equal to 1, which means that the compound task (e) If both task models tIi ,→ tOi and tOj ,→ tIj belong
ti1 ϕti2 ϕ . . . ϕtik is positive. In other words, ti1 ϕti2 ϕ . . . ϕtik to T M , where tIi ←→ tL ϕtIj , ϕ ∈ {•, ⊗, ⊕} then the task
is executed if and only if: model tL ϕtOj ,→ tOi still holds in W G.
If ϕ = •, all tasks ti1 , ti2 , . . . , tik are executed;
The previous remark allow us to identify new task models,
If ϕ = ⊗, at least one of the tasks ti1 , ti2 , . . . , tik is
as it is described below.
executed;
If ϕ = ⊕, only one of the tasks ti1 , ti2 , . . . , tik is executed. Definition 2.7: Let W G = (T, A, A0 , M ) be a workflow.
An extended task model is a model obtained by applying a
Definition 2.6: Let W G = (T, A, A0 , M ) be a workflow. Let
finite sequence of some of the rules presented in Remark 6.
ti , tj , ti1 , ti2 , . . . , tik , tj1 , tj2 , . . . , tjl ∈ T, ϕ, ψ{•, ⊗, ⊕}.
We denote by T M 0 the set of all extended task models of
A compound task model is an implication with one of the
W G.
following forms:
(1) ti ,→ tj1 ψtj2 ψ . . . ψtjl ; Example 5: In the workflow From Figure 1, bearing in mind
(2) ti1 ϕti2 ϕ . . . ϕtik ,→ tj ; that t1 ,→ t2 • t3 , t2 ,→ t4 ⊕ t5 ∈ T M, according to Remark 6
(3) ti1 ϕti2 ϕ . . . ϕtik ,→ tj1 ψtj2 ψ . . . ψtjl . we can conclude that the model t1 ,→ (t4 ⊕ t5 ) • t3 still holds
Usually we represent a compound task model by tIi ,→ tOi , in W G. Therefore, we can state that t1 ,→ (t4 ⊕ t5 ) • t3 is an
where tIi is called the incoming task and tOi is called the extended task model of W G.
outgoing task. We say that a compound task model tIi ,→ tOi
is positive if both incoming and outgoing tasks are positive, Notice we adopt the same notation of the task models to
i.e., if both tasks tIi , tOi are executed. represent the extended task models. Furthermore, the extended
In particular, the implication of the form ti ,→ tj is called task models verify the same properties of the task models. In
a simple task model. Clearly, it is positive if both tasks ti , tj particular, given an extended task model B ,→ C, necessarily
are executed. both B, C have the same Boolean value.
The set of all simple and compound task models present in Definition 2.8: Let W G = (T, A, A0 , M ) be a workflow. We
W G is called the set of task models of W G and is denoted define the closure of T M as the set of all task models and
by T M. extended task models in W G. This set is denoted by T M ∗ .
The task models have the behavior with two distinct modes: In other words, T M ∗ = T M ∪ T M 0 .
if its incoming task is true, necessarily its outgoing task is
ISBN: 978-1-61804-240-8 105
Advances
JOURNAL OF LATEX CLASS FILES, VOL. 6, NO. 1, JANUARY 2007 in Applied and Pure Mathematics 4
Example 6: As we saw in Example 4 in the workflow from [9] J. Eder, H. Groiss, and H. Nekvasil. A workflow system based on active
Figure 1, T M = {t1 ,→ t2 • t3 , t2 ,→ t4 ⊕ t5 , t4 ,→ t6 ⊕ t7 , databases. In G. Chroust and A. Benczur, editors, Proceedings of CON’
94, Workflow Management: Challenges, Paradigms and Products, pages
t5 ⊕ t6 ⊕ t7 ,→ t8 , t3 • t8 ,→ t9 }. Since t1 ,→ t2 • t3 , t2 ,→ 249–265, Linz, Austria, 1994.
t4 ⊕ t5 ∈ T M , according to Remark 6 we can deduce that [10] F. Gottschalk, W. V. M. P. v. d. Aalst, M. H. Jansen-Vullers, and
t1 ,→ (t4 ⊕ t5 ) • t3 ∈ T M ∗ . Now bearing in mind that t4 ,→ M. La Rosa. Configurable workflow models. International Journal
of Cooperative Information Systems, 17(2):223–255, 2008.
t6 ⊕ t7 ∈ T M, applying again Remark 6 we can conclude that [11] F. Gottschalk, W. V. M. P. v. d. Aalst, M. H. Jasen-Vullers, and H. M. W.
t1 ,→ ((t6 ⊕t7 )⊕t5 )•t3 ∈ T M ∗ . As (t6 ⊕t7 )⊕t5 ←→ t5 ⊕t6 ⊕ Verbeek. Protor2CPN: Using colored petri nets for configuring and
t7 we can state that t1 ,→ (t5 ⊕ t6 ⊕ t7 ) • t3 ∈ T M ∗ . Bearing testing business processes. International Journal on Software Tools for
Technology Transfer, 10(1):95–111, 2008.
in mind that t5 ⊕ t6 ⊕ t7 ,→ t8 , applying once more Remark [12] A. H. M. t. Hofstede and E. R. Nieuwland. Task structure semantics
6 we infer that t1 ,→ t8 • t3 ∈ T M ∗ . As t8 • t3 ←→ t3 • t8 , through process algebra. Software Engineering Journal, 8(1):14–20,
applying again Remark 6 we conclude that t1 ,→ t9 ∈ T M ∗ . 1993.
[13] J. Klingemann, J. Wäsch, and K. Aberer. Deriving service models in
Notice the workflow from Figure 1 logically terminates and cross-organizational workflows. In Proceedings of RIDE-Information
Technology for Virtual Enterprises (RIDE-VE’ 99), pages 100–107,
t1 ,→ t9 ∈ T M ∗ . Furthermore, we studied many other exam- Sydney, Australia, 1999.
ples of workflows that logically terminates and simultaneously [14] P. Muth, D. Wodtke, J. Weissenfels, G. Weikum, and K. Dittrich.
t1 ,→ tn ∈ T M ∗ . The analysis of these different cases led us Enterprise-wide workflow management based on state and activity
charts. In A. Dogac, L. Kalinichenko, T. Ozsu, and A. Sheth, editors,
to formulate the following conjecture. Proceedings NATO Advanced Study Institute on Workflow Management
Systems and Interoperability. Springer-Verlag, 1998.
Conjecture 1: Given a workflow W G = (T, A, A0 , M ), then [15] A. Rozinat, R. S. Mans, M. Song, and W. V. M. P. v. d. Aalst.
W G logically terminates if and only if t1 ,→ tn ∈ T M ∗ . Discovering colored petri nets from event logs. International Journal
on Software Tools for Technology Transfer, 10(1):57–74, 2008.
[16] M. P. Singh and K. van Hee. Semantical considerations on workflows:
III. C ONCLUSION An algebra for intertask dependencies. In Fifth International Workshop
In this paper we develop a formalism to describe and analyse on Database Programming Languages, Umbria, Italy, 1995. Springer.
[17] H. M. W. Verbeek, W. V. M. P. v. d. Aalst, and A. H. M. Hofstede.
the structure of workflows. Furthermore, our analysis allows Verifying workflows with cancellation regions and OR-joins: An ap-
us to study the logical termination of workflows. In particular proach based on relaxed soundness and invariants. Computer Journal,
we present conditions under which this property is valid. 50(3):294–314, 2007.
It is important to point out that our main emphasis is
the analysis of a workflow through the study of its tasks.
Another relevant aspect of our approach is the introduction
of the concept of compound tasks. This concept allows us to
identify new task models based on the existing ones. Through
these new task models we are able to describe the dynamism
present in a workflow. Clearly, the study of the dynamism of
a workflow is equivalent to analyse the sequential execution
of its tasks.
Finally, we conjecture that a workflow (W G) logically
terminates if and only if t1 ,→ tn is in the closure of W G.
R EFERENCES
[1] W. V. M. P. v. d. Aalst. The application of petri nets to workflow Glória Cravo holds a first degree in Mathematics from the University
management. Journal of Circuits, Systems and Computers, 8(1):21–66, of Lisbon, Portugal (1992), and a MSc. Degree in Mathematics from the
1998. University of Lisbon, Portugal (1997). In 2003 she obtained her PhD in
[2] W. V. M. P. v. d. Aalst. Workflow verification: Finding control-flow Mathematics, from University of Lisbon, Portugal. The main contribution of
errors using petri-net-based techniques. In W. V. M. P. v. d. Aalst, her PhD thesis is in the area of the so-called Matrix Completion Problems.
J. Desel, and A. Oberweis, editors, Business Process Management: Between 1992 and 1997, she worked at the Department of Mathematics
Models, Techniques, and Empirical Studies, pages 161–183. Springer- of the University of Madeira as auxiliary teacher. From 1997 to 2003 she
Verlag, Berlin, 2000. worked at the Department of Mathematics of the University of Madeira as
[3] P. Attie, M. Singh, A. Sheth, and M. Rusinkiewicz. Specifying and teaching assistant. Since 2003 she works at the Department of Mathematics
enforcing intertask dependencies. In Proceedings of 19th International and Engineering of the University of Madeira as assistant professor. Her main
Conference on Very Large Data Bases, pages 134–145, Dublin, Ireland, areas of research are Matrix Completion Problems, Graph Theory, Control
1993. Morgan Kaufman. Theory, Logic, Computer Science and Wireless Sensor Networks. She is an
[4] J. Cao, C. Chan, and K. Chan. Workflow analysis for web publishing active research member of the Center for Linear Structures and Combinatorics.
using a state-activity process model. Journal of Systems and Software,
76(3):221–235, 2005.
[5] G. Cravo. Applications of propositional logic to workflow analysis.
Applied Mathematics Letters, 23:272–276, 2010.
[6] G. Cravo. Logical termination of workflows: An interdisciplinary
approach. Journal of Numerical Analysis, Industrial and Applied
Mathematics, 5(3-4):153–161, 2011.
[7] G. Cravo and J. Cardoso. Termination of workflows: A snapshot-based
approach. Mathematica Balkanica, 21(3-4):233–243, 2007.
[8] U. Dayal, M. Hsu, and R. Ladin. Organizing long-running activities with
triggers and transactions. In ACM SIGMOD International Conference
on Management of Data Table of Contents, pages 204–214, Atlantic
City, New Jersey, 1990. ACM Press, New York, NY, USA.
r{t ,...,t } , with {t1 ,..., tn } ∈ {1,..., T1} × ... × {1,..., Tn } , of the type: Gaussian ones, but “shoulders” (only decreasing or only
1 n
increasing MFs) are required to model extreme linguistic terms
if x(1) is F (1) and ... and x( n ) is F ( n ) then C instead. Therefore, triangular, Gaussian and bell-shaped MFs
1 1 {1,...,1}
will not be considered here, even if Gaussian MFs require a
... , (1) lower number of parameters.
(1) (1) (n) (n) Trapezoidal MFs can be provided with some improvements,
if x is FT and ... and x is FT then C{T ,...,T }
1 n 1 n in order to reach: i) derivability, i.e. to obtain MFs smoothly
where antecedents are all possible combinations of fuzzy sets going from the plateaus to the increasing/decreasing segments;
of the partitions, and consequents are fuzzy sets [8,9]. ii) non-linearity, i.e. to obtain MFs overcoming the limits of
Each data sample xi fires the rule r{t ,...,t } with a strength linear models between input and output variables. These issues
1 n
can be settled by considering sigmoidal MFs.
( j ) ( j ) A trapezoidal or a sigmoidal MF can be represented by 4
FS{t ,...t } = T µ xi ,
j =1,...,n t j
(2)
1 n
parameters a, b, c, d, while a “shoulder” by 2 parameters. A
while the implication of the consequence is usually modeled whole partition with T MFs can be described by 2 (T − 1)
as:
parameters. Binary MFs are represented here with the same
IMP{t ,...,t } = T FS{t ,...,t } , C{t ,...,t } , (3) number of parameters. Therefore, the MFs are modeled here as
1 n 1 n 1 n
follows: right “shoulders” µ R [ x; c, d ] are binary
and different implications are aggregated as:
IMP , 0 , x ≤ (c + d ) / 2
AGG = S (4) µ RB [ x; c, d ] =
{t ,...,t }∈{1,...,T }×...×{1,...,T } {t1 ,...,tn } (c + d ) / 2 ≤ x
, (6)
1 n 1 n
1,
where T is a T-norm and S is an S-norm.
trapezoidal
Note that FS in (2) is a number, while IMP in (3) and
AGG in (4) are fuzzy sets. Suppose that there are K possible 0 , x≤c
x−c
classes C1 ,..., CK , then the fuzzy set AGG assume a set of
µ RT [ x; c, d ] = ,c<x<d , (7)
values AGG1 ,..., AGGK in correspondence of different d −c
1, d≤x
classes. Therefore, a defuzzification step is required, which
can be accomplished in different ways: usually, the class k or sigmoidal
which takes the greatest AGGk is chosen (winner-takes-all 1
µ RS [ x; c, d ] = , (8)
strategy); however, since a result constituted by different d +c x
t⋅ − 2⋅t ⋅
1+ e d −c d −c
classes with respective confidence grades should be preferred
[10], then the defuzzification can be implemented as a respective left “shoulders” are
normalization [4]: µ L [ x; a, b] = 1 − µ R [ x; a, b ] , (9)
K
while internal MFs are
DEFk = AGGk / ∑ AGGκ . (5)
µ I [ x; a, b, c, d ] = 1 − µ L [ x; a, b ] − µ R [ x; c, d ] , (10)
κ =1
with a ≤ b ≤ c ≤ d and t = Log [1 / ε − 1] , where ε 1 (in the
B. Shape of Membership Functions
following ε = 0.01 ) is a positive constant fixed to approximate
Only few works have discussed on the opportunity of using
a certain shape for MFs [1-3], however, results seem to be sigmoidal MFs to normal, i.e. ε ≤ µ [ x ] ≤ 1 − ε . Using these
application-dependent. In general, the following considerations settings, binary sets divide the variable range in intervals,
could be made. trapezoids fuzzify these intervals, and trapezoidal and
First of all, the difference between crisp intervals (binary sigmoidal MFs with the same parameters result very similar,
MFs) and fuzzy intervals (trapezoidal MFs) can be studied, the latter being smoother.
even if the advantages of fuzzy logic [11] are widely In the following, a comparison of performances will be
recognized [12,13]. made among binary, trapezoidal and sigmoidal MFs. In Fig. 1,
Fuzzy numbers (triangular MFs) can be viewed as particular the three types of partition are shown, all with a number of
cases of fuzzy intervals. Even if they are used very often, their terms T = 3 and the same parameters.
advantage with respect to fuzzy intervals can be ascribed only
to a lower number of parameters (3 instead of 4), which
reflects the difference between a number and an interval.
However, intervals are generally more appropriate to model
the terms of a linguistic variable. On the other hand, intervals
of the type x < x A or x > xB cannot be fuzzified with Fig. 1 Binary, trapezoidal, and sigmoidal fuzzy partitions.
increasing and then decreasing MFs, like triangular and
C. A Family of (Soft) T-norms and S-norms are scarcely used in fuzzy systems where interpretability is a
Many types of T-norms and S-norms were developed. In main objective, since their use undoubtedly complicates the
particular, Mamdani, Product, Łukasiewicz and Drastic T- comprehensibility of the system.
norms are the most used. However, very few attempts [4] were In the following, the advantages of axes rotation will be
done on deciding which type is the most suitable for different studied by substituting a couple of original variables x1 and
applications. At the same time, some parameterized families of x2 with another orthogonal couple obtained by linearly
norms were developed, which smoothly changes one norm into combining them through a parameter D, i.e.
the others by varying just one parameter p. Here, the Frank T-
norm family [14] is considered:
{ x1 , x2 } ⇒ { x1 + Dx2 , x2 − Dx1} . (15)
∑ aδ + α S a1,..., ad ; p .
1 a membership grade of 1) or a proper fuzzy set [8], or a
S a1 ,..., ad ; p, α = (1 − α ) (14)
d function [9]. In case of a classifier, which is examined in this
δ =1
work, each consequent can be a singleton whose support is a
In the following, the dependence of system performances on
class [8], or a fuzzy set defined on different classes [8].
both p and α parameters will be evaluated. In particular,
Suppose that there are K possible classes C1 ,..., CK . In the
their use for different T-norms and S-norms, i.e. pF and α F
simplified model, each consequent is a singleton whose
in (2), pI and α I in (3), p A and α A in (4), is investigated, support is a class. Deeper knowledge can be modeled by a
and different optimal values of them can be found. more complicated system. For example, if one knows that, in
the space restriction modeled by antecedents of a rule
D.Preferred Directions
r{t ,...,t } , different classes have respective probabilities
Some of the best regression [16,17] or classification [18,19] 1 n
systems are based on the observation that there exist a set of p{t ,...,t }−1 ,..., p{t ,...,t }− K , then these can be assigned to the
1 n 1 n
directions in the variables space, which is better than others in
rule consequent, which can be regarded as a fuzzy sets with
explaining the behavior of the system. E.g., in [16] these
different non-zero membership grades in correspondence of
preferential directions are those which maximize the variance
different classes. In this case, the result of defuzzification (5)
and result more useful to study the output variable, while in
can be viewed as the posterior probability [5]:
[18,19] they define a hyperplane which divides the space
K
∑ AGGκ
where the samples of one class are situated from the other
P Ck | x = AGGk / . (18)
where there is the other class. Generally, these directions do
κ =1
not correspond to the original variables of the dataset, but are
linear combinations of them. They can be viewed as a possible G. A Reference Method for Knowledge Extraction
reduction of dimensionality and axes rotation. The method [5], which was recently developed by authors,
On the one hand, preferred directions could be surely useful is considered as a reference one. It can be summarized as
to improve the system performances. On the other hand, they follows.
Each range of the original variables is partitioned into a probability of C B , given different values of UCS and BC, i.e.
collection of Tj fuzzy sets described by sigmoidal MFs, P [CB | x ] , in red, and that representing P [CM | x ] in green,
( j)
µt x( ) , t j = 1,..., T j , such that:
j
are showed together. The lateral view of the images is given
j
when the height of the surfaces is to be shown, while the top
Tj
view is preferred to emphasize the separation of the variables
P Ck | x ( ) ∑ λt −k ⋅ µt( ) x( j ) ,
j j
(19)
j j space into regions associated to different classes (e.g., if
t j =1
P [CB | x ] > P [CM | x ] , then the point x of the space is
with k = 1,..., K and j = 1,..., n .
assigned to the class C B and in the top view it is red).
Note that parameters of MFs and the number of fuzzy sets
Results of different systems are evaluated in terms of
constituting each partition are optimized at the same time.
classification error (CE) and squared classification error
Then, a rule base as (1) is constructed, by considering a
(SCE):
complete set of rules r{t ,...,t } , each one having a weight N
1 n
∑∑ (α
1 1 2
K SCE = k
− δ ik ) , (24)
∑ w{t ,...,t }−k
i
N K i =1 k =1
1 n
k =1
W{t ,...,t } =
1 n T1 Tn K
(20) where Cˆi is the predicted class, Ci the real one, α ik the
∑ ... ∑ ∑ w{τ ,...,τ }−k 1 n activation of k-th class for the i-th data sample, and δ ik is 1 if
τ1 =1 τ n =1 k =1
the correct class for the i-th data sample is the k-th one, 0
w{t ,...,t }− k otherwise. In particular, while CE is simply the fraction of
1 n
p{t ,...,t }−k = , (21) wrongly classified data items, SCE takes into account that high
1 n K
∑w{t ,...,t }− k
1 n
(low) confidence should be assigned to right (wrong)
solutions. Both should be minimized.
k =1
where The method described in Section II.G is used to determine
λt1 −k ⋅ ... ⋅ λtn − k parameters of membership functions and the number of terms
w{t ,...,t }−k = , (22) of each partition. In particular, only two fuzzy sets constitute
1 n n −1
P Ck each of the two partitions, and the parameters of the MFs (8)
and P Ck are prior probabilities of classes. and (9) results: 1.09 and 5.37 for UCS, 1.34 and 6.01 for BC.
The obtained partitions are shown in Fig.2.
The types of norms used are product T-norm in (2), product Moreover, the reference method is used to determine
T-norm in (3), and weighted sum for S-norm in (4). After optimal rule weights (20) and rule consequents (21), of a rule
normalization (5), it is demonstrated that the results base as (1), which are listed in the first column of Table I.A,
approximate posterior probabilities as in (18). together with the resulting performances.
In the following, this method will be used to calculate In order to evaluate the usefulness of system complications,
parameters of membership functions, rule weights and rule let us begin by considering a simple system, as described in
consequents. Its results will be compared with those of the Section II.A, where parameters p and α of norms are set as
systems obtained by simplifying it and then: i) substituting
follows: pF = pI = p A = 0 and α F = α I = α A = 1 . Original
different shapes of MFs; ii) modifying types of T-norms and S-
norms; iii) using soft T-norms and S-norms; iv) rotating axes; variables are used, no weights are assigned to rule antecedents
v) adding weights to antecedents; vi) adding weights to rules; nor to different rules, and consequents are defined as
vii) substituting singletons with fuzzy consequents. singletons, obtained by substituting the highest probability
found by (21) with 1, the others with 0.
III. RESULTS AND DISCUSSION
In this section, the Wisconsin Breast Cancer Dataset [6] is
used as a proof of concepts. 699 samples are classified into
benign (CB) and malignant (CM). Two input variables of the
dataset are used here, which are Uniformity of Cell Shape
(UCS) and Bland Chromatin (BC).
The choice of only two variables derives from the need of
showing the behavior of 3D surfaces representing the functions Fig. 2 Fuzzy partitions of two variables calculated by the reference
of the posterior probabilities of classes (CPPFs). In the method.
following (see Figs 3-7, 9), the surface representing the
The first comparison can be made among different shapes of has fundamental importance. In the following, original
MFs, going from binary, through trapezoidal, to sigmoidal variables are kept.
MFs. CPPFs given by the simple system in these three cases
are shown in Fig.3, where CPPFs reflect the shape chosen for
MFs, and the observations given in Section II.B can be
repeated. Measures of performance are listed in Table I.A,
where it can be seen that in all three cases the same CE is
obtained, while an improvement of SCE is obtained with fuzzy
systems with respect to crisp one. Using trapezoidal or
sigmoidal MFs gives very similar results, therefore for the
following runs the sigmoidal shape is chosen, in order to
ensure derivability.
If the type of norms (11) and (12) is changed by varying
pF and pI , as explained in Section II.C, then the CPPFs
result very similar. If p A is varied (see Fig. 4), then the first
difference can be detected by comparing the case of p A = 0
Fig. 3 Functions of class posterior probabilities by using (a) binary,
with others: in the null case, the separation between regions in
(b) trapezoidal and (c) sigmoidal MFs.
which different classes are associated can be drawn by
orthogonal segments connected by an angle, while in the other
cases, the segments are connected by a curve, as can be seen
by comparing Fig. 4 (a) and (c) (top views). Another
significant variation happens for values of p A < 2 , different
from 0 and 1, i.e. CPPFs flatten around the probability of 0.5,
as can be seen in Fig. 4 (b). The evaluation of performances
reveals that only varying p A has significant influence on
results, and in particular, if it assumes values of 0, 1 and +∞
the best results are obtained, as can be seen by Table I.A.
Therefore, since pF and pI can be chosen equal to 0, 1 or
+∞ , in order to have simple norms, the choices of 1 or +∞
can be done for p A in order to improve performances and
obtain curve class separation. For the following runs, the same
settings of the reference method are chosen.
If soft norms are used as in (13) and (14), by varying \
parameters α F , α I and α A as explained in Section II.C, then Fig. 4 Functions of class posterior probabilities by using (a) p A = 0 ,
the class posterior probabilities are mainly influenced by the (b) p A = 0.5 and (c) p A = 1 .
variation of α F and α I , as can be seen in Fig. 5. In
particular, surfaces flatten when these two parameters
decrease. On the other hand, in Table I.A it can be seen that
the best performances are obtained when soft norms are not
used. If a soft norm is used for aggregation, no significant
effect is gained, but a complication is added. Therefore, for the
following runs, α F = α I = α A = 1 .
In Fig. 6 (top views), the axes rotation due to the variation
of D in (15) from −0.2 to 0.2 is shown. From the analysis of
results, a value of D = −0.12 was found to minimize CE,
while a value of D = −0.08 minimizes SCE, as reported in
Table I.B. This is in accordance with the observation that
original variables are not necessarily the best, and the use of
their linear combination can often improve system
performances. However, in this case the improvements are not
great enough to justify the great loss of comprehensibility due Fig. 5 Functions of class posterior probabilities by using soft norms:
to the use of linear combination of variables in the rule base, (a) α F = 0 , α I = 1 , α A = 1 , (b) α F = 1 , α I = 0 , α A = 1 , (c) α F = 1 ,
especially for applications like medicine, where interpretability α I = 1 , α A = 0 , (d) α F = 1 , α I = 1 , α A = 1 .
TABLE I.A
PERFORMANCES OF DIFFERENT SYSTEMS
Settings
Reference Shapes Norm types Soft norms
Shape:
Binary=B
S B T S S S S S S S S
Trapezoidal=T
Sigmoidal=S
pF 1 0 0 0 0, 2 +∞ 1 1 1 1 1
pI 1 0 0 0 0, +∞ 0 1 1 1 1 1
pA +∞ 0 0 0 {0,1} 0 0, 2 +∞ +∞ +∞ +∞
αF 1 1 1 1 1 1 1 1 0 1 1
αI 1 1 1 1 1 1 1 1 1 0 1
αA 1 1 1 1 1 1 1 1 1 1 0
D 0 0 0 0 0 0 0 0 0 0 0
wUCS 1 1 1 1 1 1 1 1 1 1 1
wBC 1 1 1 1 1 1 1 1 1 1 1
W{low,low} 0.35 1 1 1 1 1 1 1 1 1 1
W{low.high} 0.02 1 1 1 1 1 1 1 1 1 1
W{high,low} 0.04 1 1 1 1 1 1 1 1 1 1
W{high, high} 0.59 1 1 1 1 1 1 1 1 1 1
p{low,low} − B 1 1 1 1 1 1 1 1 1 1 1
p{low,low} − M 0 0 0 0 0 0 0 0 0 0 0
p{low, high} − B 1 1 1 1 1 1 1 1 1 1 1
p{low, high} − M 0 0 0 0 0 0 0 0 0 0 0
p{high,low} − B 0.66 1 1 1 1 1 1 1 1 1 1
p{high,low} − M 0.34 0 0 0 0 0 0 0 0 0 0
p{high, high} − B 0 0 0 0 0 0 0 0 0 0 0
p{high, high} − M 1 1 1 1 1 1 1 1 1 1 1
Performances
CE 0.047 0.112 0.112 0.112 0.112 0.112 0.112 0.112 0.345 - 0.112
SCE 0.039 0.112 0.069 0.077 0.077, 0.079 0.084 0.079, 0.248 0.080 0.117 0.250 0.080
TABLE I.B
PERFORMANCES OF DIFFERENT SYSTEMS
Settings
Axes rotation Antecedent weights Rule weights Rule conclusions Reference
Shape:
Binary=B
S S S S S S S S
Trapezoidal=T
Sigmoidal=S
pF 1 1 1 1 1 1 1 1
pI 1 1 1 1 1 1 1 1
pA +∞ +∞ +∞ +∞ +∞ +∞ +∞ +∞
αF 1 1 1 1 1 1 1 1
αI 1 1 1 1 1 1 1 1
αA 1 1 1 1 1 1 1 1
D 0 -0.12 -0.08 0 0 0 0 0
wUCS 1 1 1 0 1 1 1 1
wBC 1 1 1 1 0 1 1 1
W{low,low} 1 1 1 1 1 0.35 1 0.35
W{low, high} 1 1 1 1 1 0.02 1 0.02
W{high,low} 1 1 1 1 1 0.04 1 0.04
W{high, high} 1 1 1 1 1 0.59 1 0.59
p{low,low} − B 1 1 1 1 1 1 1 1
p{low,low} − M 0 0 0 0 0 0 0 0
p{low, high} − B 1 1 1 1 1 1 1 1
p{low, high} − M 0 0 0 0 0 0 0 0
p{high,low} − B 1 1 1 1 1 1 0.66 0.66
p{high,low} − M 0 0 0 0 0 0 0.34 0.34
p{high, high} − B 0 0 0 0 0 0 0 0
p{high, high} − M 1 1 1 1 1 1 1 1
Performances
CE 0.112 0.102 0.103 0.345 0.345 0.052 0.112 0.047
SCE 0.080 0.077 0.076 0.135 0.117 0.041 0.060 0.039
REFERENCES
[1] N. Gupta, S. K. Jain, “Comparative analysis of fuzzy power system
stabilizer using different membership functions,” International Journal
of Computer and Electrical Engineering, vol. 2 (2), pp. 262-267, 2010.
[2] J. G. Monicka, N. Sekhar, K. R. Kumar, “Performance evaluation of
membership functions on fuzzy logic controlled AC voltage controller
for speed control of induction motor drive,” International Journal of
Computer Applications, vol. 13 (5), 2011.
[3] J. Zhao, B. Bose, “Evaluation of membership functions for fuzzy logic
controlled induction motor drive,” in Proc. 28th IEEE Annual
Conference of the Industrial Electronics Society, Svilla, Spain,
November 2002, pp. 229-234.
[4] L. Rutkowski, K. Cpalka, Flexible Neuro-fuzzy systems, IEEE
Transactions on Neural Networks, vol. 14 (3), pp. 554-574, 2003.
[5] M. Pota, M. Esposito, G. De Pietro, “Combination of Interpretable
Fuzzy Models and Probabilistic Inference in Medical DSSs,” in Proc.
8th International Conference on Knowledge, In-formation and
Creativity Support Systems, Krakow, Poland, 2013, pp. 541-552.
[6] K. Bache, M. Lichman, UCI Machine Learning Repository, 2013.
<http://archive.ics.uci.edu/ml>. Irvine, CA: University of California,
School of Information and Computer Science.
[7] P. Y. Glorennec, Algorithmes d’apprentissage pour systèmes
d’inférence floue, Editions Hermès, Paris, 1999.
[8] E. H. Mamdani, S. Assilian, “An experiment in linguistic synthesis with
a fuzzy logic controller,” International Journal of Man-Machine
Studies, vol. 7 (1), pp. 1-13, 1975.
[9] M. Sugeno, Industrial applications of fuzzy control, Elsevier Science
Pub. Co., 1985.
[10] M. Pota, M. Esposito, G. De Pietro, “Fuzzy partitioning for clinical
Fig. 9 (a) No rule weights and single class conclusions; (b) optimized DSSs using statistical information transformed into possibility-based
rule weights; (c) optimized rule conclusions; (d) reference method. knowledge,” Knowledge-Based Systems, 2014,
http://dx.doi.org/10.1016/j.knosys.2014.06.021
IV. CONCLUSION [11] L. Zadeh, “Fuzzy sets,” Information and Control, vol. 8, pp. 338-353,
1965.
Various degrees of freedom which characterize the [12] T. E. Rothenfluh, K. Bögl, K.-P. Adlassnig, “Representation and
modeling of a rule-based fuzzy system for classification were Acquisition of Knowledge for a Fuzzy Medical Consultation System,”
in Studies in Fuzziness and Soft Computing, P. S. Szczepaniak, P. J. G.
analyzed, with the aim of individuating the possible settings Lisboa, J. Kacprzyk, Eds., vol. 41, Fuzzy systems in Medicine, Physica-
and complications of a simple system which can improve the Verlag HD, 2000, pp. 636-651.
classification performances, taking into account the loss of [13] J .M. Alonso, C. Castiello, M. Lucarelli, C. Mencar, “Modelling
interpretable fuzzy rule-based classifiers for Medical Decision Support,”
interpretability connected with them. The following in Medical Applications of Intelligent Data Analysis: Research
conclusions can be drawn: i) the trapezoidal shape of fuzzy advancements, R. Magdalena, E. Soria, J. Guerrero, J. Gómez-Sanchis,
sets is preferable with respect to others frequently used, such A.J. Serrano, Eds., IGI Global, 2012, pp. 255-272.
as triangular and Gaussian, and the sigmoidal shape adds [14] M. Frank, “On the simultaneous associativity of F(x,y) and x+y–F(x,y),”
Aequationes Math., vol. 19, pp. 141-160, 1979.
derivability with the same number of parameters; ii) simple [15] R. R. Yager, D. P. Filev, Essentials of Fuzzy Modeling and Control,
norms, like Mamdani (not for aggregation), Product, and New York: Wiley, 1994.
Łukasiewicz ones can be adopted, while the use of soft norms [16] M. G. Kendall, A course in Multivariate Analysis, London: Griffin,
1957.
should be avoided; iii) axes rotation can be advantageous, but [17] A. Hōskuldsson, “PLS regression methods,” Journal of Chemometrics,
the performance improvement can be small with respect to the vol. 2 (3), pp. 211-228, 1988.
loss of comprehensibility; iv) using weights for rule [18] R. A. Fisher, “The use of multiple measurements in taxonomic
problems,” Annals of Eugenics, vol. 7 (2), pp. 179-188, 1936.
antecedents can be advantageous, but not in the examined [19] R. C. Rao, “The utilization of multiple measurements in problems of
case, however the associated loss of interpretability is biological classification,” Journal of the Royal Statistical Society, Series
significant; v) the use of rule weights strongly improves the B (Methodological), vol. 10 (2), pp. 159-203, 1948.
system performances, therefore it is recommended, even if
d 7 H P d 1 P
2 cos cos 1 sin cos 2
dt m m dt mV 2
P P P
2 sin cos 2 2 sin 3 sin 3 with P 0,
m V cos
.
mV mV 2
The laws of parameter variation α, β and Ρ with optimal d 1 1
control are determined from the Hamiltonian maximization 2 22 2 with P 0.
condition. The relations for calculation of optimal values α dt rV r
In contexts of the made assumptions in both cases
and β are obtained from the conditions H / 0 ˙
and H / 0 : 2 0, that is 2 С*.
2 ˙
tg , (9) We show that 1 t 0 with t0 t t R. The
V 1
3 cos equation for 1 with P 0 considering (9) and (10) is
tg 3 . (10)
V 1 cos 2 sin cos V 1cos
transformed in the following way:
d 1 P 1
With a help of inequation
dt
mV cos cos
sin 2 cos 2 sin 2 .
2 H / 2 0, 2 H / 2 0 we will establish the
membership of angles and to one of the two quadrants:
As the expression 1 / cos has the same sign as
cos sign1 , (11) cos (q. v. (12)), the inequality is true ψ1≥0 with P≠0.
The sign of the variable ψ1 with P=0 depends on the sign
of the constant ψ2 =C*, which is determined from the
cos sign 1 cos 2 sin
V following considerations. In order to transfer SC from earth
satellite orbit to a descending trajectory, angle α in the
sign 1 . (12) process of powered flight should be in the range of
cos / 2 . Then from the equation (11) we obtain
The engine thrust possesses the boundary values: that ψ1 ≤0, and with the help of equation (9) determine that
P Pmax with F1 0 and ψ2=C*≤0 Hence, ψ1≥0 during the coast flight. At that, the
analysis of dependence ψ1 showed that the conjugate
P0 with F1 0 13 variable ψ1 jumps at the moment of engine thrust P
Let’s prove that there are no more than two powered switching. Hence we can conclude that the variable ψ1
flight phases. during all the considered flight phase of SC changes its sign
no more than two times.
The expression has a switchover function in the thrust
Thus, considering the condition (14) we may conclude
optimal control. In order to determine the number of burns it
that the maximum number of zeros of function F1, as well as
is necessary to examine the function F1, which, according to
the number of SC engine thrust switchings, equals two.
[4], is calculated from the equation Vent and Lent,
Besides in this case the switchings are made from P=Pmax to
1
F1 ( 1 cos cos 2 sin cos P=0, and then to P=Pmax again.
m V Thus, the laws of thrust vector optimal control are
characterized by the dependencies (9), (10), (13). In order to
3
sin ) obtain the numerical evaluation of optimal trajectory it’s
V cos necessary to solve the boundary value problem which
Considering the formulae (9) and (10) we get consists in iteration of 10 N(=10) parameters at the
1 1 beginning of a trajectory (seven-parameters vector of
F1 . (14) conjugate variables ψi0 and values of control functions (P0,
m cos cos α0, β0), with which the final edge conditions are fulfilled.
Some boundary values of conjugate variables can be
IV. ANALYTIC ALGORITHM OF OPTIMAL MANEUVER obtained on account of transversability condition [4]:
CALCULATION
[ 1 1 m H t 3
In order to investigate the behavior of curve F1 we
transform the mathematical model of SC motion, as the 5 6 ]tt0k 0. 15
equation (14) cannot be solved analytically. Let’s consider As the variations δε and δφ with t=t0 are interdependent
that SC flight with the running engine is determined only by ones, the transversality condition is fulfilled in case if the
active forces and during the coast flight – only by following equality occurs
gravitational forces. Suppose that angles α and β, as well as
trajectory angle θ change insignificantly during the powered 30 60 0. (16)
flight. Then the differential equations for conjugate variable On the other hand, the variations δε and δφ are related by
ψ1, influencing the function F1, take the form: q q
0
where q cos cos C .
angle with the increase of |θent|. The start of 2nd burn is The results presented on Fig. 2, demonstrate specific
chosen so that its finish will correspond to the moment of influence of variable parameters (h0, m0, Vent, θent) on the
SC reentry (h=hent=100 km). As one would expect, the finite mass mf and on the efficiency of mass increase δmf
duration of the 1st burn is considerably greater than of the using two-burn SC deorbiting pattern. Considering them one
2nd burn. can see that the finite mass mf is increasing with the altitude
Application of such algorithm will not lead to the increase of earth satellite orbit h0, initial mass m0 and with
considerable increase in fuel consumption in reducing of SC reentry velocity Vent and absolute value of
comparison with mFmin: the differences do not exceed trajectory angle θent. Thus, the change of h0 from 400 to 600
1÷2%. The supposed control will be called suboptimal km leads to increase of mf from 1265 to 1380 kg, mass
control. change m0 from 1,5 to 2,3 t leads to increase of mf from 995
m, kg to 1660 kg, reducing of reentry velocity Vent from 7,6 to 7,2
km/s leads to increase of mf from 1310 to 1325 kg, change
of trajectory angle from -150 to -8,50 – results in the increase
of mf from 1110 to 1490 kg (Lent=300km).
mf The power characteristics of engine P and Pspec, reduced
frontal surface load Px and inclination of earth satellite orbit
ΔmF1 i0 scarcely affect the finite mass of SC.
The efficiency of δmf increase due to application of the
500 proposed scheme is provided in all the variation range of
ΔmF2 variable parameters, presented on fig. 2. Besides the high
intensity of mf increase is revealed for greater values m0,
absolute values θent and smaller altitudes h0: δmf achieves ~
0 270 kg.
Lent, km/s
150 300 450
Fig. 1. Dependencies of SC final mass and mass ΔmF1 VI. CONCLUSION
and ΔmF2 on the lateral displacement Lent (the variant of
nominal initial data) Thus the represented materials show the possibility and
Solid lines – suboptimal control; big power gain of SC two-burn deorbiting pattern in the
dashed lines – the one-burn scheme wide range of boundary conditions, design, mass and power
Figure 1 presents the dependencies of SC finite mass mf, characteristics of SC and engine. It should be noted that the
spent at the 1st and 2nd burns mF1 and mF2 , on the lateral proposed methodological approach can be applied also for
displacement Lent. For comparison there are the results of the solution of tasks of thrust vector optimal control during
fuel mass calculation mF1, necessary with the use of interorbital maneuvers and correction for maintenance of SC
spacecraft one-burn deorbiting scheme and corresponding to orbital parameters in the specified limits.
the spacecraft finite mass mf. Analyzing this data, we should
VII. LIST OF REFERENCES
note a considerable efficiency of the proposed deorbiting
scheme. Thus, the increase of SC finite mass δmf is ~ 180 [1] Avduevsky V.S., Antonov B.M., Anfimov. N.A. and
kg. others. The theory of space flight, M.:
As we should expect, with approximately optimal control Mashinostroyeniye, 1972, p.345.
the fuel mass mF1 monotonically increases with the [2] Ivanov N.M., Dmitrievsky A.A., Lysenko L.N. Ballistics
increase of value Lent: the change of Lent from 0 to 600 km and navigation of spacecrafts. M.: Mashinostroyenie.
leads to increase of mass mF1 from 430 to 660 kg. The fuel 1986, p.345.
mass mF2 changes less: in the same variety range Lent mass [3] Pontryagin L.S., Boltyanskiy V.P., Gamkrelidze R.V.,
Mishchenko E.F. Mathematical theory of optimal
mF1 reduces from 190 to 160 kg. On the whole for nominal
processes. M.: Science. 1969, p.393.
values of variable parameters (24) the SC finite mass is ~
[4] Letov A.M. Flight dynamics and control. M.: Science.
1200-1400 kg.
mf, kg
1969, p.360.
1500 h0
[5] Elyasberg, P. E. Introduction to the theory of flight of
artificial earth satellites. M.: Science. 1965, p.537.
1300
Vent
1100
m0 θent
900
400 500 600 h0, km
-8 -11 θent, deg -15
1,5 2,0 2,5 m0 ,m
The reason of this replacement took place because variable r output that is generated every time returns the three most
is considered as a very important factor in the semantic theory dominant words that are considered as keywords for every
title. In the next step we test the output in order to measure our
[17]. The replacement was made deliberately since vector Vi experiment success rate. We search in Google Scholar (in title)
represents the influence’s degree through variable r because using the exact output of the algorithm and carry out a
this will lead in a new ranking procedure according to its statistical research of the total results. In order to be more
influence. precise we will present in detail an example/sample taken from
U i = [ ri wi ]
our research. The whole process begins once we have selected
si (8) a title of an article, which was obtained by using Google
Finally, in both cases the relevance rankings of documents Scholar. Our search term was “Knowledge” and the title
in a keyword search can be calculated, using the assumptions obtained “Advances in knowledge discovery and data mining”.
of document similarities theory, by comparing the deviation of Now, the obtained title must be preprocessed, before applying
angles between each document vector and the original query the algorithm. By using POS software, we keep words that
vector where the query is represented as the same kind of based on the part of speech they belong they can be considered
vector as the documents (see fig. 1). as valid keywords. All stop words must be cleared and also
part of speech that for the purposes of this research must be
excluded. Therefore, from the original title what is left is
“Advances knowledge discovery data mining”. At this point
we can apply the algorithm to the title that is left in order to
export the three most dominant keywords of the sentence.
After the algorithm is applied, we continue the experiment by
testing whether our results are successful. To do so, we must
visit Google Scholar again and by searching only in title, with
the three most dominant keywords we check the results. The
first result we find out of 181 results is the original title.
5 1 1
Table I. The Features of Vw
6 0 0
Words Order No. Strength
7 0 0
Characters
Advances 1 8 805 8 0 0
knowledge 2 9 960 9 0 0
discovery 3 9 984 10 1 0
data 4 4 410 11 2 0
mining 5 6 642 12 1 0
13 1 0
Step 3. We applied the algorithm by using the equation 3 for
14 0 0
the extraction of the resultant vector (vs) features. This
implemented by the following code of Matlab “ u=sum(d1)” 15 0 0
and the results are presented in equation 7:
16 0
5 i 5
=Vs ∑ = Vwe( )Vs( )
i
[ 0.0085 0.0204 2.1524] (7)
17 1
i =0 i =0
As result of this procedure the words “discovery, knowledge we calculate that the robust fit polyonym is n=5 degree (see
and Advances” are selected as the optimum triple keywords. fig. 2)
R2 = 0.92192
160
120
According to previous stage, 181 triple keywords are
extracted from corresponding Scholar Google titles. In the 100
2 14 12
14 -23.8887
The Sensitivity relates to the test's ability to identify positive
15 -27.5784 results.
If f ( x) is the normal distribution χ ∈ { p1 , p2,..., p17 } then we Specificity relates to the test's ability to identify negative
calculated the normal density probability function (see results. This can also be written as:
equation 10) number _ of _ true _ negatives 11
b =
specificity = = 0.58
Pr [ a ≤ X ≤ b ] =
∫ f X ( x)dx (10) number _ of _ true _ negatives + number _ of _ false _ positives 19
a=1 and b=17 (see fig. 3), then we ascertain that the most
probable value of x is 3.
IV. CONCLUSIONS
We have presented a new keyword extraction algorithm that
detects the most important terms in the titles of scientific
papers, which are then employed as search keywords for
retrieving articles form digital scientific online resources. In
particular, the core of the problem lies in the fact that it is
rather difficult to identify a single event space and probability [4] E. Frank, G.W. Paynter, I.H. Witten, C. Gutwin, C.G. Nevill-Manning.
“Domain-specific keyphrase extraction”. In Proceedings of IJCAI,
measure within which all the related random variables can be 1999, pp. 688–673.
determined. Our algorithm is based on the vector space model [5] Y.H. Kerner, Z. Gross, A. Masa, “Automatic extraction and learning of
to represent the article titles and treats every term in a title as a keyphrases from scientific articles”. Computational Linguistics and
vector, with each vector containing three features namely, the Intelligent Text Processing, 2005, pp. 657–669.
[6] Liu F. Liu, Y Liu. “Automatic keyword extraction for the meeting
number of characters a term has, the importance of the term in corpus using supervised approach and bigram expansion”. In
the title and the relative order of the term in the title. In the Proceedings of IEEE SLT.
classical vector space model theory, using the TF/IDF [7] P. Turney, “Coherent keyphrase extraction via web mining”. In
Proceedings of IJCAI, 2003, pp. 434–439.
algorithm, the document vector is extracted by using [8] G. Carenini, R.T. Ng, X. Zhou, “Summarizing emails with
multivariate components j, while the classical vector has j conversational cohesion and subjectivity,” in Proceedings of ACL/HLT,
dimensionality. 2008.
[9] A. Janin, D. Baron, J. Edwards, D. Ellis, G. Gelbart, N. Norgan, B. Pe-
Based on the weight our algorithm computes for each of the skin, T. Pfau, E. Shriberg, A. Stolcke, and C. Wooters, “The icsi
above features it assigns every term in the article’s title with a meeting corpus,” in Proceedings of ICASSP, 2003.
score indicating the suitability of that term as a search keyword [10] Matsuo, Y., Ishizuka, M. “Keyword extraction from a single document
using word co-occurrence statistical information”. International Journal
that could retrieve its corresponding article in the top ranking
on Artificial Intelligence Tools. 2004. Vol 13, pp. 157–169.
position. The experimental evaluation of our proposed [11] Hulth, A., “Improved automatic keyword extraction given more
algorithm regarding real data proves its effectiveness in linguistic knowledge”. In Proceedings of the 2003 Conference on
detecting the most suitable keywords in the articles' title and Empirical Methods in Natural Language Processing. 2003. Association
for Computational Linguistics, pp. 216–223
indicates that title terms may be sufficient for representing the [12] P. Soucy και G. W. Mineau, ‘Beyond TFIDF weighting for text
article semantics, when it comes to scientific publications, categorization in the vector space model’. IJCAI, 2005, Vol. 5, pp
which is to be investigated in the future. Currently, we are in 1130–1135.
[13] Robertson, S. “Understanding Inverse Document Frequency: On
the process of enriching our algorithm with additional features theoretical arguments for IDF”. Journal of Documentation, 2004, Vol.
from the articles' contents such as the abstract and the 60 (5), pp. 503–520.
keywords authors indicated for their articles. Moreover, to [14] C. Blake, ‘A comparison of document, sentence, and term event spaces’,
In Proceedings of the 21st International Conference on Computational
address the issue of synonyms and polysemy, that are not Linguistics and the 44th annual meeting of the Association for
included in this experiment. In addition, we plan to employ the Computational Linguistics, 2006, pp. 601–608.
identified keywords for building an ontology that could be [15] M. Poulos, S. Papavlasopoulos, V. Chrissikopoulos, “A text
employed as a repository of search term for conducting text categorization technique based on a numerical conversion of a symbolic
expression and an onion layers algorithm”. Journal of Digital
retrieval tasks in scientific online repositories. Information, 2006, Vol. 1 (6).
Furthermore, in order to evaluate the sensitivity of our [16] Salton, G., Wong, A., Yang, C. S. “A vector space model for automatic
method we consider that the success scores are ranged between indexing”. Communications of the ACM. 1975. Vol. 18, pp. 613–620.
[17] Harispe S. et al. “Semantic measures for the comparison of units of
1 and 3 interval space, such as obtained by the probability language, concepts or entities from text and knowledge base analysis”.
density function. Arxiv. 2013. Vol. 1310, 1285. pp. 1-159.
Finally, we intend to release the ontology that will be [18] Parzen, E. “On estimation of a probability density function and mode”.
The annals of mathematical statics. 1962. Vol. 33, pp. 1065-1076
developed so that it could be employed by other researchers [19] Abramowitz, M., Stegun, I. A. (Eds.). "Probability Functions." Ch. 26 in
for performing semantic retrieval tasks. Entropy and mutual Handbook of Mathematical Functions with Formulas, Graphs, and
information in turn enables us to define a distance measure Mathematical Tables, 9th printing. New York: Dover, 1972, pp. 925-
964,
formally. With this distance a sound foundation is given for
[20] M. Poulos, G. Bokos, N. Kanellopoulos, S. Papavlasopoulos and M.
the capturing of the inherent structure of ontology. Avlonitis. “Specific selection of FFT amplitudes from audio sports and
Consequently, in our work, we attempt to use a new algorithm news broadcasting for classification purposes”. Journal of Graph
which will be created in the philosophy of TFIDF algorithm Algorithms and Applications. 2007. Vol. 11(1). pp. 277–307
giving simultaneously a solution to Shannon-like model
disharmony.
REFERENCES
[1] H. Wang, S. Liu, L.T. Chia. “Does ontology help in image retrieval? A
comparison between keyword, text ontology and multi-modality
ontology approaches”. In Proceedings of the 14th Annual ACM
International Conference on Multimedia. 2006. pp. 109–112.
[2] Y. Matsuo, M. Ishizuka, “Keyword extraction from a single document
using word co-occurrence statistical information”. International Journal
on Artificial Intelligence Tools. 2004. Vol. 13, pp. 157–169.
[3] M. Rajman, R. Besancon “Text mining – knowledge extraction from
unstructured textual data”. In Proceedings of the 6th Conference of
International Federation of Classification Societies. 1998.
Abstract- This text generally describes the architecture of Current knowledge have the task of describing the
an agent-based model for pulmonary tuberculosis in the complexity of the world we inhabit, so in this case it is
zone of Usme (Bogotá, Colombia), product of a master's intended to represent the nonlinear and interdependent
thesis in Information Sciences and Computing of District relationships of a community in the middle of a tuberculosis
University "Francisco José de Caldas" [1]. First comes the epidemic, in order to predict the scenarios of this complex
introduction, then the tools and concepts that allow the system [4, 5], explaining the construction of a Agent-Based
simulation, then the model is exposed from a conceptual Model for Tuberculosis in Usme’s zone, and thinking about
approach to technology-society relationship, and finally it is the origin and social horizon of technology.
disclosed the author’s findings.
II. TOOL AND BASIC CONCEPTS FOR THE MODEL
Key words: Agents-Based Models, Geographic Automata
System (GAS), Epidemiologic simulation, complex
A. Geographic Automata System (GAS)
Systems.
Between 1999 and 2001, Paul M. Torrens and Itzhak
I. INTRODUCTION
Benenson created Geographic Automata System (GAS) for
modeling phenomena on real spaces, from individual
The technological edge thinks forward, looking for ways to computer entities within the system: agents. On the one hand,
understand it, study and intervene in the world over time they modified the classical Cellular Automata (CA), which is
towards a particular purpose. So it makes models that attempt a system in one, two or three dimensions, consisting of
to reproduce actual facts and allows a better understanding of partitions called cells, which acquire ways to present: states,
the relationship of humans with them, appearing technologies from a default set of them, ranging from relations with its
with that purpose as Agent-Based Models and Geographic neighboring cells, their neighborhood (Figure 1), through
Automata System, among others [2]. pre-established transition- state rules, in a sequence of
moments called evolution system [6, 7].
But the future is thought from the uncertainty, because the
world is complex and multiple, set of systems with the general And on the other hand, they placed those entities that change
characteristics of not to be reversible, not to be accurately over time (A. Figure 2) on a Geographic Information System
predicted, to develop nonlinear interdependent relationships, (GIS), which is a set of layers that describe the geographic
and to have appearance of order (emergency), being characteristics of a place, from the general to the specific
necessary to go beyond the deterministic mathematical formal (Figure 2 B). Thus, agents are related to real spaces,
schemes [3, 4]. preserving the neighborhood concept in his role for the
change of state (Figure 1), but not only on generic objects
The bio-social systems, being complex, cannot be predicted such as cells, but sometimes on entities with own
with accuracy nor are reversible, so they have a constitution characteristics and mobility: specific sites , people, vehicles,
order from the simple to the complex, allowing the passage etc. (C. Figure 2) [2, 6].
from chaos to order at the time called edge of chaos, where
self-organization emerges, which is the coordination of the
parts’ behaviors of a system without any central power or
external coercion that lead them. Jhon Stwart Kauffman
formulated a hypothesis of a global connection between all
parts of a physical system, that after a certain time, due to the
energy accumulated between these, by inertia was that these
associated themselves synergistically and generate patterns Fig 1. Types of neighborhood in cellular automata which allows the
(self- adaptive systems) [3, 4]. evolution of the system. Neighborhood A: Von Newman neighborhood, with
octagonal neighbor cells. Neighborhood B: Moore neighborhood, with
octagonal and diagonal neighbor cells [5].
1
Department of Academic Research, “Manuela Beltran” University. A Geographic Automata System (G), is defined as the set of
gabrielmoreno10@gmail.com automatas (K), which vary from state (S) over time, through
2
Department of Academic Research, UNIMINUTO. transition rules (Ts), developed at specific locations (L), from
wpenapena1@hotmail.com
mobility rules (if any) of each one (ML), considering its mostly with low levels of schooling and informal jobs. In
neighborhood (N) at every time, called Tick, and the criteria addition, from 1995 to 2005, amidst the paramilitary
of relationship they have with this (RN) (Figure 3) [7, 8]. phenomenon nationwide, this zone received 8.2% of the
Where, thanks to this set, It is able to see the interdependent displaced population that came to Bogotá by the violence
relationships between the conditions of each area (population [12].
density, healthiness, etc.), and each agent (location, nutrition,
origin, economic status, etc.) [9]. Thus, according to the Unsatisfied Basic Needs Indicator
(NBI, by the Spanish acronym), which considers the
shortages at home: a) housing with physical or structural
defects, b) lack of basic services or deficiencies in drinking
water and feces disposal, c) overcrowding (2 persons/5m2),
d) high economic dependence (1 productive person / 3 or
more dependents), e) truancy (at least one child between 7 and
11 do not regularly attend a school), and f) Misery, when the
home has two or more of the above conditions, it was found
that Usme was 9, 1% of homes in NBI, with 1% in misery.
Fact also reflected in the 51% of the population below the
"poverty line" indicator that arises from considering the
minimum subsistence income for a person [13].
amount of Tick equivalent to every hour (1 hour = 1 Tick) , attribute listPerson created an arrangement for certain amount
the number of initial agents in each state, that the of agents inside, regardless of their type, verify with the
epidemiological model of pulmonary tuberculosis (SIR method getOvercrowdingCondition() the overpopulation
model) are: Susceptible (which could be infected), infected (boolean data), on the basis of the capacity of the place
(disease carrier) and Recovered (who overcomes the (method getPeople Capacity()), and the number of people in
infection), and the amount of each type of agent , which under it (method getAmountPerson()) [1].
Usme socioeconomic conditions are: Housewife, Worker,
Student, Deplaced and Homeless, which enter as parameters
via setValuesFromParameters() method of that class [1] (A.
Figure 4).
B. Contexts
Due to the needed for information on the TB outbreak Fig 4. Time and space model, and training contexts through parametric
(major focus of infection, more infected population trend of classes.
spread of the disease, etc.), which would act on this, the
dynamic generation of contexts allowed the development of The specific locations and attributes were defined with
the non-linear relationships between the different agents in a HomePlace classes (households), WorkPlace (work sites),
period of time, loading and processing the increased volume StudyPlace (study sites), EntertainmentPlace (entertainment
of information in the model thanks to the architecture of the venues) and Route (tracks), which inherit of the class Place
program, through reflection, extended the possibilities of attributes as the maximum area place (MaxArea...), high
development of the system. capacity (MaxAmount...), many people (amountPersons),
capacity (peopleCapacity), use, management, geometry,
C. Places number of people per state (amountSusceptible,
amountInfected, amountRecovered), among others.
Parametric classes system can evolve on the geographical
conditions of Usme, which is very important for the Calculation of amounts by State was performed on the
relationship between the demographic conditions of a region method step() -different class TimeProject step() method, and
and the spread of an epidemic (TB in this case), where the later exposed the Person class step() method-, and the
different types of places and their characteristics were distribution of agents by States was random from classes of
determined and verified with the abstract class Place, with the each place: Work Place, Study Place, Home Place,
Entertainment Place, trough the method 'Housewife', 'Student', 'Deplaced for violence' and 'Street
getRandomStudyPlace, getRandomHomePlace, people', through classes, Worker, Housewife, Student,
getRandomWorkPlace, and getRandom EntertainmentPlace, Deplace and Homeless, respectively.
respectfully (A. Figure 5). Previous classes inherit from the Person class attributes of
localization in each place (currentLocalization), routes of
The description of Usme geography through Shape files and movement (route), location at every moment of the system
their inclusion in the system, allowed to generate scenarios (currentTick), identification (id), employment status, medical
appropriate for studying the TB epidemic in the actual service, stratum, previous infection by HIV and TB, type of
conditions of this place, as a requirement of a model oriented person, State of health (healthFactor) and number of health
to care for the life of the population, through the generation (numberHealthFactor), which is verified by the
and the study of patterns in the spread of this disease. verifyHealthFactor() method According to the infections
activate method infect(), which passes the person State
Susceptible to infected (first passive, then contagious).
E. Rules of transition
With a total amount of 500 agents, 20% love House, 3) The model is created for the study of infection and the
workers 25%, 25%, 15% and 15% displaced students of spread of tuberculosis taking into account the social
street, 5% initial of infected, recovered 20% and 75% dynamics, adopt to different fields and learn behaviors and
susceptible, were obtained data from Table III. theoretical but reliable of a possible epidemic statistics
without ethical implications, and allowing the application of
TABLE III: AGENT TYPES OF INFECTION BY TICK. preventive methods and control large scale.
[12] Secretariat of estate of the District. Crossing Usme 2004. In: physical
Physical and socioeconomic Diagnosis of the localities of Bogota, D.C.
Major mayoralty of Bogota. Bogota DC, Colombia. 2004.
[13] Secretariat of Health of the District. The health and the quality of life.
Locality 5 - Usme. Major mayoralty of Bogota DC. 2009.
[14] Dr. S. Invertrz. Pathogeny of the tuberculosis. University of Buenos
Aires. Buenos Aires, Argentina. 2008.
[15] Dr. Ospina S. The tuberculosis, a historical - epidemiological
perspective. Magazine Infectio. Vol.5 no. 4 pp. 241-250, Bogota October -
December, 2001.
Authors
I. INTRODUCTION
B. Scenario II
The second scenario of fire considers the fire acting in the
middle of the opening on a strip of 0.50 m wide, aside the
whole width of the flanged beam (Fig. 14).
Fig. 10: Temperature distribution on concrete after two hours of fire,
Tmax=1052ºC
C. Scenario III
In the third scenario of fire it have been considered acting
on a strip of 0.50 m wide near the support, aside the whole
width of the flanged beam (Fig. 23).
Fig. 19: Temperature distribution on concrete after two hours of fire,
Tmax=978.7ºC
IV. DISCUSSION
The three scenarios taken into consideration in the
investigation presents possible situation of fire acting on the
beam. When comparing results obtained for flanged wide
reinforced concrete beam with respect to the independent wide
reinforced concrete beam [5] one can remark similar behavior
Fig. 28: Temperature distribution on concrete after two hours of fire, of the beams under the same external load, but deflection of
Tmax=969.4ºC the beam, internal stresses in concrete and reinforcements as
well as internal temperature are of more reduced values.
Comparison of the results for the three scenarios is
unnecessary since the load given by the fire in the first
scenario is incomparable with the other two scenarios. Even so
we can remark the increased risk for the stability of the
element for scenario I since after one hour it reaches excessive
deformation (beyond l/100), where the concrete cover of the
tensioned reinforcements is already inexistent. For scenarios II
and III the deformation occurred is almost negligible under the
external and fire loads.
Fig. 29: Temperature distribution on steel after two hours of fire,
Tmax=908.4 ºC
REFERENCES
[1] *** SR EN 1992-1-1-2004, Eurocode 2, Design of concrete structures.
Part 1-1: General rules and rules for buildings, 2004.
[2] *** SR EN 1992-1-2-2004, Eurocode 2, Design of concrete structures.
Part 1-2: General rules – structural fire design, 2004.
[3] ***. (1999). P118-99: Normativ de siguranță la foc a construcțiilor
[4] A. Puskas, Z. Kiss, “Testing of a wide reinforced concrete beam”, The
7th Central European Congress on Concrete Engineering,
Balatonfüred, Hungary, 22-23 September 2011, pp. 315-318
[5] A. Puskas, A. Chira, “Numerical Investigations on a Wide Reinforced
Concrete Beam Subjected to Fire”, Proceedings of the 4th International
Conference on Mathematical Models for Engineering Science - MMES
'13, Brasov, Romania, June 1-3, 2013, pp. 169-174, ISBN: 978-1-
61804-194-4
[6] Z. Kiss, K. Bálint, A. Puskás, “Steel or concrete structure –
prefabricated or cast in situ? The design of a multistory building in
Bucharest for Kika”, III, Medunarodna Savetovanke, Subotica, 8-9.
Octobar 2009, p. 79-93.
[7] Abaqus. Abaqus Analysis User's Manual.
[8] I. Moga, L. Moga, “Heat flow simulation through the window together
with the wall in which is fitted in”, IASTED conference “Applied
Simulation and Modeling” Palma de Mallorca, Spain, 29 – 31 August,
2007, ISBN: 978-0-88986-687-4
[9] A. Faris, A. Nadjai, S. Choi, “Numerical and experimental
investigations of the behavior of high strength concrete columns in fire”,
Elsevier, Engineering structures, 2010.
[10] K. Venkatesh, R. Nikhil, “A simplified approach for predicting fire
resistance of reinforced concrete columns under biaxial bending”,
Elsevier, Engineering structures, 2012.
[11] Qing-Hua Tan,Lin-HaiHan n, Hong-XiaYu, Fire performance of
concrete filled steel tubular (CFST) column to RC beam joint, Fire
Safety Journal, 2012
[12] Anil Agarwal , Lisa Choe , Amit H. Varma ,,Fire design of steel
columns: Effects of thermal gradients’’ Elsevier, Journal of
Constructional Steel Research, 2014.
cause difficulties if we want building executable code for all modeling and simulation results for different tools if they
permissible current systems of equations for whole model follow Standard.
beforehand (Fig.1). Standard should touch on classification of permissible:
• numerical methods,
• and instruments for computational experiments
x1 = Sub( x1 )
LAE: .
A ⋅ x 2 = b( t ) + G1 ( x 1 )
x1 = Sub( x1 )
ODE_general: dx 2
F( x 1 , , x 2 , t) = 0
dt
Fig 3. Modelling in OpenMVLShell. x1 = Sub( x1 )
For building and playing with Model: ODE_normal: dx
2
= F( x 1 , x 2 , t )
dt
• download a Model written in MVL (any Model Vision
tool can save graphical specification of a model as x1 = Sub( x1 )
MVL-text); ODE_linear: dx
• build executable Model with help of OpenMVLShell;
2
= A ⋅ x 2 + b + G( x 1 )
• call executable Model. dt
4. Systems of algebraic-differential equations:
x1 = Sub( x1 )
dx
DAE_semi-explicit: 2 = F( x 1 , x 2 , x 3 t )
dt
0 = G( x1 , x 2 , x 3 t )
x1 = Sub( x1 )
dx
DAE_semi-explicit_1: 2 = F( x 1 , x 2 , x 3 t )
dt
A ⋅ x 3 = b + G1 ( x1 , x 2 )
x1 = Sub( x1 )
Fig 4. A numerical experiment in OpenMVLShell.
Component «Solver» calls numerical methods in compliance DAE_semi-explicit_2: dx 2
with results of syntactic analysis of equations written by User. B ⋅ = A ⋅ x 2 + F( x 1 , t ) + b
In the rest of the paper:
dt
t ∈ ℜ1 (time); x ∈ ℜ n (unknowns );
Sub( x ) : ℜ n → ℜ n ( substitutions );
F, G : ℜ m → ℜ k ;
A, B, b − real matrices and vectors.
REFERENCES
[1] Popper N., Breitenecker F., Extended and Structural
Features of Simulators. Simulation News Europe, 2008 - Dec.,
pp.27-38.
[2] Isakov A.A., Senichenkov Yu. B. OpenMVL is a tool for
modelling and simulation. //Computing, measuring, and
control systems, St. Petersburg: SPBSPU. - 2010 - pp. 84-89.
[3] Isakov A.A., Senichenkov Yu. B. Program Complex
OpenMVL for Modeling Complex Dynamical Systems //
«Differential equations and control processes» Journal,
Mathematics and Mechanics Faculty of Saint-Petersburg State
Fig 5. Possible structures of solved systems. University, St. Petersburg - 2011.
Effectiveness depends on structure of solved system. [4] Isakov A.A., OpenMVLShell – research tool for
OpenMVLShell distinguishes between sparse and non-sparse modeling and simulation. // All-Russian competition for
systems, and systems with the special structure (Fig. 5). students and post-graduate students "Innovation technologies
Solvers of systems of linear algebraic equations take into
in Education", Belgorod state National University - 2011 - pp.
account their sparseness.
«Solver» for NAE, ODE, and DAE can call «automaton», 11-15.
standard (built-in), users’, and debugging methods (Fig. 6). [5] Fritzson P. Principles of Object-Oriented Modeling and
«Automaton» is used on default. The goal of «Automaton» is Simulation with Modelica 2.1, Wiley-IEEE Press, 939 pages.
to find solution at any cost. It may be a set of sequentially [6] Fritzson. P. Introduction to Modeling and Simulation of
executable software implementation of numerical methods.
Technical and Physical Systems with Modelica 2011. Wiley-
IEEE Press , 232 pages.
[7] https://www.modelica.org/publications
[8] Hairer E., Wanner G. (1996), Solving oridinary
differential equations II. Stiff and differential-algebraic
problems, Springer-Verlag, Berlin.
[9] A. C. Hindmarsh, "ODEPACK, A Systematized
Collection of ODE Solvers," in Scientific Computing, R. S.
Stepleman et al. (eds.), North-Holland, Amsterdam, 1983 (vol.
1 of IMACS Transactions on Scientific Computation), pp. 55-
64.
Fig. 6. Structure of a «Solver».
[10] http://www.netlib.org/ode/ddassl.
OpenMVLShell and all MvStudium Group tools have two- [11] http:/www-users.cs.umn.edu/~petzold/DASPK
level numerical library. Lower level contains software
implementation of numerical methods for NAE, LAE, ODE,
Isakov Alexander A. -- MPhil (Computer Science, 2012, SPBSTU),
DAE written in Fortran. postgraduate student of Distributed Computing and Networking Computing
NAE: Newton and Powell methods. department, Saint Petersburg state Polythecnical University.
LAE: LINPACK solvers for non-sparse systems and systems Senichenkov Yuri B. – DPhil (Numerical software, 2005, SPBSTU),
with special structure (band and so on); Sparse Solvers. Professor of Distributed Computing and Networking Computing department,
Saint Petersburg state Polythecnical University.
ODE: Solvers from ODEPACK [9]; Hairer and Wanner
Solvers from [8].
DAE: ODEPACK, DASSL [10], DASPK [11].
IV. CONCLUSION
Number and complexity of tools for modeling and simulation
of complex dynamical systems increases permanently. They
Abstract—The typical design of the weapon system using II. PROBLEM FORMULATION
separated propellant charge is not beneficial enough to the muzzle The ballistic performance of the weapon could be measured
velocity because of the ignition delay of additional charge. The new
method of the precisely timed electric ignition has been developed. by the kinetic energy of the projectile. If the energy
Contributions of the timed ignition to the ballistic output of weapon conservation equation (1) is observed and the work done by
have been modelled and model has been supported by experimental propellant gases is expressed, the weapon performance could
shootings results. The muzzle energy of the projectile fired with be enhanced by increasing the ballistic pressure of propellant
timed ignition has been increased by more than 25 percent compared gases accelerating the projectile:
to the projectile fired from typical weapon. lú
1
s p dl mq vú 2 , (1)
Keywords—Interior ballistics, gun, separated charge, muzzle 0
2
velocity, timed ignition, electric ignition. where cross-section surface is represented by s, ballistic
pressure by p, the l does stand for projectile travel, lú for
I. INTRODUCTION barrel length, mq is mass of projectile and vú is projectile’s
The described process of ignition of additional propellant is rate of the propellant gases through the ignition channel and
significantly time-consuming and does impede full utilization the generation of the propellant gases mixtures.
of system with separated propellant charge (see Fig.4). If the The flow rate of the propellant gases mpr is dependent on
ignition of additional propellant charge could be timed the difference in the internal pressure in barrel bore p and
precisely to the moment when projectile does pass the flash internal pressure in the additional chamber p1. If the
hole, the obstacle of ignition delay would be removed. condition
Therefore the pressure course should be improved and higher 1
2
muzzle velocity should be produced. p1 p (9)
2
III. MODEL is met, the propellant gases will flow by the critical velocity
and flow rate will be
Following the objective of physical action description the
p , (10)
mathematical model was assessed. The core of the model of is m pr k
T
based on the thermodynamical description of geometrical f
Tv
theory of propellant combustion consisting of following
equations: where the outflow coefficient ξ* is given by
2
z z z ,
2 3
(2)
2 2 .
1
(11)
f
1
mq v
2 2
p 2 pz , (3) If the condition (9) is not met, the coefficient ξ* is
s l l substituted by the outflow coefficient ξ given by
dv
2 1
p 2 1 2
mq s p pz , (4)
1 1 .
p 1 (12)
dt
p p
u1 m p pz
dz
, (5) The coefficient κk defines local flow conditions of the
dt Ik
particular channel design, it has been assessed according to
l l0 1
1
, (6) the Cibulevky theory.
The equations for mathematical description of the
propellant gases mixture are based on the Dalton’s law
dl
v, (7) p = ∑pi , the total pressure exerted by the mixture of non-
dt reactive gases is equal to the sum of the partial pressures of
1 v ,
2
individual gases. When βi represents partial mass of i-th
T TV 1 (8)
vlim portion
where ψ stands for the relative quantity of burnt-out i i , (13)
powder, κ, λ, μ are geometric characteristics of the powder
then the mixture could be characterized by the quantities:
grain, z is the relative burnt thickness of the powder grain, p
1
is the ballistic pressure and pz is the primer pressure, f is the f
i i
fi , (14)
specific energy of propellant, ω is the mass of the propellant
charge, mq is the mass of the projectile, φ is the fictivity f , (15)
coefficient of the projectile, s is the cross-section area of bore, f
m, u1, ν are rate of burning coefficients, Ik is the pressure
i i i
i
pressure [MPa]
numerical method only. The Runge-Kutta 4th order method 40
has been used and the appropriate procedures were built
30
under the environment of Matrix Laboratory software.
Parameters and constant values were given by 20
characteristics and properties of the gun and ammunition
10
used during following experimental shootings. The calculated
pressure course of chosen gun fired as an ordinary weapon 0
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5
(with basic propellant charge only) is illustrated in following time [ms]
40
80
30
70
20
60
10
50
pressure [MPa]
0
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5
time [ms] 40
20
The comparison of the numerical model results for the
chosen case of weapon system designed as a weapon system 10
60
Table I. – Calculated values of velocities and maximal pressures.
50
pressure [MPa]
40
30
20
10
0
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5
time [ms]
IV. EXPERIMENT The basic charge was placed inside the brass case adjusted
to the pressure measurement according to C.I.P. standards.
A. Weapon
The propellant was 7-hole single based deterred powder. The
The experimental weapon was the modified smooth-bored
mass of the basic charge was set to constant 9 g. Every charge
barrel of calibre 12.7×107 and length of 100 bores. The three
was reweighted.
additional chamber were placed alongside the bore, just first
The additional propellant charge was integrated into the
was utilized.
screw of additional chamber. It consist of electric initiator F3
(EMS PATVAG) in the hollow volume of the support casing,
composite contact casing of crezol-formaldehyde resin, and
the assembly of contact screw placed inside the composite
pin. In the case of ‘gas ignition’ the assembly without electric
equipment was used.
C. Arrangement
Fig.8 – The additional chamber. The arrangement of the measuring chain is depicted in the
Fig. 12. The action started by the time delay setup on the
The barrel was fit to the universal breech UZ-2000 and gun timing unit from personal computer by the means of the
carriage STZA providing remote initiation. developed digital interface. Then the sufficient electric charge
B. Ammunition was stored in the capacitive high-voltage unit. The
The experimental ammunition consist of projectile, basic mechanical initiation of the basic propellant charge caused
charge and additional charge. Monolithic brass projectiles the rise of the ballistic pressure. The value of the pressure was
were used instead of the common line production because of measured by the piezoelectric sensor. The charge signal was
the mass variation. transformed to voltage signal and observed by the timing unit
in real-time. After the pressure threshold had been reached,
the timing unit counted delay and sent the control signal to
the switch unit. The high-voltage circuit was closed and
initiator ignited the additional propellant. The laser gates
were used to measure projectile velocity and ballistic analyzer
recorded signals.
Fig.12 – The measuring chain (PC-personal computers, BA-ballistic Table II. – Measured values of velocities and maximal pressures.
analyzer, NZ-charge-voltage transducer, ČJ-timing unit, SJ-switch
unit, VN-high voltage unit, LH-laser gates, arrows-signal lines).
D. Results
The experimental comparison between systems of typical
gun, ‘gas ignition’ and ‘timed ignition’ has been made. The
same types and amounts of propellants were used. The
pressure courses are compared in the Fig. 13 (refers to
Fig.3-5). Analogously the ‘best solution’ pressure course was V. CONCLUSION
measured (Fig.14 does relate to Fig.6).
The timed ignition of separated propellant charge does
result in the muzzle velocity increase compared both to the
typical gun and to the gun utilizing the gas ignition of the
separated propellant charge. Although the reliability of the
model is limited, the effect to pressure curves is in
accordance. If the more detailed description of the weapon
system with separated propellant charge was focused, it
should incorporate the hydrodynamic theory, though the
solution would become complicated due to its demands for
values acquaintance of considerable amount of unknown
quantities. The utilization of one additional chamber with
timed ignition does produce equivalent increase in muzzle
velocity compared to use of three additional chambers ignited
by hot gases [13].
REFERENCES [8] P. Oosthuisen, “Compressible fluid flow (Book style)”, New York,
[1] M. Kovarik, “Serial chamber gun (Published Conference Proceedings McGraw-Hill, 1997.
[9] E. Hairer, The numerical solution of differential-algebraic systems by
style)”, in Proceedings of the 17 th International Conference New Trends
Runge-Kutta methods. Berlin, Springer, 1989.
and Research of Energetic Materials, Pardubice, 2014.
[10] J. Kusák, Thermodynamical devices of high velocity of ejected body
[2] M. Kovarik, “Projectile velocity increase by the use of separated propellant
(Study style)”, Prototypa ZM, Brno, 2007.
charge (Published Conference Proceedings style)”, in Proceedings of the
[11] L. Jedlička, “Modelling of pressure gradient in the space behind the
4th International Conference on Fluid Mechanics and Heat & Mass
projectile”, In: Proceedings of the 7th WSEAS international conference
Transfer). World Scientific and Engineering Academy and Society,
on System science and simulation in engineering. World Scientific and
Dubrovnik, 2013.
Engineering Academy and Society, 2008.
[3] E. Davis, “Advanced propulsion study”, Warp Drive Metrics, Las Vegas,
[12] M. Hajn, “Possible advantages of the weapon system using separated
2004.
propellant charge (Published Conference Proceedings style)”, in Proc.
[4] L. Stiefel, “Gun propulsion technology (Book style), New Jersey,
Armament and Equipment of Land Forces, Lipt. Mikuláš, 2012.
AIAA,1997.
[13] M. Hajn, “Projectile velocity increase by the use of separated propellant
[5] A. Horst, “Brief Journey Through The History of Gun Propulsion (Book
style)”, Army Research Lab, Aberdeen proving ground MD, 2005. charge (Published Conference Proceedings style)”, in Proceedings of the
[6] A. Siegel, Theory of high-muzzle-velocity guns, AGARDogograph 4th International Conference on Fluid Mechanics and Heat & Mass
Vol 91, US Naval Ordnance Laboratory, Silver Spring, 1965. Transfer). World Scientific and Engineering Academy and Society,
[7] А. Златин, “Баллистические установки и их примениние в Dubrovnik, 2013.
экспериментальных исследованниях (Book style),” Наука, 1974. [14] S. Ying, “The mechanism analysis of interior ballistics of serial chamber
gun (Report style)”, Ballistic Laboratory of China, 2008.
Abstract—To investigate and analyze the obesity or overweight adolescents. In [12], this paper aimed to investigate the ability
situation among 7 to 12 year old children in 2013 and its relation to of BMI and waist circumference, single and combined, in
blood pressure, pulse, vital capacity index. To provide scientific identifying children who are at risk of hypertension and in
evidence for different level governments to make corresponding influencing absolute blood pressure values. High blood
preventative strategies and intervention measures. This paper take 7 pressure is strongly associated with excess weight. Waist
to 12 year old children in 11 primary schools as investigation objects; circumference improves the ability of BMI to identify
measure their height, weight, blood pressure, pulse and vital capacity hypertension in obese children. Waist circumference is related
index and screen the overweight and obesity according to the to absolute blood pressure values in all weight classes. In [13],
standard made by WGOC. The overweight rate of 7 to 12 city boys,
this paper aimed to determine the prevalence of overweight in
city girls, country boys, country girls in 2013 is respectively
14.54%,7.06%,7.35%,1.97%. The obesity or overweight rate of them
US children using the most recent national data with measured
is respectively 23.85%,13.02%,11.38%,3.82%. The overweight rate weights and heights and to examine trends in overweight
and the obesity rate have different trends as the age grows. The prevalence. The prevalence of overweight among children in
overweight rate and the obesity rate at different ages are different in the United States is continuing to increase, especially among
genders and town-country difference. The overweight children’s and Mexican-American and non-Hispanic black adolescents. In
obesity children’s blood pressure, pulse of are higher than the normal [14], this paper aimed to examine the extent of blood lipid
children’s, but their vital capacity index is lower than the normal abnormalities in overweight children and to determine whether
children’s. BMI is very positively related to the diastolic blood the prevalence of dyslipidemia is different in overweight
pressure and systolic blood pressure (except girls), while BMI is children with elevated blood pressure (BP) compared with
obviously negatively related to vital capacity index. The school and overweight children with normal BP (NBP). The high
the children’s parents cooperate to strengthen the food management prevalence of dyslipidemia found in this overweight sample
of obesity or overweight children, train them good dietary habits, do supports recent recommendations to collect plasma lipid levels
exercise to build their physique, improve their health level. in not only overweight children with BP≥90th percentile but
also in all overweight children. In [15], this paper aimed to
Keywords: Children obesity or overweight, physical health index, examine tracking and predictiveness of childhood lipid levels,
Western China blood pressure, and body mass index for risk profile in
adulthood and the best age to measure the childhood risk factor
I. INTRODUCTION levels. Childhood blood pressure, serum lipid levels, and body
In recent years, as the dietary pattern and life style changes mass index correlate strongly with values measured in middle
and the living standards improves, Childhood obesity is age. These associations seemed to be stronger with increased
developing in Children, which seriously influences the age at measurements. In [16], although the influence of obesity
children’s health, and it has been an important public health on ventilatory function has long been recognized, the nature of
matter. According to the survey, childhood obesity cannot only the relationship and the mechanisms are not yet clear. The
cause social psychological problems to the children [1], but purpose of this report was to examine the effects of overall
also increase the risk of obesity to them in their adulthood and obesity and fat distribution on ventilatory function. Body fat
blood pressure, diabetes, dyslipidemia and other diseases distribution has independent effects on ventilatory function
related to obesity which often happen in adulthood occur in after adjustment for overall obesity in men. The finding that
their childhood [2-5]. age modifies this association has implications for future
research.
In [11], in comparisons among age–sex–BMI percentile The paper analyzed the 7-12 year old children’s physical
groups, systolic and diastolic blood pressure values were health monitoring data in Lanzhou, gained the children’s
higher in obese and overweight groups than in normal weight overweight, obesity popularity situation among different
groups for both sexes. Although BMI among girls was higher economic social groups in Lanzhou cities and countries,
than among boys in all three percentile groups, there were no analyzes the relation between childhood obesity and blood
significant differences between sexes with respect to blood pressure, pulse, vital capacity index and then provide scientific
pressure values. The present findings emphasize the evidence for different level government to make corresponding
importance of the prevention of obesity in order to prevent preventative strategies and intervention measures.
future related problems such as hypertension in children and
overweight + obesity rate at other ages group are higher in the overweight + obesity rate , 7 to 9-year old group obesity rate is
city girls. By comparing the city girls and country girls at the greatly different from the country girls’(P<0.05 or P<0.01).
same age group, the 7-year old group city girls’ overweight rate,
TABLE I. 7 TO 12 YEAR OLD TOWN AND COUNTRYSIDE CHILDREN’S OVERWEIGHT AND OBESITY POSITIVE RATE IN LANZHOU (2013)
Boy Girl
Area Age Overweight and Testing Overwei Overweight
Testing obesity
Overweight obesity obesity Person ght and obesity
Person
16(8.47) 15(7.94
26(12.04)△ △ 46(21.30)△ △ 189 31(16.40)△ △
City 7 216 20(9.26) △ )△ △
10(5.92) 8(4.73)
29(12.24)*△ △ 51(21.52)**△ △ 169 18(10.65)△ △
8 237 22(9.28)△ △ △
11(5.58) 15(7.61
31(13.14)**△ △ 50(21.19)*△ △ 197 26(13.20)△ △
9 236 19(8.05) △ )△ △
18(11.18
44(19.47)*△ △ 69(30.53)**△ △ 161 7(4.35) 25(15.53)△ △
10 226 25(11.06)*△ )△ △
70(7.06) 59(5.95
186(14.54)**△ △ 119(9.30)**△ △ 305(23.85)**△ △ 991 129(13.02)△ △
Total 1279 △△ )△ △
15(1.85
Total 1116 82(7.35)** 45(4.03)** 127(11.38)** 812 16(1.97) 31(3.82)
)
Note: the number in ( ) is the possible rate/%. Comparison between children in different genders: Comparison between children
in town and in country: Δ means P,0.05; ΔΔ means P<0.01.
4) Children’s overweight, obesity possible rate at different
B. Comparison of Obesity-or-overweigh children’s and
age groups normal children’s blood pressure, vital capacity index
Boys’ overweight rate grows as the age increases. City
boys’ obesity rate falls as the age increases. The country boys’ The boys’ and girls’ diastolic blood pressure, systolic blood
obesity rate stays at about 4 percent. City boys’ overweight + pressure and pulse: the obesity group > the overweight group>
obesity possible rate first grows and then falls. The country the normal weight group. Vital capacity index: the normal
boys overweight + obesity possible rate first falls and then group > overweight group > obesity group. Through
grows. Through inspection, the city (F= 12.82, P=0.025), the comparison, the boys’ and the girls’ overweight group, obesity
country (F=15.42, P=0.009) boys’ overweight rate at different group and the normal weight group’s diastolic blood pressure,
ages are greatly different. systolic blood pressure, pulse ( except girls), vital capacity
index is statistically different (P<0.05 or P< 0.01).
The city girls’ overweight, obesity, overweight + obesity
possible rate all falls as the age grows. The country girls’
overweight, obesity, overweight + obesity possible rate grows
as the age grows. Through inspection, the country girls’
overweight rate (F=37.95, P=0.0000), obesity rate (F=37.95,
P=0.0000), overweight + obesity rate (F=58.91, P= 0.0000) at
different age groups are greatly different.
TABLE II. TABLE 2 COMPARISON OF LANZHOU OVERWEIGHT GROUP, OBESITY GROUP AND NORMAL WEIGHT GROUP CHILDREN’S BLOOD PRESSURE, PULSE,
VITAL CAPACITY INDEX (±S)
Boy Girl
Overweight 268 72.17±9.04** 112.54±10.13** 87.63±7.62* 40.29±8.47** 86 72.38±7.79** 111.63±9.01** 87.23±7.81 35.69±7.93**
obesity 164 75.36±8.46** 117.44±10.09** 88.99±8.04** 34.48±7.98** 74 74.63±8.25** 115.27±9.53* 88.47±10.52 29.12±4.93**
Vital capacity is one of the most important indexes to [10] Christodoulos, A. D., Douda, H. T., & Tokmakidis, S. P. (2012).
reflect the human beings’ lung function. The vital capacity is Cardiorespiratory fitness, metabolic risk, and inflammation in
children.International journal of pediatrics, 2012.
not only related to the anatomy capacity of the lungs, but also
[11] Gundogdu Z. Relationship between BMI and blood pressure in girls and
related to the ventilator function and the contractility of the boys.Public Health Nutr, 2008,11(10):1085-1088.
lungs. Obesity causes the ventilator function to fall, thus the [12] Genovesi S, Antolini L, Giussani M, et al. Usefulness of waist
blood circulation property becomes abnormal and the circumference for the identification of childhood hypertension. J
hypoxemia forms, which finally damages the ventilator Hypertens, 2008,26(8):1563-1570.
function. The research founds that overweight, obesity [13] Ogden CL, Flegal KM, Carroll M, et al. Prevalence and trends in
children’s vital capacity index is lower than the normal weight overweight among US children and adolescents 1999-
children’s. BMI is negatively related to the vital capacity index. 2000[J].JAMA,2002,288(14):1728-1732.
Note: the obesity children’s ventilator functions grows far [14] Boyd, G. S., Koenigsberg, J., Falkner, B., Gidding, S., & Hassink, S.
behind the normal children’s, because the extra fat (2005). Effect of obesity and high blood pressure on plasma lipid levels
in children and adolescents. Pediatrics, 116(2), 442-446.
accumulating in the chest and the belly can take mechanism
[15] Juhola J, Magnussen CG, Viikari JS, et al, Tracking of serum lipid
action to the chest and the belly, action of the chest and the levels, blood pressure, and body mass index from childhood to
diaphragm is limited, then the ventilator block grows causing adulthood : the cardiovascular risk in young finns Study. J
the ventilator function to fall, finally the vital capacity and the Pediatr,2011,159(4):584-590
ventilator adjustment changes [16]. The research has not found [16] Lazarus, R., Sparrow, D., & Weiss, S. T. (1997). Effects of obesity and
the relation between pulse and BMI, but the close relation fat distribution on ventilatory function the normative aging
between pulse and individual action may influence the relation study. CHEST Journal,111(4), 891-898.
between pulse and BMI.
Childhood obesity mainly is a kind of pure obesity. The
parents and the school should strengths the health education
onto their children or students; help them to believe in the
conception that health is the most important. They should take
some necessary long-term interfere actions to control the
dietary habit of “high energy, high fat”, increase their exercise
strength, change their bad lifestyle to prevent or control obesity
and other complications.
ACKNOWLEDGMENT
Chengguan of Lanzhou of Science and Technology Project:
Application of bioelectrical impedance measuring body
composition feasibility study in the diagnosis of childhood
obesity (2012-1-11).
REFERENCE
[1] DANIELS SR.Complications ofobesity in children and adolescents.Int
JObes(Lond), 2009,33(Suppl1):S60-S65.
[2] Christodoulos AD, Douda HT, Tokmakidis SP, Cardiorespiratory
fitness, metabolic risk and inflammation in children.INT J
Pediatr,2012,2012:270515
[3] Chen X, Wang Y, T racking of blood pressure from childhood to
adulthood: a systematic review and met aregression
analysis[J].Circulation,2008,117(25):3171-3180
[4] Weiss R, Dziura J, Burgert TS, et al. Obesity and the Metabolic
Syndrome in Children and Adolescents.N Engl J Med, 2004, 350:2362-
2374.
[5] DANIELSSR. The consequences of childhood overweight and obesity.
Future Child, 2006, 16(1):47-67.
[6] Working Group on Obesity in China (WGOC). Guidelines for
prevention and control of overweight and obesity in Chinese adults.
Biom Environ Sci,2004,17(Suppl):1-36.
[7] OGDEN CL,CARROLL MD,CURTIN LR,et al. Prevalence of
overweight and obesity in the United States,1999-
2004[J].JAMA,2006,295(13):1549-1555
[8] Ogden CL, Carroll MD, Kit BK, et al.Prevalence of obesity and trends
in body mass index Among US children and adolescents, 1999-
2010[J].JAMA,2012,307(5):483-490.
[9] Liu W, Lin R, Lin U, et al.Prevalence and association between obesity
and metabolic syndrome Among Chinese elementary school children: a
school-based survey. BMC Public Health, 2010,10:780-785
Ioana-Raluca Edu1,*,
Felix-Constantin Adochiei2,
Radu Obreja2,
Constantin Rotaru1,
Teodor Lucian Grigorie2
1
Military Technical Academy, Faculty of Mechatronics and Armament Integrated Systems, Bucharest, Romania
2
University of Craiova, Faculty of Electrical Engineering, Department of Electric, Energetic and Aerospace Engineering
*Corresponding author (E-mail: edu_ioana_raluca@yahoo.com)
Abstract — The current research is aimed at implementing and This noise component that overlaps over the output of the
validating software procedures by proposing new algorithms that sensors, cannot be totally eliminated but it can be influenced
receive data from inertial navigation systems/sensors (data acquisition by stochastic processes [5].
equipment) and provide accurate navigation information. In the first The development and also the optimization of advanced
phase, our concern was to implement a real-time evaluation criterion algorithms for improving the performances of miniaturized
with the intention of achieving real-time data from an accelerometer.
It is well known that most errors in the detection of position, velocity
inertial sensor and inertial measurement units are extremely
and attitude in inertial navigation occur due to difficult numerical important topics in the field of aerospace navigation systems.
integration of noise. The main goal of this study was to propose a The practical challenge of the study under discussion was to
signal processing algorithm, based on the Wavelet filter, used along develop and validate a complex algorithm that would process
with a criterion for evaluating and updating the Wavelet’s optimal the signals received from accelerometers, and later from INS,
level of decomposition, for reducing the noise component. We remove the noise and offer precise information regarding the
performed numerical simulations using signals received from an vehicle displacement.
accelerometer and analyzed the numerical results to see whether the Therefore, an improved wavelet filter was proposed, used
improved Wavelet (proposed method) can be used to achieve more to remove the noise detected during measurements, in order to
precise information on a vehicle.
obtain a better accuracy of the measurements. The optimal
Keywords — signal processing; wavelet transform, partial directed order of the wavelet filter (the optimal decomposition level)
coherence method. was calculated using a correlation analysis function applied to
the signals achieved from the accelerometers and the real
I. INTRODUCTION speed signals applied to the accelerometers (considered as
The aim of the our scientific research is to develop reference signals).
advanced algorithms able to determine the optimal level of
II. PROPOSED METHOD
decomposition for the Wavelet method and to implement these
algorithms in a miniaturized inertial measurement units in order The Wavelet transform is a very powerful tool for the
to obtain accurate data regarding the vehicle displacement. signal feature extraction and noise reduction and offers
Apart from the indisputable benefits of size reduction, effective localization in time and frequency domains.
reliability, manufacturing costs and power consumption, the In order to analyze signals, the continuous Wavelet
miniaturization of sensors and by default of inertial transform (CWT) can be considered as a tree decomposition of
measurement systems (INS) caused a number of problems the signal (the Wavelet decomposition tree), a combination of
related to their performance degradation. As a result of a set of basic functions, obtained by means of dilation and
miniaturization, stochastic (noise of the system) and translation of a single prototype Wavelet function ψ (t ) called
deterministic errors occurred [1-4]. the mother Wavelet, as illustrated in Fig. 1 [6].
The inertial sensors noise, major source of errors for In practice, the dual tree consists of two discrete Wavelet
inertial navigation systems, is characterized by a constant transforms working in parallel. The branches of the tree are
power throughout the frequency spectrum, that reflects the interpreted as the real part, respectively, the imaginary one of
dynamics of mobile systems - which are intended to be a complex wavelet. Thus the complex wavelet transform is
monitored (generally in the range 0-100 Hz). Therefore, this obtained.
type of noise filtering in the 0-100 Hz band is not The continuous Wavelet transform (Wψ f )( s, τ ) , of the signal
recommended. f (t ) ∈ L2 ( R) can be determined as:
1 ⎛ t −τ ⎞ (1) , ∑ |(4)
(W f )(s, τ ) =
+∞
ψ ∫ f (t ) ⋅ ⋅ψ ⎜ ⎟ ⋅ dt
−∞ s ⎝ s ⎠ where I is the identity matrix, a number of time-varying
where, ψ denote the complex conjugate of ψ . measurements of connectivity can be established.
The PDC coupling estimation between two time series (Xi
and Xj) was defined by Baccala et al. [11] as:
Δ Aij (n, f ) (5)
π ij (n) =
H
a (n, f )a j (n, f )
j
Where the w vector is the white noise and Ar matrices are where, is the Wavelet’s actual level and
expressed by means of the next formulation: is the Wavelet’s optimal level of decomposition.
These measures provide information on wavelet coefficient as
11 , 1 ,
follows: a) if the previous value is lower than the current value
, (3) then the order of the wavelet decomposition is equal to the
1 , , previous value plus 1. b) if the previous value is higher than or
with r=1,.., p model order. parameters represent the linear equal to the current value, then the optimal order of
interaction effect of onto . They are estimated decomposition is equal to the previous value.
by means of an adaptive autoregressive approach [10], the The main idea of the optimization algorithm is illustrated in
main advantage of which is the possibility of analyzing time- Fig. 2, where a signal received from an accelerometer is
varying signals by updating the calculations for each time processed and analyzed by using the Wavelet transform until
sample under investigation. an optimal level of decomposition is established and the useful
By calculating the Fourier transform of Ar(n) matrix, more signal is achieved [10]. This elementary structure was
specific by calculating A(n, t) coefficient matrix in frequency proposed and studied in order to obtain a further general tuning
domain: method for the inertial sensor denoising, with the wavelet
(Fig. 3). 2
Reference 0
signal -2
-4
-6
-8
Signal from Wavelet Filtered signal
PDC -10
accelerometer Denoising 200 400 600 800 1000 1200 1400 1600 1800 2000
-6
level of coupling between the signals, CL values ≈ 1. A higher
-8
level of coupling can be seen in figure 8, for CL = 0.851520,
-10
200 400 600 800 1000 1200 1400 1600 1800 2000
where the predominantly red color suggests the presence of a
strong level of coupling between the two signals. This level of
Fig 4 - Original signal
coupling indicates that, for 0.851520
10 is the .
Ch1 <-- Ch 2
1 the visual inspection we concluded that sinusoidal signal
30
0.9 achieved for =10, which may be seen in Fig. 11
25
0.8
is/ was the optimum level of decomposition - .
0.7
10
20 0.6
8
0.5
15 6
0.4
10 0.3 4
0.2 2
5
0.1 0
0 0 -2
0 0.2 0.4 0.6 0.8 1
-4
-8
0.9
30
0.8 Fig. 10 - Signal achieved for =1
25 0.7
10
20 0.6
8
0.5
15 6
0.4
4
10 0.3
2
0.2
5 0
0.1
0 -2
0
0 0.2 0.4 0.6 0.8 1
-4
-8
Ch1 <-- Ch 2
1 -10
200 400 600 800 1000 1200 1400 1600 1800 2000
0.9
30
0.8 Fig. 11 - Signal achieved for =3
25 0.7
10
20 0.6
8
0.5
15 6
0.4
4
10 0.3
2
0.2
5 0
0.1
-2
0 0
0 0.2 0.4 0.6 0.8 1 -4
10 -8
-10
200 400 600 800 1000 1200 1400 1600 1800 2000
Ch1 <-- Ch 2
1
0.9
Fig. 12 - Signal achieved for =6
30
0.8 10
25 0.7 8
20 0.6 6
0.5 4
15
0.4 2
10 0.3
0
0.2
-2
5
0.1
-4
0 0
0 0.2 0.4 0.6 0.8 1 -6
-8
10
Sensor Applying in Inertial Navigation Systems,"
8 Research - Development and Application on
6 Electronics, Telecommunications and Information
4 Technology, No. 2 , Posts, Telematics & Information
2 Technology Journal, pp. 19-24, 2007.
0 [4] D. G. Eqziabher, "Design and Performance Analysis
-2 of a Low-Cost Aided Dead Reckoning Navigator," A
-4 Dissertation submitted to the Department of
-6 Aeronautics and Astronautics and the committee on
-8
graduate studies of Stanford University in partial
-10
200 400 600 800 1000 1200 1400 1600 1800 2000 fulfillment of the requirements for the degree of
doctor of philosophy, Stanford University, February,
Fig. 14 - Signal achieved for =14 2004.
[5] H. Haiying, " Modeling inertial sensors errors using
By using the proposed configuration from Figure 3, the Allan variance," UCEGE reports number 20201,
correlation between useful signals, signals received from an Master’s thesis, University of Calgary, September,
accelerometer or an inertial measurement unit/INS can be 2004.
tracked and corrected by using signals received from a GPS [6] T. Chan and C. J. Kuo, "Texture Analysis and
device (in a pre-calibration phase of the INS). This correlation Classification with Tree-Structured Wavelet
becomes less clear when the signals achieved from the INS Transform," IEEE Transactions on Image Processing,
become correlated with the signals received from GPS, the Vol. 2, No. 4, October, 1993.
correlation level reaches values close to 1, resulting in reduced [7] H. Khorrami and M. Moavenian, "A comparative
errors of the navigation system (caused by the noise). study of DWT, CWT and DCT transforms in ECG
arrhythmias classification " Expert Systems with
IV. CONCLUSIONS Applications, 2010.
This is a topical issue which brings significant [8] M. Alfaouri and K. Daqrouq, "ECG Signal Denoising
improvement in the inertial navigation systems signal By Wavelet Transform Thresholding," in American
processing having a clear-cut role in positioning investigations. Journal of Applied Sciences 5 (3): pp: 276 – 281,
The purpose of this research was to improve the 2008.
performance of inertial navigation systems and their level of [9] F.-C. Adochiei, S. Schulz, I.-R. Edu, H. Costin, and
accuracy for situations when the GPS becomes unavailable. An A. Voss, "A new normalised short time pdc for
improved version of the Wavelet filter was proposed for dynamic coupling analyses in hypertensive pregnant
filtering/denoising the signals received from an accelerometer. women.," BMT 2013, GRAZ, 19-21 September, 2013.
We intend to implement this algorithm for pre-calibrating [10] T. Milde, L. Leistritz, L. Astolfi, W. H. R. Miltner, T.
a two-dimensional navigation system in the horizontal plan in Weiss, F. Babiloni, and H. Witte, "A new Kalman
order to improve its accuracy in positioning. By establishing filter approach for the estimation of high-dimensional
the best coupling level of signals received from INS and GPS, time-variant multivariate AR models and its
using the GPS signal as reference, the optimal level of application in analysis of laser-evoked brain
decomposition of the wavelet transform can be established and potentials " Neuroimage 50 960-9, 2010.
the proposed algorithm can be implemented in the inertial [11] L. A. Baccala and K. Sameshima, "Partial directed
measurement unit as a real-time evaluation criterion with the coherence: a new concept in neural structure
purpose of achieving real-time data. determination.," Biol. Cybern., 84, 463-474.
REFERENCES
( )
K. Metaxiotis is an Associate Professor at the Decision Support Systems = 2−
Laboratory, Department of Informatics, University of Piraeus, 80, Karaoli &
Dimitriou Str., 18534 Piraeus, Greece (phone: +30210 4142378; fax: +30210
4142264; e-mail: kmetax@unipi.gr). 3. Create a random number u between 0 and 1.
K. Liagkouras is a PhD Candidate at the Decision Support Systems
Laboratory, Department of Informatics, University of Piraeus, 80, Karaoli &
Dimitriou Str., 18534 Piraeus, Greece (e-mail: kliagk@unipi.gr). u [0, 1];
4. Find a parameter with the assistance of the mechanism by recalling the simulated binary crossover (SBX)
following polynomial probability distribution: as shown in section II, as the first two steps are common for
both methods. Indeed as shown below first we calculate the
spread factor β and then the parameter α in the same manner
⎧( ⁄( ) 1 as the SBX.
) ≤ ,
⎪ However, in step 3 we follow a different strategy. As shown
= in section II that illustrates the SBX operator, a random
⎨ 1
⁄( )
⎪ number u ∈ [0, 1] is generated. If u ≤ 1/a, it samples to the left
⎩ ℎ
2− hand side (region between ( ) and ( ) , otherwise if u > 1/a it
samples to the right hand side (region between ( ) and ( ),
where ( ) is the ith parent solution.
The above procedure allows a zero probability of creating any In ECO at this particular point as shown below we follow
children solutions outside the prescribed range ( ) , ( ) . a different methodology. Specifically, instead of generating a
Where ηc is the distribution index for SBX and can take any random number ∈ [0, 1], we generate two random
nonnegative value. In particular, small values of ηc allow numbers, u ∈ [0, 1/ ] to sample the left hand side and a
children solutions to be created far away from parents and
random number u ∈ (1/ , 1] to sample the right hand side of
large values of ηc allow children solutions to be created near
the probability distribution. From the aforementioned process
the parent solutions.
emerge two values of , the that samples the left hand
5. The children solutions are then calculated as follows: side of the polynomial probability distribution and the that
samples the right hand side of the polynomial probability
( ) ( ) ( ) ( ) ( ) distribution. Next, as shown below in step 5 with the
= 0.5 + − | − |
assistance of and are formulated two variants for each
( ) ( )
( ) ( ) ( ) ( ) ( )
child solution. Specifically, c and c are the two variants
= 0.5 + + | − | ( )
that emerge by substituting the β and β to c .
( ) ( )
Respectively c and c are the two variants that emerge by
The probability distributions as shown in step 4, do not create ( )
substituting the and to c .
any solution outside the given bounds ( ) , ( ) instead they Then, by substituting to the parent solution vector at the
scale up the probability for solutions inside the bounds, as position of the selected variable to be crossovered,
shown by the solid line in Fig. 1. ( ) ( )
respectively the c and c we create two different child
( ) ( )
solution vectors (csv), the csv and csv . Thanks to the
( ) ( )
generated csv and csv we are able to perform fitness
evaluation for each one of the corresponding cases. As soon as
we complete the fitness evaluation process, we select the best
( ) ( )
child solution between the two variants c and c with the
assistance of the Pareto optimality framework. The same
( ) ( )
procedure is followed for c and c . The proposed
methodology allows us to explore more efficiently the search
space and move progressively towards higher fitness
solutions. Whenever, there is not a clear winner i.e. strong or
( ) ( )
weak dominance, between the c and c , or respectively
( ) ( )
between the c and c the generation of a random number
allows the random choice of one of the two alternative child
Fig. 1 Probabilities distributions for bounded and unbounded cases solutions.
( ) ( )
The procedure of computing children solutions c and c
from two parent solutions ( ) and ( ) under the Exploratory
III. EXPLORATORY CROSSOVER OPERATOR (ECO) Crossover Operator (ECO) is as follows:
The Evolutionary Algorithms (EAs) in order to be able to
maintain a satisfactory level of progression towards higher 1. Calculate the spread factor β:
fitness regions of the search space is necessary to exploit and
explore efficiently the search space. The exploitation of the 2 ( ) () ( ) ( )
=1+ ( − ), ( − )
search space is done mainly through the selection operator and ( )
−
( )
the exploration through the crossover operator. In this study
we propose a new crossover operator named Exploratory
Crossover Operator (ECO) due to its ability to explore 2. Calculate parameter a :
efficiently the search space. We will start analyzing ECO
uR (1/α, 1];
V. PERFORMANCE METRICS
ZDT3 use 30 decision variables and ZDT4 and ZDT6 use 10 The results in the Tables I and II have been produced by
decision variables respectively. using JMetal [3] framework. Table I presents the results of
ZDT1-4, 6 test functions. Specifically, it presents the mean,
Zitzler-Deb-Thiele’s function N.1 ( ZDT1) problem: standard deviation (STD), median and interquartile range
(IQR) of all the independent runs carried out for Hypervolume
( )= (HV) and Epsilon indicator respectively.
⎧
( )= ( )ℎ( ( ), ( )) Clearly, regarding the HV [4], [7] indicator the higher the
⎪
⎪ 9
( )= 1+ ∑ value (i.e. the greater the hypervolume) the better the
= 29
⎨ computed front. HV is able of capturing in a single number
⎪ ( )
⎪ℎ ( ), ( ) = 1 − both the closeness of the solutions to the optimal set and to a
⎩ ( ) certain degree, the spread of the solutions across the objective
0 ≤ ≤1 1 ≤ ≤ 30 space [8]. The second indicator, the Epsilon [5] is a measure
of the smaller distance that a solution set A, needs to be
changed in such a way that it dominates the optimal Pareto
Zitzler-Deb-Thiele’s function N.2 ( ZDT2) problem: front of this problem. Obviously the smaller the value of this
indicator, the better the derived solution set.
( )=
⎧ Table II use boxplots to present graphically the
⎪ ( )=
( )ℎ( ( ), ( ))
⎪ 9 performance of NSGAII under two different configurations,
= ( )= 1+ ∑ ECO and SBX respectively, for HV and Epsilon performance
29
⎨
⎪ ( ) indicators. Boxplot is a convenient way of graphically
⎪ ℎ ( ), ( ) = 1 −
⎩ ( ) depicting groups of numerical data through their quartiles.
0 ≤ ≤1 1 ≤ ≤ 30
TABLE I
MEAN , STD, MEDIAN AND IQR FOR HV AND EPSILON
Zitzler-Deb-Thiele’s function N.3 ( ZDT3) problem
Problem: ZDT1 NSGAII
( )=
⎧ ECO SBX
( )= ( )ℎ( ( ), ( ))
⎪
⎪
HV. Mean and Std 6.60e-012.9e-04 6.59e-013.2e-04
9 HV. Median and IQR 6.60e-013.5e-04 6.59e-014.8e-04
= ( )= 1+ ∑
29 EPSILON. Mean and Std 1.26e-022.0e-03 1.35e-022.4e-03
⎨ EPSILON. Median and IQR 1.25e-023.0e-03 1.29e-022.5e-03
⎪ ( ) ( )
⎪ℎ ( ), ( ) = 1 − − sin 10 ( )
⎩ ( ) ( ) Problem: ZDT2 NSGAII
ECO SBX
0 ≤ ≤1 1 ≤ ≤ 30
HV. Mean and Std 3.27e-012.6e-04 3.26e-012.9e-04
HV. Median and IQR 3.27e-013.6e-04 3.26e-014.0e-04
EPSILON. Mean and Std 1.28e-022.1e-03 1.37e-022.5e-03
Zitzler-Deb-Thiele’s function N.4 ( ZDT4) problem EPSILON. Median and IQR 1.24e-022.9e-03 1.30e-022.9e-03
TABLE II
BOXPLOTS FOR HV AND EPSILON
VII. ANALYSIS OF THE RESULTS - CONCLUSIONS
Problem ZDT1 NSGAII In this section, we analyze the results obtained by applying
HV EPSILON the Exploratory Crossover Operator (ECO) and the Simulated
ECO SBX ECO SBX
Binary Crossover (SBX) operator respectively to the NSGAII
for solving the ZDT1-4, 6 benchmark problems. The
0.02 assessment of the performance of the proposed crossover
0.660
indicator.
0.659
0.01
REFERENCES
Problem ZDT3 NSGAII [1] Koza, J.R. Genetic Programming. MIT Press, Cambridge; 1992.
HV EPSILON [2] Deb, K. & Agrawal, R.B. (1995) Simulated binary crossover for
ECO SBX ECO SBX continuous search space, Complex Systems 9 (2) 115–148.
[3] Durillo, J. J. & Nebro, A. J. (2011). jMetal: A Java framework for multi-
objective optimization, Advances in Engineering Software 42, 760–771
[4] Zitzler, E., Brockhoff, D. & Thiele. L.(2007). The Hypervolume
Indicator Revisited: On the Design of Pareto-compliant Indicators Via
0.514
0.2
[5] Zitzler, E., Thiele, L., Laumanns, M., Fonseca, C.M. & Da Fonseca,
0.512
47, 1998.
0.06
[8] Zeng, F., Decraene, J., Hean Low, M.Y., Hingston, P., Wentong, C.,
Suiping, Z. & Chandramohan, M. (2010). Autonomous Bee Colony
0.645
0.02 0.04
0.01
Abstract—The computational prediction of non-stationary be- is based on the balance law of classical thermodynamics.
haviour of cementitious composites exposed to high temperature, However, it is not trivial to supply a resulting system of
supported by proper physical and mathematical formulations, is evolution by reasonable effective constitutive relations, valid
a rather complicated analysis, dealing with the coupled heat and
(liquid and vapour) mass transfer and related mechanical effects. at the microstructural level, despite of the complicated (typi-
The macroscopic mass, momentum and energy conservation cally porous) material structures. Most classical approaches to
equations need to exploit available information from micro- evaluate temperature and moisture redistributions rely on the
structural considerations. The derivation and implementation (semi-)empirical relations from [1]. The attempts to improve
of a computational model is always a compromise between the them can be seen in various directions, with still open prob-
model complexity and the reasonable possibility of identification
of material characteristics of all components and their changes. lems and difficulties everywhere. The physical and/or math-
The paper presents a possible approach to such analysis, supplied ematical homogenization approaches applied to a reference
by an illustrative example, confronting computational results volume element seem to be useful. The process of cement
from a simplified model (taking only dominant physical pro- hydration, distinguishing between anhydrous cement scale,
cesses into account), with experimental ones, performed in the cement-paste scale, mortar scale and macro-scale, handled
specialized laboratory at the Brno University of Technology.
by specific physical and mathematical formulations, with a
Index Terms—Cementitious composites, high temperature be- posteriori scale bridging applying least squares arguments, has
haviour, poro-mechanics of multi-phase media, computational
been described in details, evaluating the hydration of particular
simulation.
clinker phases, in [14]. However, the thermal gains and losses
from chemical reactions cannot be considered as deterministic
I. I NTRODUCTION ones, moreover corresponding computations just in the case of
supplied by appropriate initial and boundary conditions, 6. if R , V and T are sufficiently small, return to 3,
e. g. for a priori known values of all variables for t = 0 of 7. stop the computation if the final time is reached, other-
the Dirichlet, Cauchy, Neumann, Robin, etc. types, for local wise return to 2.
unit boundary outward normals n(x) = (n1 (x), n2 (x), n3 (x))
in particular Consequently all practical computational tools make use of
strong simplifications. The rather general approach of [8] in-
• σij nj = ti with imposed tractions ti ,
a s a a a troduces the additive linearized strain decomposition instead of
• (ρa (vi − vi ) + ri )ni = r with imposed air fluxes r ,
w s w v s v w v the multiplicative finite strain one, ignores some less important
• (ρw (vi −vi )+ri +ρv (vi −vi )+ri )ni = r +r +β(ρv )
terms, as the kinetic energy (the first additive term in φε ) in
with imposed liquid water and vapour fluxes rw , rv and
the energy balance interpretation of (1), as well as the Dufour
some mass exchange function β,
w s ε c and Soret effects in (3) and (4), and reduces the number of
• (ρw (vi − vi )hv − λij T,j − ξij p,j )ni = q̄ + α(T ) with
variables, comparing (2) with the differences viε − vis with ε ∈
imposed heat fluxes and some heat exchange function α,
{v, w, a} from the momentum balance interpretation of (1),
e. g. by the Stefan - Boltzmann law, proportional to T 4 .
and presents a computational scheme of the above sketched
This seems to be a correct and complete formulation for the type, applying the Galerkin formulation and the finite element
analysis of time development of R, V and T . technique. Nevertheless, most engineering approaches, as [2]
or [12], endeavour to obtain a system of 2 equations of
III. C OMPUTATIONAL PREDICTION evolution for T and pc (or some equivalent quantity) only, pre-
Unfortunately, a lot of serious difficulties is hidden in the eliminating or neglecting all remaining ones, using arguments
above presented formulation, e. g. the still missing “Millenium of various levels: from micro-mechanically motivated physical
Prize” existence result on the solvability of Navier - Stokes considerations to formal simplification tricks.
equations (cf. the “mysteriously difficult problem” of [15], The reliability of computational results depends on the
p. 257), the physical and mathematical incompatibilities like quality and precision of input data, including the design
[7], as well the absence of sufficiently robust, efficient and and identification of all material characteristics. Probably all
reliable numerical algorithms based on the intuitive time- authors take some of them from the literature, not from their
discretized computational scheme: extensive original experimental work; the variability of forms
1. set R, V and T by the initial conditions for t = 0, of such characteristics and generated values, accenting those
2. go to the next time step, preserving R, V and T , from (2), including the formulae from [8], has been discussed
3. solve some linearized version of (1) with the mass in [5]. Nevertheless, the proper analysis of uncertainty and
balance choice, evaluate and perform the correction R , significance of particular physical processes lead to even more
4. solve some linearized version of (1) with the momentum complicated formulations, analyzed (for much easier direct
balance choice, evaluate and perform the correction V , model problems) in [17] using the spectral stochastic finite
5. solve some linearized version of (1) with the energy element technique, or in [10] the Sobol sensitivity indices, in
balance choice, evaluate and perform the correction T , both cases together with the Monte Carlo simulations.
ISBN: 978-1-61804-240-8 165
Advances in Applied and Pure Mathematics
ACKNOWLEDGMENT
This research has been supported by the research project
FAST-S-14-2490 at Brno University of Technology, in collab-
oration with the research project FAST-S-14-2346.
R EFERENCES
[1] Z. P. Bažant, W. Thonguthai, Pore pressure and drying of concrete at
high temperature. Journal of the Engineering Mechanics Division 104,
1978, pp. 1059–1079.
[2] M. Beneš, R. Štefan and J. Zeman, Analysis of coupled transport
phenomena in concrete at elevated temperatures. Applied Mathematics
and Computation 219, 2013, pp. 7262–7274.
[3] A. Bermúdez de Castro, Continuum Thermomechanics. Birkhäuser,
2005.
[4] D. Cioranescu and P. Donato, An Introduction to Homogenization.
Oxford University Press, 1999.
Fig. 4. Time development of moisture content ρw + ρv . [5] C. T. Davie, C. J. Pearce and N. Bićanić, Aspects of permeability in
modelling of concrete exposed to high temperatures. Transport in Porous
Unlike most experiments in civil and material engineering in Media 95, 2012, pp. 627–646.
laboratories and in situ, nearly no relevant results for advanced [6] J. H. Ferziger and M. Perić, Computational Methods for Fluid Dynamics.
Springer, 2002.
numerical simulations are available from real unexpected fires, [7] T. Fürst, R. Vodák, M. Šı́r and M. Bı́l, On the incompatibility of
as the most dangerous conditions for building structures, Richards’ equation and finger-like infiltration in unsaturated homoge-
crucial for the reliable prediction of behaviour of concrete neous porous media. Water Resource Research 45, 2009, W0340, 12 pp.
[8] D. Gawin and F. Pesavento, An overview of modeling cement based
and similar structures. Since also laboratory experiments with materials at elevated temperatures with mechanics of multi-phase porous
cementitious composites under the conditions close to a real media. Fire Technology 48, 2012, pp. 753-793.
fire, as that documented on Figure 1, performed at the Brno [9] D. Gawin, F. Pesavento and B. A. Schrefler, Simulation of damageperme-
ability coupling in hygro-thermo-mechanical analysis of concrete at high
University of Technology, are expensive, producing incom- temperature. Communications in Numerical Methods and Engineering
plete and uncertain quantitative data anyway, the reasonable 18, 2002, pp. 113-119.
goal of numerical simulation is to test simple numerical [10] J. Gottvald and Z. Kala, Sensitivity analysis of tangential digging forces
of the bucket wheel excavator SchRs 1320 for different terraces. Journal
models with acceptable correlation with real data, as the first of Civil Engineering and Management 18, 2012, pp. 609–620.
step for the development of more advanced multi-physical [11] P. Jarošová and S. Šťastnı́k, Modelling of thermal transfer for energy
models. saving buildings. International Conference on Numerical Analysis and
Applied Mathematics 2013 – Proceedings, American Institute of Physics,
Figures 2, 3 and 4 are the outputs from a rather simple 2013, pp. 1000–1003.
two-dimensional isotropic model, neglecting mechanical loads, [12] V. Kočı́, J. Maděra, T. Korecký, M. Jerman and R. Černý, Computational
strains and stresses, compatible with the slightly modified and analysis of coupled heat and moisture transport in building envelopes:
the effect of moisture content. Thermophysics 2013 – Conference Pro-
revisited approach of [2], combining the finite element and ceedings in Podkylava (Slovak Republic). Slovak Academy of Sciences
volume techniques together with the iterated time discretiza- in Bratislava, 2013, pp. 62–71.
tion scheme (the numerical construction of Rothe sequences), [13] G. Nguetseng and N. Svanstedt, σ-convergence. Banach Journal of
Mathematical Analysis, 5, 2011, pp. 101–135.
implemented in the MATLAB environment. The fire is consid- [14] Ch. Pichler, R. Lackner and H. A. Mang, Multiscale model of creep of
ered as the boundary thermal radiation on the left and upper shotcrete – from logarithmic type viscous behaviour of CSH at the µ-
edges of the rectangle. Surprisingly some phenomena observed m scale to macroscopic tunnel analysis. Journal of Advanced Concrete
Technology 6, 2008, pp. 91–110.
in situ can be explained from the deeper analysis of results of [15] T. Roubı́ček, Nonlinear Partial Differential Equations with Applications.
such seemingly simple calculations. Birkhäuser, 2005.
[16] L. Sanavia, B. A. Schrefler and P. Steinmann, A formulation for an unsat-
urated porous medium undergoing large inelastic strains. Computational
IV. C ONCLUSION Mechanics 1, 2002, pp. 137–151.
Development of reasonable models and computational algo- [17] N. Zabaras, Inverse problems in heat transfer. In: Handbook on Numer-
ical Heat Transfer (W. J. Minkowycz, E. M. Sparrow and J. S. Murthy,
rithms for the prediction of thermal, mechanical, etc. behaviour eds.), Chap. 17. John Wiley & Sons, 2004.
of cementitious composites and whole building structures, is
strongly required by the applied research in civil engineering,
Jiřı́ Vala (*1956) is professor of the Institute of Mathematics and Descriptive
which cannot be ignored, although the formal mathematical Geometry at the Brno University of Technology, Faculty of Civil Engineering.
verification of such models is not available and the practical Anna Kučerová (*1985) and Petra Rozehnalová (*1984) are Ph.D. students
validation suffers from the lack of data from observations at the same university.
Abstract—In fault detection, data-driven methods are limited Regarding industrial processes, failures usually occur
by the lack of knowledge about faults. This paper presents an in sensors, actuators or processes. The main purpose of
approach based in statistical modeling, which uses normal oper- fault detection is to evaluate symptoms, which indicate the
ation information for training a Gaussian Mixture Model. Such difference between normal and faulty states. As such, and
model has been widely used in speaker verification, because is taking into account the performed observation, the methods
capable of mapping the observations into a multidimensional
membership function, this feature makes this algorithm very
of failure detection are divided into three categories:
convenient because works in a unsupervised way. This technique signal threshold, signal model and process model based
uses the normal behavior data to estimate the parameters and approaches, see [6].
then classify future data according to the model. The proposed
approach consists of two stages: an offline stage, where the From the taxonomy of failure detection presented
Gaussian model is trained by using Expectation-Maximization previously, the proposed approach in this paper falls within
algorithm; and the online stage, where the readings likelihood the signal model-based category, as it extracts the statistical
is evaluated according to the normal operation model. If the features of the process from the sensor readings, where the
likelihood is found to be inferior to a certain threshold, a training is taking place off-line. As such, this approach is
fault is detected. This approach may be extended to multiple suitable for real-time applications, because the operations
processes in science and engineering; for instance, a case study
required for classification can be calculated in small batches
is presented using the DAMADICS benchmark, where this
approach was validated. while new data is being read.
Keywords—Fault Detection, Gaussian Mixture Models, Indus- This paper is organized as follows. Firstly, A brief re-
trial Processes view of GMM is presented, fault detection by using GMM
is described in Section 3. The proposed scheme is vali-
I. I NTRODUCTION dated by simulation examples of Chemical Plant Benchmark
S the industrial processes complexity keeps increasing, (DAMADICS) in Section 4. Finally, we present the conclu-
A fault detection has become a critical issue in order to
meet safety, enviromental and productivity requirements.
sions obtained from the implementation and discuss future
work.
In this regard, several approaches have been proposed,
particularly, being artificial intelligence methods the current II. G AUSSIAN M IXTURE M ODELS
trend. However, one of the main drawbacks of these Gaussian mixture models are commonly used in
methods is the implementation complexity for a real statistical analysis. In most applications, its parameters are
industrial environment, which limits its appeal for most determined by using Maximum Likelihood and Expectation
industrial applications. Hence, approaches that may incur Maximization algorithm.
in a lower complexity while keeping acceptable results are
highly favored. The approach presented in this paper aims As industrial processes are influenced by multiple facts,
at this paradigm. like temperature, noise and environmental conditions, it is
required to include adequate modelling to these variables.
Computer-based solutions have been proposed to meet The GMM has the capability of representing a large class
the aforementioned requirements, for instance, machine of sample distributions, which is a very useful tool for
learning methods, see [1], like Bayesian Learning methods the characterization of multiple problems, like speaker
[2], Self-Organizing Maps [3], or PCA [4],etc. The main recognition, see [7].
objective of these methods is to transform the monitoring
data (usually sensors) into statistical models, in order to In the context of fault detection systems, the usage of a
classify and detect anomalies. Since the analytical model GMM for representing features distributions, is motivated
is not taken into account, the method remains flexible for by the notion that the individual component densities can
implementation in different environments. model a set of hidden classes. In the same context, we
assume that the features in the normal operation space
However, a shortcoming of artificial intelligence methods correspond to different hidden classes. We assume that if we
is the need of data from different anomalies or failures to learn a Gaussian Mixture Model with sufficient variability of
train the model. Some systems have low on-site failure rate, normal operation behavior, then the model will be capable
and as a consequence, the amount of failure data available of distinguish between any state different than the normal
is limited, see [5]. In this paper, a GMM approach for fault one. Because all the features used to train the GMM are
detection is proposed, to overcome the lack of failure data, as unlabeled, the healthy state classes are hidden in a class of
this method can be trained using only normal operation data. an observation that is unknown.
M
X Fig. 1. Block diagram for fault detection by using GMM
wi = 1 (2)
i=1
where wi are the mixing coefficients and the parameters of 2) Evaluate probabilistic distance using the current pa-
the of the component density functions p(x|λ) which usually rameter values
vary with i. T
1X
wi = Pr(i|xt , λ) (6)
(x − µi )T (x − µi )
1 T t=1
g(x|µi , Σi ) = exp −
(2π)D/2 |Σi |1/2 2Σi PT
(3) Pr(i|xt , λ)xt
where g(x|µi , Σi ) are the component Gaussian densities. The µi = Pt=1
T
(7)
set of parameters is λi = (wi , µi , Σi ), where i = 1, . . . , M , t=1 Pr(i|xt , λ)
M is the GMM model size, µi and Σi are the mean and PT 2
t=1 Pr(i|xt , λ)xt
covariance matrix of the ith GMM component, respectively: σ 2i = P T
− µi (8)
Finally, D is the dimension of vector x. t=1 Pr(i|xt , λ)
3) Calcule the aposteriori data probabilities.
To determine the ”membership” of the data point x to the
ith GMM component, the loglikelihood of the data point, wi g(xt |µi , Σi )
Pr(i|xt , λ) = PM (9)
of each individual model distribution is calculated using the k=1 wk g(xt |µk , Σk )
Equation 4 . An implementation of this algorithm for Matlab
can be found at [8]. 4) Calcule the error, which is the negative of the likeli-
hood function.
T
X 5) Go back to step 2 until the error change is very small
log p(x|λ) = log p(xt |λ) (4) or the number of iterations is reached (usually 15-20).
t=1 where σ 2 ,xt , and µi , refer to arbitrary elements of the vectors
σ 2 , xt , and µi respectively.
A. Expectation Maximization Algorithm for Gaussian Mix-
tures III. FAULT DETECTION USING GMM
Given a set of training vectors and a GMM configura- The main advantage of using Gaussian Mixture Models,
tion (model size, minimum variance, which are determined is that the model can be trained with normal operation data,
heuristically),the parameters of the GMM λ are left to be without further information about the multiple failure states,
estimated. These parameters, should lead to the best match which can be unknown at the moment. This advantage,
(in the maximum likelihood sense) for the distribution of allows the detection of multiple failure states even if a
the training features vectors. There are several techniques particular failure is happening for the very first time.
available for estimating the parameters of the GMM, being
Maximum Likelihood (ML) the most used method. The aim The signals can be used directly or transformed to extract
of ML estimation is to find the model parameters which more interesting features for the detection task. In this article
maximize the likelihood of the GMM given the training data. we use the raw signals to learn the model.
For a sequence of T training vectors X = {x1 , . . . , xT }, The principle of this method consists of two phases: a
the GMM likelihood, assuming independence between the first phase during which an statistical model is learned using
vectors, can be written as the Expectation Maximization algorithm; and a second
T phase where the learned model is used to score the current
system condition, as it is shown in Figure 1. If the maximum
Y
p(X|λ) = p(xi |λ) (5)
t=1
likelihood score exceeds certain threshold, then the detection
system triggers a failure alarm.
Unfortunately, for this expression direct maximization is not
possible. This equation, however, can be solved iteratively The proposed method described in the Algorithm 1, is
using a special case of the EM algorithm. The goal of this a data-based technique which mainly uses the sensor data
algorithm is to maximize the likelihood function (Equation (consisting as well of normal operating data called xtr )
4) according to the parameters: means, covariances and the gathered during a significant period of time to adjust the
mixing coefficients for all the components. The process is parameters of a Gaussian mixture model called λ of size
divided in the following stages: M . The threshold is obtained using validation data (also
1) Initialize the means µi , covariances σi and mixing normal operation data), called xv . The obtained model uses
coefficients wi , and evaluate the initial value of the log Likelihood to distinguish between data that is ”normal”,
likelihood. and the data that is ”anomalous”, the data set for testing
14: end if
0.9
15: end for
16: end procedure Normalized Log likelihood
0.8
0.5
This benchmark was proposed in 2003 [9] to help the 0.3 Fault 16
the model, and the failure data to test the results. The faulty
data is scarce, but we simulated 30 cases to test our approach
Fig. 3. Normalized Loglikelihood for data set, corresponding to November
with the simulink tool available in the repository. 17th, where multiple types of faults occurred. In this picture we can see that
three faults occurred after the time window number 50.
V. E XPERIMENTS
To validate the functionality of the algorithm we use the
DAMADICS benchmark. This database was created to test sensors and valves readings, whose size depends on the
fault detection algorithms with real data and simulations. failure mode, for this article we only consider the 16, 17 and
The authors provided libraries to simulate 19 types of 18 faults, which are general external failures. The model size
faults from three actuators with two different degrees was determined heuristically, as well the minimum variance
of ocurrence (Abrupt, Incipient). In this experiment we and detection threshold.
monitor three signals: controller output, upstream pressure
and downstream pressure.
B. Model identification
In Table I, the analysed fault types are presented, and The goal of this stage is to obtain a GMM for normal
the fault scenarios that are derived from them. As the operation data. The data set concerning the failures is used
monitoring of root causes for a faulty event is a very only in the testing stage. The first step uses the k-means
difficult task for the operators, it follows to infer the state algorithm to locate the centroids for the initial parameters.
of the process, from key variables belonging to it. The second step relies in the EM algorithm to estimate
4
x 10 Frequency Distribution Normal Operation Data (Three variables case)
14
12 LC5103CV
LC5103X
Frequency
10
8 LC5103PV
6
4
2
25 30 35 40 45 50 55
4
x 10 Estimated Gaussian Mixture Model Normal Operation Data (Three variables case)
14
12
Frequency
10
8
6
4
2
25 30 35 40 45 50 55
seconds
Fig. 4. Upper: Histogram of Normal operation data set corresponding to November 1st to 5th for training purposes. Below: Histogram of learned Gaussian
Mixture model.
the GMM parameters, where the covariance matrices are to 0.82 for normalized likelihood. Every one of the faults
set to be diagonal. The appropriate size of the model is expected in every day, was detected by the algorithm. To
determined, in order to obtain trustworthy results. Also, the prove the robustness of the algorithm, we use simulated
best minimum covariance threshold is found, to ensure that data, for normal behavior data we use 100 tests and for
the model converges and remains sensitive to variances in faulty behavior we use 30 test. The results for false alarm
data. The model selection is based on the likelihood function was 6% and false rejection was 1%. This numbers change
evaluated over the validation set. The result is a model for according to the threshold but these were the best found in
the normal operation data which detects anomalies in the the experiments.
process by using loglikelihood function.
VII. C ONCLUSION
C. Testing the Model
A data set representing an operating condition as shown A fault detection system based on GMM was developed.
in Figure 4 is fed to the GMM model. This data set is Using the DAMADICS benchmark actuator system it was
composed by data of the three previously mentioned signals, possible to verify its capabilites regarding monitoring tasks.
between November 1st-4th. The whole data set has 345600 With a model of normal operations was able to detect
data points for each signal, which is equivalent to 4 days of fault events immediately after its occurrence. This pattern
operation. recognition tool plays an important role when fault event
data is scarce, and we have plenty of normal operation data.
From the data set, we calculate the loglikelihood score and Besides, by using a temporal sequence of observations as
normalize it by the number of samples in the window (set to model input, the model may be trained again if there are any
1000 secs). If the data set score is lower than the threshold , changes in the behavior of the process, with the posibility
the algorithm detects an anomaly. The model is tested with of adding even more signals, which can be a daunting task
different types of faults, being important to note that the user with model based methods.
has to set the threshold according to the sensibility desired.
The results are very promising concerning to the
application of GMM in chemical process monitoring. In the
VI. R ESULTS AND D ISCUSSION future, these results can be used along with decision trees
A set of observation sequences belonging to a particular to include a fault diagnosis stage, as a complement for the
fault scenario are fed to the model. The results for both proposed system. It follows from the results obtained, that
monitoring tasks and detection, are shown in the Figures 2 the proposed approach may have further applications in the
and 3. These figure show how the score decreases drastically monitoring of other industrial procesess as well, leading to
when a fault is detected. The fault scenario corresponds to prospective opportunities for research and development.
October 30th, and 17th of November. We set the threshold
MAY 2014 5
R EFERENCES
[1] Y. Guo, J. Wall, and J. L. andSam West, “A machine learning approach
for fault detection in multi-variable systems,” in ATES in conjunction
with Tenth Conference on Autonomous Agents and Multi-Agent Systems
(AAMAS) AAMAS 2011, 2011.
[2] J. P. Matsuura, A. Iees, P. Marechal, E. Gomes, T. Yoneyama,
R. Kawakami, and H. Galvo, “Learning bayesian networks for fault
detection,” in In Proceedings of the IEEE Signal Processing Society
Workshop, 2004, pp. 133–142.
[3] T. Chopra and J. Vajpai, “Classification of faults in damadics bench-
mark process control system using self organizing maps,” International
Journal of Soft Computing and Engineering (IJSCE) ISSN: 2231-2307,
Volume-1, Issue-3, July 2011, vol. 1 (3), pp. 85–90, 2011.
[4] J. Sun, Y. Li, and C. Wen, “Fault diagnosis and detection based on com-
bination with gaussian mixture models and variable reconstruction,” in
Innovative Computing, Information and Control (ICICIC), 2009 Fourth
International Conference on, 2009, pp. 227–230.
[5] F. Nelwamondo and T. Marwala, “Faults detection using gaussian
mixture models, mel-frequency cepstral coefficients and kurtosis,” in
Systems, Man and Cybernetics, 2006. SMC ’06. IEEE International
Conference on, vol. 1, 2006, pp. 290–295.
[6] A. Shui, W. Chen, P. Zhang, S. Hu, and X. Huang, “Review
of fault diagnosis in control systems,” in Proceedings of
the 21st annual international conference on Chinese control
and decision conference, ser. CCDC’09. Piscataway, NJ,
USA: IEEE Press, 2009, pp. 5360–5369. [Online]. Available:
http://dl.acm.org/citation.cfm?id=1714810.1715162
[7] D. A. Reynolds, T. F. Quatieri, and R. B. Dunn, “Speaker verification
using adapted gaussian mixture models,” in Digital Signal Processing,
2000, p. 2000.
[8] I. Nabney, NETLAB: Algorithms for Pattern Recognition, ser. Advances
in Computer Vision and Pattern Recognition. Springer, 2002. [Online].
Available: http://books.google.com.co/books?id=LaAAJP1ZxBsC
[9] M. Syfert, R. Patton, M. Bartys, and J. Quevedo, “Development and
application of methods for actuator diagnosis in industrial control
systems (damadics): A benchmark study.” Proceedings of the IFAC
Symposium Safe Process, pp. 939–950, 2003.
I. INTRODUCTION
intervertebral disc can be considered a ball joint with three Figure 4 represents the lumbar segment in the coronal plane.
degrees of freedom corresponding to rotation after three axes. The L1, L2 ... L5 vertebras are the CrT12-L1CrL2-L3-L3
In figure 3 are shown as an example, L3 and L4 vertebras CRL1-L2CrL2, CrL2-L3CrL3-L4-S1 ... CrL4-L5CrL5
centers as CL3 and CL4 points, and the rotation centers of the segments. The relative rotation between two vertebras is given
L2, L3, L4 and L5 vertebras, as CrL2-L3-L4 and CrL4 CrL3- by the angle αi between the segments representing the two
L5. vertebras.
Fig. 4 – The lumbar spine with the segments representing L1, L2 ...
Fig. 2. – The spine in the coronal plane related to the coordinate L5 vertebras.
system xOy.
The motion law of the spine in the coronal plane can be
expressed as a function of the vehicle speed (va), the curved
trajectory radius (rtr) and the upper body mass (mcs), function
that returns the values of the αi angles.
(1)
Fig. 3 - L3 and L4 vertebras centers (CL3 and CL4), and the rotation
centers of the L2, L3, L4 and L5 vertebras (CrL2-L3-L4 and CrL4
CrL3-L5).
Considering the vertebras in the coronal plane as
represented by segments connecting the rotation centers, the
shape of the spine may be given by the angles αi of these
segments.
Fig. 5 - The route used in the experiment.
The track used in the experiment is the same track used to stage the position in time of one of the three points is
determine the dynamic cornering ability of the vehicles (fig. determined. In each stage the vehicle is traveling with a
5).[29] constant speed of 15km/h according to the vehicle dynamic
In the experiment we used the motion sensor manufactured steering ability tests. [5]
by PASCO scientific and the PASCO CI-6400 Science
Workshop 500 Interface (fig. 7).
The motion sensor MotionSensor II (fig. 6) operates on the
sonar principle. The sensor can measure distances between
0.15m and 8m, between it and the object of interest (fig. 6).
different amplitudes.
Fig. 10 - The series of measurements for the C1 vertebra. Fig. 13 – Graphical representation of the positions in time of the
three points.
cp := um := lb :=
0 0 0
0 0.477 0 0.39 0 0.418
1 0.476 1 0.399 1 0.418
Fig. 11 - The series of measurements the T4 vertebra. 2 0.479 2 0.401 2 0.418
3 0.478 3 0.403 3 0.419
4 0.482 4 0.405 4 0.419
5 0.479 5 0.407 5 0.419
6 0.474 6 0.409 6 0.419
7 0.48 7 0.411 7 0.42
8 0.473 8 0.411 8 0.42
9 0.473 9 0.411 9 0.42
10 0.475 10 0.411 10 0.42
11 0.48 11 0.411 11 0.421
12 0.469 12 0.41 12 0.422
13 0.476 13 0.412 13 0.424
14 0.485 14 0.413 14 0.425
15 ... 15 ... 15 ...
Fig. 12 - The series of measurements for the L1 vertebra. The cp string corresponds to the C1 point, um string
corresponds to the T4 point and lb string corresponds to the L1
point.
The next step is to determine the frequency of each string.
tcpmax := i if max( cp ) cp
i i
0 otherwise
max( tcpmax) = 98
tcpmin := i if min ( cp ) cp
i i (7)
0 otherwise
(8)
max( tcpmin ) = 83
(2)
(3)
(4)
(6)
Fig. 15 - Graphical representation of the yum sinusoidal function Fig. 18 – The sinusoidal functions describing the position variation in
compared with the um string. time of the C1, T4 and L1 vertebras.
(13)
(14)
(15)
V. CONCLUSIONS
The amplitude values of the sinusoidal functions describing
Fig. 16 - Graphical representation of the ylb sinusoidal function the variation in time of angles between the vertebras gives an
compared with the lb string. image regarding the deformation degree of the intervertebral
discs.
A nonergonomic posture of the driver’s body seated in the
vehicle’s seat implies the spine to be in a shape that subjects
the intervertebral discs to uneven tensions causing
deformations that in some cases can exceed the limits at which
the musculoskeletal affections of the spine can be avoided or
treated by physiotherapy.
REFERENCES
[1] Adams M. ; Bogduk N. ; Burton K. ; Dolan P. (2006). The
Biomechanics of Back Pain Second Edition, Churchill Livingstone
Elsevier.
[2] Borozan I. S.; Maniu I.; Kulcsar R. M.; „Ergonomic analysis on driving
an Automatic Gearbox equipped vehicle”, SACI 2012 IEEE 7th
International Symposium on Applied Computational Intelligence and
Fig. 17 – The sinusoidal functions describing the position variation in Informatics, May 24-26, 2012, Timisoara, Romania.
time of the C1, T4 and L1 vertebras points. [3] Borozan I. S.; Kulcsar R. M.; “Vertebral column bioengineering
analysis at bending and torsion”, International Conference on Human-
Machine Systems, Cyborgs and Enhancing Devices HUMASCEND,
Iasi, Romania, June 14-17, 2012.
[4] Goran Devedžić, Saša Ćuković, Vanja Luković, Danijela Milošević, K.
Subburaj, Tanja Luković, "ScolioMedIS: web-oriented information
system for idiopathic scoliosis visualization and monitoring", Journal of
Computer Methods and Programs in Biomedicine, Vol.108, No.-, pp.
736-749, ISSN -, Doi 10.1016/j.cmpb.2012.04.008, 2012.
[5] Hilohi C., Untaru M., Soare I., Druţa Gh., „Metode si mijloace de
incercare a automobilelor”, Editura Tehnică București, 1982.
[6] Hinza B., Seidel H., „The significance of using anthropometric
parameters and postures of European drivers as a database for finite-
Abstract—We use a measure for distances of neighbors’ of a II. M ULTIDIMENSIONAL POINT PROCESSES AND FRACTAL
given point that is based on lp metrics and a scaling exponent. We BEHAVIOR
show that if the measure scales with scaling exponent mentioned,
then distribution function of this measure converges to Erlang A. Point processes
distribution. The scaling of distances is used for design of a
classifier. Three variants of classifier are described. The local Let there be an ,,underlying process“ UP . This process
approach uses local value of scaling exponent. The global method is sampled randomly and independently so that random d-
uses the correlation dimension as the scaling exponent. In the dimensional data
IINC method indexes of neighbors of the query point are essential.
Results of some experiments are shown and open problems of
classification with scaling are dicussed. P = x1 , x2 , . . ., xi ∈ X ⊂ Rd (1)
Keywords—Multivariate data, nearest neighbor, Erlang distribu-
tion, multifractal, scaling exponent, classification, IINC. arose. This data (without respect to time or order in which
individual samples xi was taken) forms spatial point process in
Rd and individual samples xi are called points, in applications
often events [6], samples, patterns or so.
I. I NTRODUCTION We are interested in distances from one selected fixed point
x to others; especially distance to the k-th nearest neighbor.
In this paper we use a model that there is some underlying From now we use numbering of points according to their order
process and in the process some events occur. We suppose as neighbors of point x; xk being the k-th nearest neighbor of
that events occur randomly and independently one of another. point x. To distance lk from x to its k-th nearest neighbor a
The only information we have is d-dimensional data arising probability is assigned. There is introduced
from events, i.e. by (often rather approximate) measurement
or sampling.
Sk (l) = P r{l < lk } = P r{N (lk ) < k}
Important notion is a scaling characterized by scaling
exponent denoted also as fractal dimension. This dimension q
is lesser than space dimension d, and usually is not an integer. i.e. probability that a distance to the k-th nearest neighbor
The space dimension d is often called embedding dimension is larger than l that is equal to probability of finding k-1
using concept that fractal is a q-dimensional formation plunged points within distance lk [4]. For k = 1 it is called avoidance
into larger d-dimensional space [16]. This concept can be probability and often denoted P0 . Function
applied to volume V of a ball of radius r. There is V = cq rq
for q-dimensional ball in d-dimensional space; cq is a constant Fk (l) = 1 − Sk (l)
dependent on q and metrics used. Usually q = d but the same
holds for integer q < d, e.g. two dimensional circle in three is the distribution function of distance l to the k-th neighbor.
dimensional Euclidean space. Keeping the concept consistent, A scaling function is a real-valued function c : Rd →
q need not be an integer but there is no intuition how, say, R+ , that satisfies a self-similarity property with respect to a
2.57-dimensional ball looks like. group of affine transformations [20]. There are several types
The goal of this study is to analyze the distances of nearest of scaling functions, shifting, scaling, eventually reflections.
neighbors from given point (location) in a multidimensional General equation for scaling can have form
spatial point process in Rd with exponential scaling [5]. The
result is that when using scaled measure for distance of the µ(~x + ~a) = cθ (~x)
k-th neighbor, the distance can have the Erlang distribution of
order k. We show here that scaling leads to simple polynomial and in less general (fractal) case of exponential scaling
transformation z = rq . With the use of this transformation a
classifier can be designed.
µ(~x + ~a) = ah(~x)
The work was supported by Ministry of Education of the Czech Republic
under INGO project No. LG 12020.
Here θ is l-dimensional parameter vector. When the scaling
M. Jiřina is with the Institute of Computer Science AS CR, Pod Vodarenskou is location dependent, we speak about locally dependent point
vezi 2, 182 07 Praha 8, Czech Republic (e-mail: marcel@cs.cas.cz) process.
that means λk
N
f (µk ) = (µk )k−1 exp(−λµk ) .
1 X k!
CI (l) = lim D(x0i , l) . Proof: It is omitted here.
N →∞ N
i=1
We found that when one can find a scaling of neighbors’
Thus the correlation integral is a mean of probability distribu- distances measure, in form z = rq , q is the distribution
tion mapping functions for all points of W ⊂ P . mapping exponent, then one can find a “Poisson process-
We introduce a local scaling exponent q according to the like” behavior, i.e. Erlang distribution of neighbors’ distances
following definition. measure. Usually, a measure is considered that may depend
N
on the embedding space dimension d (integer), while we use there is p(0|x) + p(1|x) = 1 and from it Sq =
P
1/riq Thus
more general distribution mapping exponent q that is a positive i=1
real number. we get the form suitable for practical computation
1/riq
P
III. C LASSIFICATION USING SCALING xi ∈Uc
p̂(c|x) = N
(2)
Here we show basic idea of multidimensional data classifi-
1/riq
P
cation using scaling and three variants of this approach. i=1
(The upper sum goes over the indexes i for which the
A. Data corresponding samples of the learning set are of class c). At the
Let the learning set U of total N samples be given. Each same time all N points of the learning set are used instead of
sample xt = {xt1 , xt2 , xtd }; t = 1, 2, ...N, xtk ∈ R; k = some finite number as in the k-NN method. Moreover, we do
1, 2, ..., d corresponds to a point in d-dimensional metric space not use the nearest point (i = 1) usually. It can be found that
Md , where d is the sample space dimension. For each xt ∈ U its influence is more negative than positive on the probability
a class function T : Rd → {1, 2, ...C} : T (xt ) = c is estimate here.
introduced. With the class function the learning set U is 2) Theory: Here we come from an assumption that the best
decomposed into disjoint classes Uc = {xt ∈ U |T (xt ) = approximation of the probability distribution of the data is
C
S closely related to the uniformity of the data space around the
c}; c ∈ {1, 2, ..., C}, Uc , ∩Ub = ∅; c, b ∈ 1, 2, ..., C; c 6= b. query point x. In cases of uniform distribution - at least in the
c=1
Cardinality of set Uc let be Nc . As we need to express neighborhood of the query point - the best results are usually
which sample is closer or further from some given point x, obtained. Therefore we approximate (polynomially expand)
we can rank points of the learning set according to distance the true distribution so that at least in the neighborhood of
ri of point xi from point x. Therefore, let points of U be the query point the distribution density function appears to be
indexed (ranked) so that for any two points xi , xj ∈ U constant.
there is i < j if ri < rj ; i, j = 1, 2, ...N , and class Now a question arises why influences of individual points of
Uc = {xi ∈ U |T (xi ) = c}. Of course, the ranking depends on a given class to the final probability that point x is of the class
point x and eventually metrics of Md . We use Euclidean (L2 ) are inversely proportional to the z = riq . Let there be Z, the
and absolute (Manhattan, L1 ) metrics here. In the following largest of all z for a given class. We have shown that variable
indexing by i means ranking just introduced. z = rq has uniform distribution with some density pz . It
holds Zpz = 1 because the integral of the distribution density
function over its support (0, Z) equals to one. If support
B. The DME method would be (0, Z1 ), Z1 < Z, then the density must be larger
This classifier uses the distribution mapping exponent al- proportionally to Z/Z1 . It means that shift of each point closer
ready introduced. to point x will enlarge the density so that it will be inversely
1) Intuitive explanation: Let us consider the partial influ- proportional to the distance of a point from point x.
ences of the individual points to the probability that point x is Theorem 2: Let the task of classification into two classes
of class c. Each point of class c in the neighborhood of point x be a given. Let the size of the learning set be N and let both
adds a little to the probability that point x is of class c, where classes have the same number of samples. Let q , 1 < q < d
c = 0, 1 is the class mark. Suppose that this contribution is the be the distribution mapping exponent, let i be the index of the
larger the closer the point considered is to point x and vice i-th nearest neighbor of point x (without respect to class), and
versa. Let p(c|x, i) be a partial contribution of the i-th nearest ri > 0 its distance from point x. Then
point to the probability that point x is of class c. Then:
1/riq
P
For the first (nearest) point i = 1 and ther is p(c|x, 1) '
1 xi ∈Uc
Sq r1q
, where we use the distribution mapping exponent q in- p(c|x) = lim N
(3)
N →∞
stead of the data space dimensionality d; Sq is proportionality 1/riq
P
constant dependent on the dimensionality and metrics used. i=1
For the second point i = 2 there is p(c|x, 2) ' Sq1rq ... And (the upper sum goes for all points of class c only) is probability
2
so on; generally for point No. i p(c|x, i) ' Sq1rq . that point x belongs to class c.
i
We add the partial contributions of individual points together Proof: can be found in [13]
by summing up 3) Generalization: Up to now we suposed two classes only
and the same number of samples of both classes in the learning
X 1 X set. For general number of C classes and of the different
p(c|x) ' p(c|x, i) = 1/riq
Sq number of the samples N1 , N2 , ...NC of individual classes
xi ∈Uc xi ∈Uc
formula (3) must be completed. In fact, the last is only a
(The sum goes over the indexes i for which the corresponding recalculation of the relative representation of the different
samples of the learning set are of class c). For both classes number of the samples in classes.
1/riq )
P
lim (1/Nc
N →∞ xi ∈Uc
p(c|x) = C
(4)
1/riq )
P P
lim (1/Nk
k=1 N →∞ xi ∈Uk
1/riq
P
1/Nc
xi ∈Uc
p̂(c|x) = C (5)
1/riq )
P P
(1/Nk
k=1 xi ∈Uk
C. CD method - correlation dimension based approach The sum goes over indexes i for which the corresponding
samples of the learning set are of class c. The estimation of
In this method it is supposed that distribution mapping
the probability that the query point x belongs to class c is
exponents for individual query points differ only slightly and
that one can use the value of correlation dimension ν instead. P
1/i
Computation has then two steps, in the first step the estimate of xi ∈Uc
correlation dimension ν is computed using any known suitable p̂(c|x) = N .
P
method and then one uses formulas (2) or (5) where ν instead 1/i
of q is used. i=1
Again, as in Section III-B we exclude the first nearest In the denominator is the so-called harmonic number HN ,
neighbor of the query point. The convergence of sums is the sum of truncated harmonic series. The hypothesis above
equally fast as in the DME method. is equivalent to the assumption that the influence of individual
A relative advantage of this approach is that estimate of points of the learning set is governed by Zipfian distribution
the correlation dimension is more exact than estimate of the (Zipf’s law) [27], [23]. There is an interesting fact that the use
distribution mapping exponent and that computation of the of 1/i has a close connection to the correlation integral and
correlation dimension is done once only in difference of the correlation dimension and thus to the dynamics and true data
DME that must be computed for each query point anew. dimensionality of processes that generate the data we wish to
1) Correlation dimension estimation: For the approximation separate.
of probability of class at a given point and classification 2) Theory:
described above, a fast and reliable method for correlation Theorem 4: Let the task of classification into two classes
dimension estimation is needed. Methods for the estimation be given. Let the size of the learning set be N and let both
of correlation dimension differ by approaches used and also classes have the same number of samples. Let i be the index
by some kind of heuristics that usually optimize the size of of the i-th nearest neighbor of point x (without considering
radius r to get a realistic estimation of correlation dimension the neighbor’s class) and ri be its distance from point x. Then
[17], [3], [25] as mentioned above. P
1/i
Averaging method x ∈U
The basic problem of correlation dimension estimation is the p(c|x) = lim i N c (6)
N →∞ P
large number of pairs that arise even for a moderate learning 1/i
set size. The idea of the correlation dimension estimation i=1
described below is based on the observation that distances (the upper sum goes over indexes i for which the corresponding
between all pairs of points can be divided into groups, each samples are of class c) is the probability that point x belongs
group associated with one (fixed) point of the learning set. to class c.
Theorem 3: Let there be a learning set of N points (sam- Proof: For proof see [15].
ples). Let the correlation integral be CI (r) and let D(x, r) be In the formula above it is seen that the approach is, in the
the distribution mapping function corresponding to point x. end a kernel approach with rather strange kernel in difference
Then, CI (r) is a mean of D(x, r) for all points of U to kernels usually used [12], [22].
Proof: For proof see [15]. It is easily seen that
We have found that for sufficiently good estimation of the P
correlation dimension one can use part of the data set only, for C C 1/i
x ∈U
X X
each point to estimate the distribution mapping exponent, and p(c|x) = lim i c =1
take the average. The part of the data set may be some number c=1 c=1
N →∞ HN
of points randomly selected from the data set. It suffice to use
100 points. The method of averaging need not be limited to and p(c|x) is a ”sum of relative frequencies of occurrence” of
the Grassberger-Procaccia algorithm. We use it analogically points of a given class c. A ”relative frequencies of occurrence”
for Takens’ algorithm [25] as well. of point i , i.e. of the i-th neighbor of query point x, is
1/i
f (i; 1, N ) =
HN
D. IINC method - the inverted indexes of neighbors classifier
In fact, f (i; 1, N ) is a probability mass function of Zipfian
1) Intuitive basis: Similar way as in Section III-B1 let us distribution (Zipf’s law). In our case p(c|x) is a sum of
assume that the influence on the probability that point x is of probability mass functions for all appearances of class c.
class c of the nearest neighbor of class c is 1, the influence of From these considerations Theorem 4 above was formulated.
the second nearest neighbor is 1/2, the influence of the third 3) The Classifier construction: Let samples of the learning
nearest neighbor is 1/3 etc. Again we add the partial influences set (i.e. all samples regardless of the class) be sorted according
of individual points together by summing up to their distances from the query point x. Let indexes be
X X assigned to these points so that 1 is assigned to the nearest
p̂(c|x) = p1 (c|x, ri ) = K 1/i. neighbor, 2 to the second nearest neighbor etc. This sorting
xi ∈Uc xi ∈Uc is important difference to both methods described before
that need no sorting when distribution mapping exponent averaging over 100 different runs. It is seen that in this
or cortrelation dimension are known. Let us compute sums synthetic experiment, the DME based method presented here
S0 (x) = N10 1/i and S1 (x) = N11 reliably outperforms all other methods shown and for a large
P P
1/i, i.e. the
xi ∈U0 xi ∈U1 number of samples fast approaches to the Bayes limit.
sums of the reciprocals of the indexes of samples from class
c = 0 and from class c = 1. N0 and N1 are the numbers of
samples of class 0 and class 1, respectively, N0 + N1 = N .
The probability that point x belongs to class 0 is
S0 (x)
p̂(c = 0|x) = (7)
S0 (x) + S1 (x)
and similarly the probability that point x belongs to class 1 is
S1 (x)
p̂(c = 1|x) = (8)
S0 (x) + S1 (x)
When some discriminant threshold θ is chosen then if p̂(c =
1|x) > θ point x is of class 1 else it is of class 0. This is the
same procedure as in other classification approaches where the
output is an estimation of probability (naive Bayes) or any real
valued variable (neural networks). The value of threshold can Fig. 3. Comparison of the classification errors of the synthetic data for the
be optimized with respect to minimal classification error. The different approaches in dependence on the size of the learning set. In the
default value of the discriminant threshold here is θ = 0.5. legend, 1-NN(L2) means the 1-NN method with Euclidean metrics, CW, PW,
4) Generalization: Formulas above hold for two class prob- and CPW are three variants of the method by Paredes and Vidal [19]; the
lem with equal number of samples of both classes in the points are estimated from this reference. Bayes means the Bayes limit, DME
means the basic method presented here.
learning set. For larger number of classes and a different
number of samples of classes formula has the form similar
Note that in this test, the error of the DME estimation is
to (5):
combined with numerical errors, and with a negative influence
P of the low number of the samples giving the results presented
1/Nc 1/i
xi ∈Uc
in Fig. 3.
p̂(c|x) = C (9)
P P
(1/Nk 1/i) B. Data from Machine Learning Repository
k=1 xi ∈Uk
The classification ability of the algorithm (DME) was
It is only a recalculation of the relative representation of
tested using real-life tasks from the UCI Machine Learning
different numbers of samples of one and the other class. For
Repository; see [1]. Seven databases have been used for the
classification into more than two classes we use this formula
classification task, see Table 1.
for all classes and we assign to the query point x a class c for
which p̂(c|x) is the largest. TABLE 1. Classification mean square errors for four different meth-
ods including DME.
IV. E XPERIMENTS
We demonstrate the features and the power of the classifier
both on synthetic and real-life data.
A. Synthetic Data
Synthetic data according to Paredes and Vidal [19] is two-
dimensional and consists of three two-dimensional normal
distributions with identical a-priori probabilities. If µ denotes
the vector of the means and Cm is the covariance matrix, there
is
Class A : µ = (2, 0.5)t , Cm = (1, 0; 0, 1) (identity matrix)
Class B : µ = (0, 2)t , Cm = (1, 0.5; 0.5, 1)
Class C : µ = (0, −1)t , Cm = (1, −0.5; −0.5, 1).
Fig. 3 shows the results obtained by the different methods
for the different learning sets sizes from 8 to 256 samples and For the Shuttle data, the learning and testing sets are directly
a testing set of 5000 samples all from the same distributions at hand and were used as they are. For smaller data sets a
and independent. Each point in the figure was obtained by cross validation of 10 or 9 was used. The Iris data set was
VI. D ISCUSSION
We found that when one can find a scaling of neighbors’
distances measure, in form z = rq , q is the distribution
mapping exponent, then one can find a “Poisson process-
like” behavior, i.e. Erlang distribution of neighbors’ distances
measure. Usually, a measure is considered that may depend
on the embedding space dimension d (integer), while we use
more general distribution mapping exponent q that is a positive
real number.
Because the Erlang distribution converges to Gaussian dis-
tribution for index k → ∞, the result according to Theorem
Fig. 4. Comparison of the classification errors for four different methods 1 also relates to some results of e.g. [2], [8], [24] about
including the DME. Note that for the Mushroom data, both the 1-NN and convergence of near-neighbor distances.
DME algorithms give zero error. For the Shuttle data, the errors are ten times
enlarged.
The correlation dimension, eventually multifractal dimen-
sion, singularity (or Hölder) exponent or singularity strength,
is often used for characterization of one dimensional or two-
V. O PEN PROBLEMS dimensional data, i.e. for signals and pictures. Our results are
valid for multidimensional data that need not form a series
We have shown the use of scaling for data classification. The because in this respect data is considered as individual points
three classifiers presented here were tested and can be used as in a multidimensional space with proper metrics.
they are similarly as one sometimes uses the 1-NN or k-NN Our model of the polynomial expansion of the data space
methods. On the other hand, preprocessing like data editing comes from the demand to have a uniform distribution of
or some kind of learning may essentially enhance classifier‘s points, at least locally. There is an interesting relationship be-
behavior. tween the correlation dimension and the distribution mapping
exponent. The former is a global feature of the fractal or data
A. Editing generating process; the latter is a local feature of the data
This is a way of learning data set modification that tries set and is closely related to a particular query point. On the
enhance especially borders between classes to make class other hand, if linear regression were used, the computational
recognition easier. Original idea of editing (or preclassification) procedure is almost the same in both cases. Moreover, it can
[26] is to classify a sample of the learning set by the method be found that the values of the distribution mapping exponent
for which edited learning data will be used. If classification lie sometimes in a narrow, sometimes in a rather wide interval
result does not correspond to the sample class, remove this around its mean value. Not surprisingly, the mean value of
sample from the learning set else leave it there. After this the distribution mapping exponent over all samples is not far
is done, use the edited learning set for data classification from the correlation dimension. Introducing the notion of the
by standard way. There are another ingenious methods that distribution mapping exponent and the polynomial expansion
modify originally simple methods with help of learning, e.g. of the distances may be a starting point for a more detailed
the learning weigting method [19] modifies the learning set description of the local behavior of the multivariate data and
by weighting classes and features and then uses simple 1-NN for the development of new approaches to the data analysis,
method similarly as [26]. including classification problems.
Our experiments demonstrate that the simplest classifier [10] G RASSBERGER , P., P ROCACCIA , I. (1983) Measuring the strangeness
based on the ideas introduced here can outperform other of strange attractors, Physica, Vol. 9D, pp. 189-208.
methods for some data sets. In all the tasks presented here, [11] G UERRERO , A., S MITH , L.A. (2003) Towards coherent estimation of
the distribution-mapping-exponent-based method outperforms correlation dimension. Physics letters A, Vol. 318, pp. 373-379.
or is comparable to the 1-NN algorithm and in six of the seven [12] H ERBRICH , R. (2002) Learning Kernel Classifiers. Theory and Algo-
rithms. The MIT Press, Cambridge, Mass., London, England.
tasks, outperforms naive Bayes algorithm being only a little
[13] J I ŘINA , M. (2013) Utilization of singularity exponent in nearest
bit worse for the Splice data. All of these comparisons include neighbor based classifier. Journal of Classification (Springer) Vol. 30, No.
an uncertainty in the computation of the distribution mapping 1, pp. 3-29.
exponent. By the use of the notion of distance, i.e. a simple [14] J I ŘINA , M. (2014) Correlation Dimension-Based Classifier. IEEE
transformation En → E1 , the problems with dimensionality Transaction on Cybernetics Vol. 44, in print.
are easily eliminated at a loss of information on the true [15] J I ŘINA , M., J I ŘINA , M., JR . (2014) Classification Using the Zipfian
distribution of the points in the neighborhood of the query Kernel. Journal of Classification (Springer). Vol. 31, in print.
point which does not seem to be fundamental. [16] M ANDELBROT, B. B. (1982) The Fractal Geometry of Nature. W. H.
Freeman & Co; ISBN 0-7167-1186-9.
[17] M O , D., H UANG , S.H. (2012) Fractal-Based Intrinsic Dimension
VII. C ONCLUSION Estimation and Its Application in Dimensionality Reduction. IEEE Trans
This work was motivated by observation that near neighbors on Knowledge and Data Engineering. Vol. 24 no. 1, pp. 59-71.
distances in homogenous Poisson processes in Rd have, in fact, [18] O SBORNE ,A. R., P ROVENZALE ,A.(1989) Finite correlation dimension
the Erlang distribution modified so, that independent variable for stochastic systems with power-law spectra. Physica D, Vol. 35, pp.
357-381, (1989).
is substituted by term Krd , where K is a constant, r the
[19] R. PAREDES ,R., V IDAL ,E.(2006) Learning Weighted Metrics to Mini-
distance of the neighbor and d the space dimension. This is mize Nearest Neighbor Classification Error. IEEE Transactions on Pattern
the scaling function in exponential form. Here we answer a Analysis and Machine Intelligence, Vol. 20, No. 7, pp. 1100-1110.
question, what if point process has arisen from underlying [20] P ROKE ŠOV À , M., H AHN , U., AND J ENSEN , E. B. V. (2006) Statistics
process with scaling exponent lesser than space dimension d. for locally scaled point processes. In Baddeley, A., Gregori, P., Mateu, J.,
This problem is solved by introduction of a distribution Stoica, R., and Stoyan, D. (editors), Case Studies in Spatial Point Process
mapping function and its power approximation. It has been Modelling, Lecture Notes in Statistics, Springer, New York, Vol. 185, pp.
99-123.
shown that the distribution mapping exponent of the power
[21] P ROKE ŠOV À , M. (2010) Inhomogeneity in spatisal Cox point processes
approximation is very close to the scaling exponent known - location dependent thinning is not the only option. Image Anal Stereol
from the theory of fractals and multifractals. It leads simplified, Vol.29, pp. 133-141.
in the end, to strange scale measured by scaling exponent- [22] S CHLKOPF, B., S MOLA , A.J. (2002) Learning with Kernels. Support
power of neighbors’ distances. It was then found that when Vector Machines, Regularization, Optimization, and Beyond. The MIT
using thus scaled measure for distance of the k-th neighbor one Press, Cambridge, Mass., London, England.
can construct simple and effective classifier; we have presented [23] S CHMULAND , B.(2003) Random Harmonic Series. American Mathe-
here three its variants and discussed some open problems. matical Monthly Vol 110, pp. 407-416, May 2003.
[24] S ILVERMAN , B. W. (1976) Limit theorems for dissociated random
variables. Advances in Applied Probability, Vol. 8, pp.806-819.
R EFERENCES [25] TAKENS , F. (1985) On the Numerical Determination of the Dimension
[1] BACHE , K., L ICHMAN , M.(2013) UCI Machine Learning Repository. of the Attractor. In: Dynamical Systems and Bifurcations, in: Lecture
Irvine, CA: University of California, School of Information and Computer Notes in Mathematics, Vol. 1125, Springer, Berlin, p. 99-106.
Science, [online], 2013. Available: ¡http://archive.ics.uci.edu/ml¿. [26] W ILSON ,D.L. (1972) Asymptotic properties of nearest neighbor rules
[2] B ONETTI M, PAGANO M. (2005) The interpoint distance distribution using edited data. IEEE Trans on System Man and Cybernetics, Vol. SMC-
as a descriptor of point patterns, with an application to spatial disease 2, No. 3, pp. 408-421. (July 1972)
clustering. Stat. Med. Vol. 24, No. 5, pp. 753-773. [27] Z IPF, G.K.(1968) The Psycho-Biology of Language. An Introduction
[3] C AMASTRA , F. (2003) Data dimensionality estimation methods: a to Dynamic Philology . The MIT Press, 1968.
survey. Pattern Recognition, Vol. 6, pp. 2945-2954.
[4] DALEY, D. J., V ERE -J ONES , D. (2005) An Introduction to the Theory
of Point Processes. Volume I, Elementary theory and methods. Second
edition, Springer.
[5] DALEY, D.J., V ERRE -J ONES , D. (2008) An Introduction to the Theory of
Point Processes. Volume II, General Theory and Structure. Second Edition,
Springer.
[6] D IGGLE , P.J. (2003) A statistical Analysis of Spatial Point Processes.
Arnold, London.
[7] DVORAK ,I., K LASCHKA ,J. (1990) Modification of the Grassberger-
Procaccia algorithm for estimating the correlation exponent of chaotic
systems with high embedding dimension. Physics Letters A, Vol. 145,
No. 5, pp. 225-231.
[8] E VANS , D. (2008) A law of large numbers for nearest neighbour
statistics. Proc. R. Soc. A. Vol. 464, pp. 3175–3192.
[9] F RIEDMANN ,J. H. (1994) Flexible Metric Nearest Neighbor Classifica-
tion. Technical Report, Dept. of Statistics, Stanford University, 32 p.
Keywords: Data mining, Association rule, Factor analysis, Principal component analysis, Cluster, K-means and Euclidian distance,
Hamming distance.
form the association rule using the principal component In first section, abstract and introduction have been
analysis and factor analysis. furnished. In second section, brief methodologies of the
Markus Z¨oller[16] has discussed the ideas, assumptions and models have been furnished. In third section, implementation
purposes of PCA (Principal component analysis) and FA. A has been furnished in detail. In fourth section, result and
comprehension and the ability to differ between PCA and FA conclusion have been furnished.
have also been established.
Hee-Ju Kim [17] has examined the differences between 2. Methodology:
common factor analysis (CFA) and principal component 2.1 Data Mining
analysis (PCA). Further the author has opined that CFA A formal definition of data mining (DM) is also known as data
(Common factor analysis) provided a more accurate result as fishing, data dredging (1960), knowledge discovery in
compared to the PCA (Principal component analysis). databases (1990), or depending on the domain, as business
Diana D. Suhr [18] has discussed similarities and differences intelligence, information discovery, information harvesting or
between PCA (Principal component analysis) and EFA. data pattern processing.
Examples of PCA (Principal component analysis) and EFA Knowledge Discovery in Databases (KDD) is the non-trivial
with PRINCOMP and FACTOR have also been illustrated and process of identifying valid, novel, potentially useful, and
discussed. ultimately understandable patterns in data.
Several authors have used data mining techniques [2- By data the definition refers to a set of facts (e.g. records in a
15] for association rule generation and the selection of best database), whereas pattern represents an expression which
rules among the extracted rule. Certain authors have made a describes a subset of the data, i.e. any structured
comparative study regarding the performance principal representation or higher level description of a subset of the
component analysis [16]-[18] and factor analysis [16]-[18]. data. The term process designates a complex activity,
During the extraction of association rule, it has been comprised of several steps, while non-trivial implies that some
observed that some of rules have been ignored. In many cases, search or inference is necessary, the straightforward derivation
the algorithms generate an extremely large number of of the patterns is not possible. The resulting models or patterns
association rules, often in thousands or even in millions. should be valid on new data, with a certain level of
Further, the component of association rule is sometimes very confidence. The patterns have to be novel for the system and
large. It is nearly impossible for the end users to comprehend that have to be potentially useful, i.e. bring some kind of
or validate such large number of complex association rules, benefit to the analyst or the task. Ultimately, these should be
thereby limiting the usefulness of the data mining results. interpretable, even if this requires some kind of result
Several strategies have been proposed to reduce the number of transformation.
association rules, such as generating only “interesting” rules, An important concept is that of interestingness, which
generating only “nonredundant” rules, or generating only normally quantifies the added value of a pattern which
those rules satisfying certain other criteria such as coverage, combines validity, novelty, utility and simplicity. This can be
leverage, lift or strength. expressed either explicitly, or implicitly, through the ranking
In order to eliminate the problems, the technique of performed by the DM (Data Mining) system on the returned
factor analysis and principal component analysis has been patterns. Initially DM (Data Mining) has represented a
applied on the available data to reduce the number of variables component in the KDD (Knowledge Discovery in Databases)
in the rules. Further it has been observed that if data are not in process which is responsible for finding the patterns in data.
clean form poor result may be achieved. Therefore the data 2.2 Data Preprocessing
mining preprocessing techniques like data cleansing, data Data preprocessing is an important step in the data
integration, data transformation and data reduction have to be mining process. The phrase "garbage in, garbage out" is
applied on the available data to clean it in proper form to form particularly applicable to data mining. Data gathering methods
the association rule in data mining. Thereafter, method of handle resultant in out-of-range values (e.g., Income: 100),
factor analysis and principal component analysis has been impossible data combinations (e.g., Gender: Male, Pregnant:
applied on the dataset. The comparative study of factor Yes), missing values, etc. Analyzing data that has not been
analysis and principal component analysis has been made by carefully screened for such problems can produce misleading
forming the different number of clusters on the total effect results. Thus, the representation and quality of data is
value as formed using factor analysis and principal component necessary to be reviewed before running an analysis.
analysis. The distance measure function has been applied on If there exits irrelevant and redundant information or noisy
the different number of cluster as formed using factor analysis and unreliable data, knowledge discovery during the training
and principal component analysis to select the preferable phase becomes difficult. Data preparation and filtering steps
model. The bit equal distance measure function has been can take considerable amount of processing time. Data
proposed to select the clustering elements based on the same cleaning (or data cleansing) routines attempt to fill in missing
number of bit of two elements. Here Iris Flower data set has values, smooth out noise while identifying outliers, and
been used as data set. correct inconsistencies in the data. The data processing and
data post processing depends on the user to form and
represent the knowledge of data mining.
Data preprocessing includes the following techniques: consequently these are not purchased so often as the threshold
(1) Data cleaning: This technique includes fill in missing required, but association rules between those expensive items
values, smooth noisy data, identify or remove outliers and are as important as other frequently bought items to the
resolve inconsistencies. retailer.
(2) Data integration: This technique includes integration of Confidence of an association rule is defined as the
multiple databases, data cubes, or files. percentage/fraction of the number of transactions that contain
(3) Data transformation: This technique includes X Y to the total number of records that contain X, where if
normalization and aggregation. the percentage exceeds the threshold of confidence an
(4) Data reduction: This technique is used to obtain reduced interesting association rule X ⇒Y can be generated.
representation in volume but to produce the same or similar
analytical results
(5) Data discretization: This is the part of data reduction but it Confidence is a measure of strength of the association rules,
is importance especially for numerical data.
suppose the confidence of the association rule X⇒Y is 80%, it
2.3 Association Rule
means that 80% of the transactions that contain X and Y
The formal statement of association rule mining problem has
together, similarly to ensure the interestingness of the rules
been discussed by Agrawal et al. [18] in the year 1993. Let I=
specified by minimum confidence is also pre-defined by users.
{I1, I2. … Im} be a set of m distinct attributes, T be transaction
Association rule mining is to find out association rules that
that contains a set of items such that T ⊆ I, D be a database satisfy the pre-defined minimum support and confidence from
with different transaction records Ts. An association rule is an
a given database. The problem is usually decomposed into two
implication in the form of X ⇒Y, where X, Y⊂ I are sets of sub problems. One is to find those item sets whose
items called item sets, and X ∩ Y = ɸ . X is called antecedent occurrences exceed a predefined threshold in the database,
while Y is called consequent, X implies Y has formed as rule. those item sets are called frequent or large item sets. The
There are two important basic measures for association rules, second problem is to generate association rules from those
support(s) and confidence(c). Since the database is large and large item sets with the constraints of minimal confidence.
users concern only those frequently used items, usually Suppose one of the large item sets is Lk. Lk = { I1, I2,…, Ik-1,
thresholds of support and confidence are predefined by the Ik} association rules with this item sets are generated in the
users to drop those rules that are not so important or useful. following way, the first rule is ={ I1, I2,…, Ik-1}⇒ {Ik}, by
The two thresholds are called minimal support and minimal checking the confidence this rule can be determined as
confidence respectively, additional constraints of interesting important or not. Then other rules are generated by deleting
rules also can be specified by the users. Support(s) of an
the last items in the antecedent and inserting it into the
association rule is defined as the percentage/fraction of
consequent item, further the confidence of the new rules are
records that contain X Y to the total number of records in checked to determine the importance of them. Those processes
the database. The count for each item is increased by one in are iterated until the antecedent item becomes empty.
every time when the item is encountered in different 2.4 Factor Analysis
transaction T in database D during the scanning process. It Factor analysis is a statistical method used to
means the support count does not take the quantity of the item describe variability among observed, correlated variables in
into account. For example in a transaction a customer buys terms of a potentially lower number of unobserved variables
three bottles of beers but the support count number of {beer}is called factors. In other words, it is possible, that variations in
increased by one, in another word if a transaction contains an fewer observed variables mainly reflect the variations in total
item then the support count of this item is increased by one. effect. Factor analysis searches for such joint variations in
Support(s) is calculated by the following formula: response to unobserved latent variables. The observed
variables are modeled as linear combinations of the potential
factors, plus "error" terms. The information gained about the
From the definition it can be mentioned that the support of an interdependencies between observed variables can be used
item is a statistical significance of an association rule. Suppose later to reduce the set of variables in a dataset.
the support of an item is 0.1%, it means only 0.1 percent of the Computationally this technique is equivalent to low rank
transaction contains purchasing of this item. The retailer will approximation of the matrix of observed variables. Factor
not pay much attention to such kind of items that are not analysis originated in psychometrics, and is used in behavioral
bought so frequently, obviously a high support is desired for sciences, social sciences, marketing, product
more interesting association rules. Before the mining process, management, operations research, and other applied sciences
users can specify the minimum support as a threshold, which that deal with large quantities of data.
means the user are only interested in certain association rules 2.5 Principal component analysis (PCA) is a statistical
that are generated from those item sets whose support value procedure that uses orthogonal transformation to convert a set
exceeds that threshold value. However, sometimes even the of observations of possibly correlated variables into a set of
item sets are not as frequent as defined by the threshold, the values of linearly uncorrelated variables called principal
association rules generated from them are still important. For components. The number of principal components is less than
example in the supermarket some items are very expensive, or equal to the number of original variables. This
transformation is defined in such a way that the first principal 1. Place K points into the space represented by the
component has the largest possible variance (that is, accounts objects that are being clustered. These points
for as much of the variability in the data as possible), and each
represent initial group of centroids.
succeeding component in turn has the highest variance
possible under the constraint that it be orthogonal to (i.e., 2. Assign each object to the group that has the closest
uncorrelated with) the preceding components. Principal centroid.
components are guaranteed to be independent if the data set is 3. When all objects have been assigned, recalculate the
jointly normally distributed. PCA is sensitive to the relative positions of the K centroids.
scaling of the original variables. 4. Repeat Steps 2 and 3 until the centroids no longer
2.6 Clustering move. This produces a separation of the objects into
The Clustering method deals with finding a structure in a
groups from which the metric to be minimized can be
collection of unlabeled data. A loose definition of clustering
could be “the process of organizing objects into groups whose calculated.
members are similar in some way”. A cluster is therefore a Although it can be proved that the procedure will always
collection of objects which are “similar” between them and are terminate, the k-means algorithm does not necessarily find the
“dissimilar” to the objects belonging to other clusters. The most optimal configuration, corresponding to the global
graphical example of cluster has shown in figure 1.9. In this objective function minimum. The algorithm is also
case it is easily identified the 4 clusters into which the data can significantly sensitive to the initial randomly selected cluster
be divided, the similarity criterion is that two or more objects centers. The k-means algorithm can be run multiple times to
belong to the same cluster if they are “close” according to a reduce this effect. K-means algorithm is a simple algorithm
given distance (in this case geometrical distance). This is that has been adapted to many problem domains.
called distance-based clustering. 2.6.3 Distance Measure
2.6.1 K-Means Clustering An important component of a clustering algorithm is the
K-Means (MacQueen, 1967) clustering algorithm is one of the distance measure between data points. If the components of
simplest unsupervised learning algorithms that solve the well the data instance vectors are in the same physical units it is
known clustering problem. The procedure follows a simple possible that the simple successfully group similar data
and easy way to classify a given data set through a certain instances. However, even in this case the Euclidean distance
number of clusters (assume k clusters) with a priori. The main can sometimes be misleading.
idea is to define k centroids, one for each cluster. These 2.6.3.1 Euclidean Distance
centroids should be placed in a cunning way because different According to the Euclidean distance formula, the distance
location causes different result. So, the better choice is to between two points in the plane with coordinates (x, y) and (a,
place them as much as possible far away from each other. The b) is given by
next step is to take each point belonging to a given data set dist ((x, y), (a, b)) = √(x - a)² + (y - b)²
and associate it to the nearest centroid. When no point is
pending, the first step is completed and an early groupage is 2.6.3.2 Hamming Distance
done. At this point it is needed to re-calculate k new centroids The Hamming distance between two strings of equal length is
as barycenters of the clusters resulting from the previous step. the number of positions at which the corresponding symbols
After these k new centroids have been obtained a new binding are different. In another way, it measures the minimum
has to be done between the same data set points and the number of substitutions required to change one string into the
nearest new centroid. A loop has been formed. As a result of other, or the minimum number of errors that could have
this loop it may be noticed that the k centroids change their transformed one string into the other.
location step by step until no more changes are done. In other The Hamming distance between:
words centroids do not move any more. • 1011101 and 1001001 is 2.
Finally, this algorithm aims at minimizing an objective • 2173896 and 2233796 is 3.
function, generally a squared error function. The objective
2.6.3.3 Proposed Bit equal
function as It has been observed that hamming distance based on the
minimum number of errors that could have transformed one
where is a chosen distance measure between a
string into the other. But clustering techniques based on
data point and the cluster centre . grouping of object of similar types therefore here an effort has
2.6.1.2 K-Means Algorithm been made to group the object of similar type rather than
based on the dissimilarity of strings (hamming distance). The
bit equal distance measure function has been proposed to
count the same number of bit of two elements. For an
example, consider the strings 1011101 and 1001001, the
hamming distance is 2, but in bit equal (Consider the strings
have equal number bit) the similarity is 5. Here we can say
that five bits of the elements are same based the similarity 21 2.4 3.3 1 4.9
properties of clustering. 22 2.9 4.6 1.3 6.6
23 2.7 3.9 1.4 5.2
3. Implementation: 24 2 3.5 1 5
The available data contains the information of Iris flowers 25 3 4.2 1.5 5.9
with various items which have been furnished column wise in 26 2.2 4 1 6
table 1. The data related to iris flower containing 40 data 27 3.3 6 2.5 6.3
values which have been shown in table 2. It is to note that the 28 2.7 5.1 1.9 5.8
quality of flower depends on the sepal length of the flower. If 29 3 5.9 2.1 7.1
the sepal length of a particular flower is known, the quality of 30 2.9 5.6 1.8 6.3
flower can be ascertained. Therefore the sepal length of flower 31 3 5.8 2.2 6.5
(D) has been chosen as the objective item (consequent item) of 32 3 6.6 2.1 7.6
the flowers. The other parameters i.e. sepal width (A), petal 33 2.5 4.5 1.7 4.9
length (B) and petal width (C) have been chosen as the 34 2.9 6.3 1.8 7.3
depending items (antecedent items). 35 2.5 5.8 1.8 6.7
The purpose of this work is to correlate the items A, B and C 36 3.6 6.1 2.5 7.2
with D, so that based on any value of A, B and C, the value of 37 3.2 5.1 2 6.5
D can be estimated. From the value of D the quality of that 38 2.7 5.3 1.9 6.4
type of flower can be ascertained. 39 3 5.5 2.1 6.8
Table 1 40 2.5 5 2 5.7
Flower Characteristics
Item Name Item description The data mining preprocessing techniques like data cleansing,
A Sepal width data integration, data transformation and data reduction have
B Petal length been applied on the available data as follows:
C Petal width Data Cleansing
D Sepal length The data cleansing techniques include the filling in the
missing value, correct or rectify the inconsistent data and
identify the outlier of the data which have been applied on the
Now it is necessary to check whether the data items are proper
available data.
or not. If the data items are proper, extraction of information is
It has been observed that each data set does not contain any
possible otherwise the data items are not suitable for the
extraction of knowledge. In that case preprocessing of data is missing value. The said data item does not contain any
necessary for getting the proper data. The set of 40 data inconsistent data i.e. any abnormally low or any abnormally
high value. All the data values are regularly distributed within
elements has been taken which have been furnished in table 2.
the range of that data items. Therefore the data cleansing
Table 2
techniques are not applicable for the available data.
Available Data
Data Integration
Serial
Number A B C D The data integration technique has to be applied if data has
1 3.5 1.4 0.2 5.1 been collected from different sources. The available data have
2 3 1.4 0.2 4.9 been taken from a single source therefore the said technique is
3 3.2 1.3 0.2 4.7 not applicable here.
4 3.1 1.5 0.2 4.6 Data transformation
5 3.6 1.4 0.2 5 The data transformation techniques such as smoothing,
6 3.9 1.7 0.4 5.4 aggregation, normalization, decimal scaling have to be applied
7 3.4 1.4 0.3 4.6 to get the data in proper form. Smoothing technique has to be
8 3.4 1.5 0.2 5 applied on the data to remove the noise from the data.
9 2.9 1.4 0.2 4.4 Aggregation technique has to be applied to summarize the
10 3.1 1.5 0.1 4.9 data. To make the data within specific range smoothing and
11 3.7 1.5 0.2 5.4 normalization technique have to be applied. Decimal scaling
12 3.4 1.6 0.2 4.8 technique has to be applied to move the decimal point to
13 3 1.4 0.1 4.8 particular position of all data values. Out of these data
14 3.2 4.7 1.4 7 transformation techniques, smoothing and decimal scaling
15 3.2 4.5 1.5 6.4 techniques have been applied on the data as furnished in table
16 3.1 4.9 1.5 6.9 2 to get the revised data as furnished in table 3.
17 2.3 4 1.3 5.5 Table 3
18 2.8 4.6 1.5 6.5 Revised Data
19 2.8 4.5 1.3 5.7 Serial
20 3.3 4.7 1.6 6.3 Number A B C D
3 653 600 have been assigned a range which has been furnished in table
4 653 600 6.
5 703 600 Table 6
6 703 600 Assigned Range of the Data Element
7 703 600 Serial A B C D Range of Data
8 703 600 Number
9 653 600
1 3500 1400 200 5100 A3 B1 C1 D1
10 653 600
2 3000 1400 200 4900 A2 B1 C1 D1
11 703 600
3 3200 1300 200 4700 A2 B1 C1 D1
12 703 600
4 3100 1500 200 4600 A2 B1 C1 D1
13 653 600
5 3600 1400 200 5000 A3 B1 C1 D1
14 753 700
6 3900 1700 400 5400 A3 B1 C1 D1
15 753 650
7 3400 1400 300 4600 A3 B1 C1 D1
16 803 700
8 3400 1500 200 5000 A3 B1 C1 D1
17 703 650
9 2900 1400 200 4400 A2 B1 C1 D1
18 753 650
10 3100 1500 100 4900 A2 B1 C1 D1
19 753 650
11 3700 1500 200 5400 A3 B1 C1 D1
20 803 650
12 3400 1600 200 4800 A3 B1 C1 D1
21 753 600
13 3000 1400 100 4800 A2 B1 C1 D1
22 753 700
14 3200 4700 1400 7000 A2 B1 C2 D3
23 753 600
15 3200 4500 1500 6400 A2 B1 C2 D2
24 703 600
16 3100 4900 1500 6900 A2 B2 C2 D3
25 753 650
17 2300 4000 1300 5500 A1 B2 C2 D2
26 703 650
18 2800 4600 1500 6500 A2 B2 C2 D2
27 903 650
19 2800 4500 1300 5700 A2 B2 C2 D2
28 853 650
20 3300 4700 1600 6300 A3 B2 C2 D2
29 853 700
21 2400 3300 1000 4900 A1 B2 C2 D1
30 853 650
22 2900 4600 1300 6600 A3 B2 C2 D3
31 853 650
23 2700 3900 1400 5200 A2 B2 C2 D1
32 853 700
24 2000 3500 1000 5000 A1 B2 C2 D1
33 753 600
25 3000 4200 1500 5900 A2 B2 C2 D2
34 853 700
26 2200 4000 1000 6000 A1 B2 C2 D2
35 803 700
27 3300 6000 2500 6300 A3 B3 C3 D2
36 903 700
28 2700 5100 1900 5800 A2 B3 C3 D2
37 853 650
29 3000 5900 2100 7100 A2 B3 C3 D3
38 853 650
30 2900 5600 1800 6300 A2 B3 C3 D2
39 853 700
31 3000 5800 2200 6500 A2 B3 C3 D2
40 803 650
32 3000 6600 2100 7600 A2 B3 C3 D3
33 2500 4500 1700 4900 A1 B3 C3 D1
Step 4 34 2900 6300 1800 7300 A2 B3 C3 D3
It is to note that numerical values have been assigned to each 35 2500 5800 1800 6700 A1 B3 C3 D2
item by dividing the items values into certain range of values. 36 3600 6100 2500 7200 A3 B3 C3 D3
Here the items A, B, C and D have been divided into three 37 3200 5100 2000 6500 A2 B3 C3 D2
ranges of values. Further it is to note that it may happen that 38 2700 5300 1900 6400 A2 B3 C3 D2
each item may have to divide into unequal range of values. In 39 3000 5500 2100 6800 A2 B3 C3 D3
that case it may be difficult to handle range of values. Apart 40 2500 5000 2000 5700 A1 B3 C3 D2
from that, it may happen that a particular consequent item may
be related to a number of cumulative antecedent items with Here A1, A2 and A3 occurred 11, 22 and 7 times respectively.
huge difference in data values. Similarly B1, B2, B3, C1, C2, C3, D1, D2 and D3 have been
The support and confidence has to be found on Iris data for the occurred 15, 12, 13, 13, 13, 14, 17, 15 and 8 respectively.
application of Apriori algorithm. In case of Iris flower data all According to Apriori algorithm support of A1 will be 11/40
the data in numeric form where it is not possible to calculate (27.5%), A2=22/40 (55%), A3=7/40 (17.5%), B1=15/40
the confidence and support using Apriori algorithm method (37.5%), B2=12/40, (30%) B3=13/40, (32.5%) C1=13/40
mentioned by Agrawal et al. [18] in the year 1993. Therefore (32.5), D1= (42.5%), D2=15/40(37.5%) and D3=8/15(20%).
an effort has to be made to calculate the confidence and If support is less than 30% then ignore the rule. Here A3 will
support in following ways. To calculate the support and be ignored from the set of rule. Therefore it is necessary to
confidence all the data have to be converted into range which ignore the rule where A3 has occurred. Therefore, seven rules
has been furnished in table 4. Base on the table 6 the all data have to be ignored from the rule set. This action further
ignored the rules which are associated with A3. For an
Antecedent vs Consequent
ean 9.2 6 6.7 3.85 7.5 6.2 5.7 .41
Ham 369 347 314 269 403 399 366 292
ming
Cumulative
20000 Antecedent Item Table 17
(FA) Data used Boston city Data Set
15000 Cumulative PCA vs. Factor Analysis Using Different Clusters
Range of data
everyday problems belong to R, and mostly to R+. Probability According to Chapman – Kolmogorov relation for discrete
of occurrence of an event always belongs to the range [0,1]. In Markov matrices (Karlin and Taylor (1975)), it can be proved
this paper, it is discussed that to solve a special class of that
Markov chain which should have solution in real world, we
are confronted with "analytic probabilities"!. Though the name (4)
probability applies to the values between zero and one, we Pn that is P to the power n is a Markov matrix if P is
define a special analogue measure of probability as complex Markov.
probability where the conventional probability is a subclass of Now, suppose that we intend to derive the t-step transition
this newly defined measure. probability matrix P(t) where t≥0 from the above (3) and (4)
Now define the well-known discrete time Markov chain {Yn } definition of n-step transition probability matrix P. That is, to
find the transition probability matrix for incomplete steps. On
a Markov stochastic process whose state space is
the other hand, we are interested to find the transition matrix
s = {1,2,..., N } for which T = {0,1, 2,...} . Refer to the value P(t) when t is between two sequential integers. This case is not
of Yn as the outcome of the nth trial. We say Yn being in state i just a tatonnement example. To clarify the application of this
if Yn = i. The probability of Yn+1 being in state j, given that Yn phenomenon, consider the following example.
is in state i (called a one–step transition probability) is denoted Example 1. Usually in population census of societies with N
n , n +1 distinct regions, migration information is collected in an NxN
by Pij , i.e.,
migration matrix for a period of ten years. Denote this matrix
+1
Pijn , n= P=
ij Pr {Y n +=
1 j Y=
n i } (1) by M. Any element of M, mij is the population who leaved
region i and went to region j through the last ten years. By
Therefore, the Markov or transition probability matrix of deviding each mij to sum of the ith row of M, a value of Pij is
the process is defined by computed as an estimate of probability of transition from ith to
jth regions. Thus, the stochastic matrix P gives the probabilities
of going from region i to region j in ten years (which is one–
step transition probability matrix). The question is: how we
Bijan Bidabad is WSEAS Post Doctorate Researcher, (No. 2, 12th St.,
Mahestan Ave., Shahrak Gharb, Tehran, 14658, IRAN. (Tel.: +98-21- can compute the transition probability matrix for one year or
88360810, Mobile: +98-912-1090164, Fax: +98-21-88369336, email: one-tenth step and so on.
bijan@bidabad.com , web: http://www.bidabad.com )
Behrouz Bidabad, Faculty of mathematics, Polytechnics University, Hafez
If we knew the generic function of probabilities in very
Ave., Tehran, 15914, IRAN (email: bidabad@aut.ac.ir, web: small period of time we would be able to solve problems
http://www.aut.ac.ir/official/main.asp?uid=bidabad) similar to example 1. But the generic function (5) is not
Nikos Mastorakis, Technical University of Sofia, Bulgaria, Department of obtainable. If it were, we would apply the continous time
Industrial Engineering, Sofia, 1000, BULGARIA (email: mastor@tu-sofia.bg,
web: http://elfe.tu-sofia.bg/mastorakis) Markov procedure using the generic NxN matrix A as:
probability matrix can be written as, Where λi , i ∈ S are the eigenvalues of A defined by (5), and
−1
P = XΛ X
{ }
t t
(11) Λ (t ) = diag exp (λ1t ),...,exp (λNt (19)
where,
{ }
And X is the corresponding eigenmatrix of A. Take the
Λt = diag λ1t ,...., λNt natural logarithm of (18),
(12)
ln P (t ) = XΦ (t ) X −1 (20)
By comparison of (22) and (24) conclude that With the above information, consider the following
discussions.
ln P (t ) = t ln P (1) (25)
a ) λi ∈(0,1] ∀i ∈ S
or,
P (t ) = Pt (1) (26) In this case all λti ≥ 0 for t ≥ 0 and no imaginary part
Given (15), equation (26) is the same as (16) Q.E.D. occurs in matrix Pt. λi are all positive for i belonging to S if
we can decompose the matrix P to two positive semi-definite
Result 1. Matrix Pt fulfils definition 1. Thus, Pt ⊆ Q . This and positive definite matrices B and C of the same size
comes from the following remarks. (Mardia, Kent, Bibby (1982)) as
P = C−1 B
Remark 3. Sum of each row of Pt is equal to one. Since Pt b ) λi ∈[ −1,1] , λi ≠ 0, ∀ i ∈ S
satisfies Markovian properties (theorem 1).
λti , t ≥ 0 belongs to sets of real and imaginary numbers
Remark 4. Sum of imaginary parts of each row is equal to based on the value of t. In this case Pt belongs to the class of
zero. This immediately comes from remark 3.
generalized stochastic matrix Q of definition 1. For λi ∈R , it
is sufficient that P be positive definite.
Remark 5. If qij denotes the ijth element of Pt for t ≥ 0,
c ) λi ∈C , λi ∈(0,1] ∀ i ∈S
then q ij ≤ 1 for all i and j belonging to S. This remark can be
Pt in this case for t ≥ 0 and t ∉N belongs to the class of
concluded form theorem 1.
generalized Markov matrices of definition 1.
d ) t ∈ N (Natural numbers)
Remark 6. If=
Q Pt ,t ≥ 0 equals to the complex matrix In all cases of a, b, and c we never coincide with complex
defined by (13), then V jk ≤1 ∀j , k ∈S . Since, probabilities. Since Pt can be drived by simply multiplying P, t
times.
1 ≥ q jk = u jk + iv jk ⇒ 1 ≥ u 2 jk + iv 2 jk e ) t ∈ Z (Integer numbers)
In this case, Pt is a real matrix but does not always satisfy
1 ≥u 2jk +v 2jk ⇒1≥ v jk .
condition 2 of definition 1.
f ) t ∈R −
Remark 7. Given Q as in remark 6, then ujk∈[0,1]. This also
Pt is a complex matrix but does always satisfy conditions 2
comes immediately from theorem 1.
and 3 of definition 1.
III. DISCUSSION ON BROKEN TIMES
IV. COMPLEX PROBABILITY JUSTIFICATION
The broken time discrete Markov chain is not always a
Interpretation of the "Complex probability" as defined by
complex probability matrix defined by definition 1. Matrix Pt
definition 1 is not very simple and needs more elaborations.
has different properties with respect to t and eigenvalues. λi The interesting problem is that, it exists in operational works
may be real (positive or negative) or complex depending on of statistics as the example 1 discussed. Many similar
the characteristic polynomial of P. examples like the cited may be gathered.
Since P is a non–negative matrix, Forbenius theorem With this definition of probability, the moments of a real
(Takayama (1974), Nikaido (1970)) assures that P has a random variable are complex. Although the t–step distribution
positive dominant eigenvalue πt of initial distribution π0 with respect to Pt may be
complex, they have the same total as π 0 . That is, if
(35)
And, sum of t–step distribution is
N N N
∑ π tj ∑ π oj (u j 1 + ... + u jN ) + i ∑ π oj (v j 1 ,...,v jn ) (36)
=
=j 1=j 1 =j 1
V. SUMMARY
By summarizing the discrete and continuous times Markov
stochastic processes a class of real world problems was
introduced which cannot be solved by each of the procedures.
The solutions of these problems coincide with “Complex
probabilities” of transitions that are inherent in mathematical
formulation of the model. Complex probability is defined and
some of its properties with respect to the cited class are
examined. Justification of the idea of complex probability
needs more work and is left for further research.
ACKNOWLEDGEMENTS
The authors are indebted to Dr. A. Monajemi who read the
manuscript and gave valuable remarks.
REFERENCES
[1] R. Bellman (1970), Introduction to matrix analysis, McGraw–
Hill.
[2] P.J. Dhrymes (1978), Mathematics for econometrics. Springer-
Verlag.
[3] W. Feller (1970, 1971), An Introduction to probability theory
and its applications, vols. 1,2, Wiley, New York.
[4] P.G. Hoel, S.C. Port, C. J. Stone (1972), Introduction to
stochastic processes. Houghton Mifflin, New York.
[5] S. Karlin, H.M.Taylor (1975), A first course in stochastic
processes. Academic Press.
For and
(3.2)
• The Sumudu transform can be defined for functions definition 3 we define partial derivative of order for
which are discontinuous at the origin. In that case the respect to the function
two branches of the function should be transformed
separately. (3.4)
(4.2) ,
Using the property of the Sumudu transform, we have
Finally, we approximate the analytical solution by the
truncated series:
(4.10)
(4.3)
Now applying the Sumudu inverse on both sides of (4.3) we
obtain:
The above series solution generally converges very rapidly
[33]
Or in general by putting
We obtain:
(4.12)
Substituting (4.5) and (4.6)
(4.8)
(4.14)
and initial condition
and the nonlinear term can be decomposed as Example 2
Consider the following time-fractional derivative in
(4.14) plane as (5.2)
This equivalent to the exact solution in closed form: Example 3: Apply the steps involved in HPSTM as presented
(5.12) in section 4.1 to equation (5.4) Kilicman et al [ 33] obtained
the following:
where is the Mittag-Leffler function. ,
Example 2: Applying the steps involved in HPSTM as ,
presented in section 4.1 to equation (5.2) we obtain:
,
,
,
, (5.15)
,
Therefore the series solution is given as:
(5.13)
which is the first four terms of the series expansion of the Therefore the approximate solution of equation for the first
exact solution is given below as:
. (5.18)
Example 2
V.II. Solution via HDM Following the discussion presented earlier, applying the initial
conditions and comparing the terms of the same power of p,
Example 1: Apply the steps involved in HDM as presented in integrating, we obtain the following solutions:
section 4.2 to equation (5.1) we obtain the following
(5.19)
(5.21)
Which are the first four terms of the series expansion of the
exact solution
The asymptotic solution is given by
Example 3:
,
,
(5.20)
,
(5.23)
This is the exact solution of (5.1) when .
.
(5.27)
Therefore the approximate solution of equation for the first
is given below as:
Example 4: Following carefully the steps involved in the
HDM, we arrive at the following equations
(5.29)
Table 1: Numerical results of equation (5.2) via mathematica be noted that only the fourth-order term of the HDM and
HPSTM were used to evaluate the approximate solutions for
HPSTM HPSTM HPSTM
and and and HDM
Figures 1. It is evident that the efficiency of the present
HDM HDM method can be noticeably improved by computing additional
t x y Exact Errors terms of when the HDM is used.
Abstract: Measurements using the high resolution array-comparative genomic hybridization (HR-CGH) array are
accompanied with large noise which strongly affects the estimates of the copy number variations (CNVs) and re-
sults in segmental errors as well as in jitter in the breakpoints. Based on the probabilistic analysis and algorithm
designed, we show that jitter in the breakpoints can be well approximated with the discrete skew Laplace distri-
bution if the local signal-to-noise ratios (SNRs) exceed unity. Using this distribution, we propose an algorithm
for computing the estimate upper and lower bounds. Some measurements and estimates tested using these bounds
show that the higher probe resolution is provided the more segmental accuracy can be achieved and that larger
segmental SNRs cause smaller jitter in the breakpoints. Estimates of the CNVs combined with the bounds pro-
posed may play a crucial role for medical experts to make decisions about true chromosomal changes and even
their existence.
1 Introduction 1000 base pairs of DNA [6]. The human genome with
23 chromosomes is estimated to be about 3.2 billion
base pairs long and to contain 20000 − 25000 distinct
The deoxyribonucleic acid (DNA) of a genome es-
genes [1]. Each CNV may range from about one kb to
sential for human life often demonstrates structural
several megabases (Mbs) in size [2].
changes called copy-number variations (CNVs) asso-
ciated with disease such as cancer [1]. The sell with One of the techniques employing chromosomal
the DNA typically has a number of copies of one or microarray analysis to detect the CNVs at a resolution
more sections of the DNA that results in the struc- level of 5–10 kbs is the array-comparative genomic
tural chromosomal rearrangements - deletions, dupli- hybridization (aCGH) [7]. It was reported in [8] that
cations, inversions and translocations of certain parts the high-resolution CGH (HR-CGH) arrays are accu-
[2]. Small such CNVs are present in many forms in rate to detect structural variations (SV) at resolution of
the human genome, including single-nucleotide poly- 200 bp. In microarray technique, the CNVs are often
morphisms, small insertion-deletion polymorphisms, normalized and plotted as log2 R/G = log2 Ratio,
variable numbers of repetitive sequences, and ge- where R and G are the fluorescent Red and Green
nomic structural alterations [3]. If genomic aberra- intensities, respectively [9]. An annoying feature of
tions involve large CNVs, the process was shown to such measurements is that the Ratio is highly contam-
be directly coupled with cancer and the relevant struc- inated by noise which intensity does not always al-
tural changes were called copy-number alterations low for correct visual identification of the breakpoints
(CNAs) [4]. A brief survey of types of chromosome and copy numbers and makes most of the estimation
alterations involving copy number changes is given techniques poor efficient if the number of segmental
in [5]. The copy number represents the number of readings is small. It was shown in [10] that sufficient
DNA molecules in a cell and can be defined as the quality in the CNVs mapping can be achieved with
number of times a given segment of DNA is present in tens of millions of paired reads of 29–36 bases at each.
a cell. Because the DNA is usually double-stranded, Deletions as small as 300 bp should also be detected
the size of a gene or chromosome is often measured in some cases. For instance, arrays with a 9-bp tiling
in base pairs. A commonly accepted unit of measure- path were used in [8] to map a 622-bp heterozygous
ment in molecular biology is kilobase (kb) equal to deletion. So, further progress in the probe resolution
I = [i1 i2 . . . iL ]T ∈ RL . (1)
xj > αl , 2 2 ,
σl = σl+1
words, with the probability of 99%, the breakpoint nl Alj is (6)
2 2 ,
can be found at any point between n = 40 and n = 52 αl < xj < βl , σl < σl+1
that may be too rough for medical conclusions, espe-
βl < xj < αl , σl2 < σl+1
2 ,
cially if r̄ is large. Let us add that simple averaging 2 2 ,
Blj is xj < αl , σl = σl+1 (7)
which is optimal for the estimation of PWC changes 2 2
between the breakpoints is able to reduce the noise (xj < αl ) ∧ (xj > βl ) , σl > σl+1 .
variance by the factor of Nl . Noise reduction may
The inverse events meaning that at least one of the
thus also be insufficient for medical applications if Nl
points do not belong to the relevant interval are Ālj =
is small. So, effect of noise needs more investigations
1 − Alj and B̄lj = 1 − Blj .
and the CNVs estimate bounds are required.
Both Alj and Blj can be united into two blocks
Let us now think that jitter occurs at some point where φl is specified by (16) and 0.5 <
nl ± k, 0 6 k 6 N , and assign two additional blocks {P (Al ), P (Bl )} < 1. By substituting ql =
of events P −1 (Al )−1 and dl = P −1 (Bl )−1, we find P (Al ) =
1/(1 + ql ) and P (Bl ) = 1/(1 + dl ), provide the
Alk = {Ail −N . . . Ail −1−k } , transformations, and finally go from (17) to the dis-
Blk = {Bil +k . . . Bil +N −1 } . crete skew Laplace distribution (5) in which κl and
νl still need to be connected to (17). To find κl and
The probability Pk− , Pk− (Alk Āl(il −k) . . . Āil −1 Bl ) νl , below we consider three points k = −1, k = 0,
that jitter occurs at kth point to the left and k = 1. By equating (5) and (17), we first
from nl (left jitter) and the probability obtain (1−d1−d
l )(1−ql )dl
l ql
= φ1l 1−P (Bl )
P (Bl ) for k = 1 and
Pk+ , Pk+ (Al B̄l(il +1) . . . B̄l(il +k−1) Blk ) that (1−dl )(1−ql )ql
= φ1l 1−P (Al )
for k = −1 that gives us
jitter occurs at kth point to the right from nl (right 1−dl ql P (Al )
1−κ2
jitter) can thus be written as, respectively, νl = κl ln µl l , where
Pk− = P N −k (Al )[1 − P (Al )]k P N (Bl ) , (12) P (Al )[1 − P (Bl )]
µl = . (18)
Pk+ N
= P (Al )[1 − P (Bl )] P k N −k
(Bl ) . (13) P (Bl )[1 − P (Al )]
(1−dl )(1−ql )
By normalizing (12) and (13) with (9), we arrive For k = 0, we have 1−dl ql = φ1l and trans-
at a function that turns out to be independent on N : form it to an equation (1+µl )
x2l − φl1+φ l
1−φl
x − 1+φ l
µl = 0,
which proper solution is
[P −1 (Al ) − 1]|k| , k < 0 , (left) Ã s !
fl (k) = 1 , k = 0, φl (1 + µl ) 4µl (1 − φ2l )
−1 k xl = 1− 1+ 2 (19)
[P (Bl ) − 1] , k > 0 . (right) 2(1 + φl ) φl (1 + µl )2
(14)
Further normalization of fl (k) to have a unit area −
κ2
l
leads to the pdf pl (k) = φ1l fl (k), where φl is the sum and which xl = µl
1−κ2
gives us
l
well for the allowed probability of jitter-free detec- which mean value is E{âj } = aj and variance is
tion of 90%. It narrows the jitter bounds with about
±2 points for 99%. Observing another example illus- σj2
trated in Fig. 3 for γl− = 9.25625 and γl+ = 2.61186, σ̂j2 = . (26)
Nj
Figure 3: The jitter pdf approximated using the discrete Note that κ in (30) and (31) should be specified
skew Laplace distribution and found experimentally (cir- in the θ-sense as in (27) and (28).
cled) using a ML estimator over 104 runs for γl− = 9.25625 UB and LB Masks and Algorithm. By combin-
and γl+ = 2.61186: (a) pdfs and (b) approximation errors. ing (27), (28), (32), and (33), the UB mask BnU and
the LB mask BnL can now be formed to outline the
region for true genomic changes. The relevant algo-
The UB for segmental estimates can r
be formed in rithm was designed in [41]. Its inputs are measure-
σj2 ments yn , breakpoints estimates n̂l , tolerance param-
the θ-sigma sense as âUB = E{âj } + θ Nj , where
j eter θ, number L of the breakpoints, and number of
θ > 1 is commonly integer. However, neither an ac- readings M . At the output, the algorithms produces
tual aj = E{âj } nor multiple measurements neces- two masks: BnU and BnL .
sary to approach aj by averaging are available. We The UB and LB masks have the following basic
thus specify UB and LB approximately as applied properties:
s • The true CNVs exist between BnU and BnL with
σj2 the probability determined in the θ-sigma sense.
âUB ∼
= âj + θ , (27)
j
Nj
s • If BnU or BnL covering two or more breakpoints is
σj2 uniform, then there is a probability of no changes
âLB ∼
= âj − θ . (28)
j
Nj in this region.
• If both BnU and BnL covering two or more break-
where θ = 1 guarantees an existence of true changes points are uniform, then there is a high probabil-
between UB and LB with the probability of 68.27% or ity of no changes in this region.
error probability of κ = 0.3173 that is 31.73%; θ = 2
of 95.45% or κ = 0.0555 that is 5.55% and θ = 3 of We notice again that the jitter bounds in BnU and
99.73% or κ = 0.0027 that is 0.27%. BnL may have enough accuracy if (γl− , γl+ ) > 1.
They may be considered in the lower bound sense if
Jitter Bounds. The jitter left bound (JLB) JlL
(γl− , γl+ ) < 1. However, the approximation error is
and the jitter right bound (JRB) JlR can be determined
with respect to nl as follows. Because a step is unity commonly large if (γl− , γl+ ) < 0.5. For details, see
with integer k, we specify the jitter probability at the Section 2.2.
kth point using (5) as
4 Applications
Pk (γl− , γl+ ) = p[k|d(γl− , γl+ ), q(γl− , γl+ )] . (29)
In this section, we test some CNVs measurements
We then equate (29) to κ and solve it for the right and and estimates by the UB and LB masks computed
in the three-sigma sense, θ = 3, using the algorithm Because of large noise, estimates of the CNVs
[41–43]. Because the algorithm can be applied to any may bring insufficient information to experts and must
CNVs data with supposedly known breakpoints, we be tested by UB and LB masks. To form such masks,
choose the 1st chromosome measured using the HR- the jitter distribution must be known. We have shown
CGH array in [28] and available from [44]. that jitter in the breakpoints can be modeled using
The CNVs structure has 34 segments and 33 the discrete skew Laplace distribution if the segmen-
breakpoints. Most of the segments have the SNRs tal SNRs exceed unity. Otherwise, the approximation
exceeding unity meaning that the UB and LB masks errors can be large and more profound investigations
will have enough accuracy. The SNRs in segments of jitter will be required. The UB and LB masks pro-
â18 and â21 range between 0.5 and unity which means posed in this paper in the θ-sigma sense outline the
that real jitter can be here about twice larger. The re- region within which the true changes exist with a high
maining segments â23 , â28 , â31 and â32 demonstrate probability (99.73% in the three-sigma sense). Pro-
the SNR below 0.5 that means that the jitter bounds vided the masks, information about CNVs is more
cannot be estimated with sufficient accuracy. We just complete and sometimes can be crucial for medical
may say that jitter can be much larger in the relevant experts to make a correct decision about true struc-
breakpoints. ture. Testing some measurements and estimates by
Let us consider the CNVs measurements and esti- the UB and LB masks has revealed large errors ex-
mates in more detail following Fig. 4. As can be seen, ceeding (30...50)% in many segments. It was also
there are two intervals with no measurements between demonstrated that jitter in some breakpoints is redun-
the breakpoints î15 and î16 and the breakpoints î28 and dantly large for making any decision about their true
î29 . A part of measurements covering the breakpoints locations. We finally notice that further investigations
from î5 to î14 is shown in Fig. 5a. Its specific is that must be focused on the jitter statistics at low SNR val-
the segmental SNRs are all larger than unity and the ues that is required to sketch a more correct proba-
masks thus can be used directly for practical appli- bilistic picture of the CNVs.
cations. The masks suggest that errors in all of the
segmental estimates reach tens of percents. In fact, â5
and â10 are estimated with error of about 50%. Error References:
exceeds 30% in the estimates â7 , â9 , â12 , and â13 . A [1] International Human Genome Sequencing Consor-
similar problem can be observed in the estimates of tium, “Finishing the euchromatic sequence of the
almost all of the breakpoints in which left and right human genome,” Nature, vol. 431, pp. 931-945,
jitter reaches several points. Oct. 2004.
A situation even worse with a part of the chromo- [2] P. Stankiewicz and J. R. Lupski, “Structural Varia-
some covering the breakpoints from î17 to î26 . The tion in the Human Genome and its Role in Disease,”
segmental errors exceed 50% here over almost all Annual Review of Medicine, vol. 61, pp. 437-455,
segments. Furthermore, the UB is placed above LB Feb. 2010.
around î17 , î20 , and î22 . That means that there is a [3] A. J. Iafrate, L. Feuk, M. N. Rivera, M. L. Lis-
probability that these breakpoints do not exist. On the tewnik, P. K. Donahoe, Y. Qi, S. W. Scherer, and C.
other hand, estimates in the part covering î24 –î26 are Lee, “Detection of large-scale variation in the hu-
not reliable. Thus there is a probability of no changes man genome,” Nature Genetics, vol. 36, no. 9, pp.
in this region as well. 949-951, Sep. 2004.
[4] J. R. Pollack, T. Sorlie, C. M. Perou, C. A. Rees, S.
S. Jeffrey, P. E. Lonning, R. Tibshirani, D. Botstein,
5 Conclusions A. L. Borresen-Dale, and P. O. Brown, “Microarray
analysis reveals a major direct role of DNA copy
Effect of measurement noise on the HR-CGH array- number alteration in the transcriptional program of
based estimates of the CNVs naturally results in seg- human breast tumors,” Proc. Natl Acad. Sci (PNAS),
mental errors and jitter in the breakpoints due to typ- vol. 99, no. 20, pp. 12963-12968, Oct. 2002.
ically low SNRs. Errors can be so large that medical [5] R. Pique-Regi, A. Ortega, A. Tewfik, and S. As-
expert would hardly be able to arrive at correct con- gharzadeh, “Detection changes in the DNA copy
number,” IEEE Signal Process. Mgn., vol. 29, no.
clusions about real CNVs structures irrespective of the
1, pp. 98-107, Jan. 2012.
estimator used. Two rules of thumb for designers of
measurement equipment are thus the following: the [6] A. F. Cockburn, M. J. Newkirk, and R. A. Fir-
tel, “Organization of the ribosomal RNA genes of
higher probe resolution the more segmental accuracy
dictyostelium discoideum: mapping of the non-
and the larger segmental SNRs the lower jitter in the
trascribed spacer regions,” Cell, vol. 9, no. 4, pp.
breakpoints.
605-613, Dec 1976.
Figure 4: Measurements and estimates of the 1st chromosome changes taken from archive “159A–vs–159D–cut” available
in [44].
[12] E. Ben-Yaacov and Y.C. Eldar, “A fast and flexible [25] A. K. Alqallaf and A. H. Teqfik, “DNA copy num-
method for the segmentation of aCGH data”, Bio- ber detection and Sigma filter,” In Proc. GENSIPS,
statistics, vol. 24, no. 16, pp. i139–i145, 2008. pp. 1–4, 2007.
[13] V. Katkovnik and V.G. Spokoiny, “Spatial adaptive [26] S. Ivakhno, T. Royce, A. J. Cox, D. J. Evers, R. K.
estimation via fitted local likelihood techniques,” Cheetham, and S. Tavare, “CNAseg–A novel frame-
IEEE Trans. Signal Process., vol. 56, no. 3, pp. work for identification of copy number changes in
873–886, 2008. cancer from second-generation sequencing data,”
[14] A. Goldenshluger and A. Nemirovski, “Adaptive Bioinform., vol. 26, no. 24, pp. 3051–3058, Dec.
de-noising of signals satisfying differential inequal- 2010.
ities,” IEEE Trans. Inf. Theory, vol. 43, no. 3, pp.
[27] J. Chen and A. K. Gupta, Parametric Statistical
872–889, 1997.
Change Point Analysis with Applications to Ge-
[15] I.E. Frank and J.H. Friedman, “A Statistical View netics, Medicine, and Finance, 2nd Ed., Springer,
of Some Chemometrics Regression Tools,” Techno- 2012.
metrics, vol. 35, pp. 109-148, 1993.
[16] A.E. Hoerl and R.W. Kennard, R.W., “Ridge re- [28] R. Lucito, J. Healy, J. Alexander, A. Reiner, D.
gression: biased estimation for nonorthogonal prob- Esposito, M. Chi, L. Rodgers, A. Brady, J. Sebat,
lems,” Technometrics, vol. 12, pp. 55-67, 1970. J. Troge, J. A. West, S. Rostan, K. C. Nquyen,
S. Powers, K. Q. Ye, A. Olshen, E. Venkatra-
[17] R. Tibshirani, “Regression Shrinkage and Selection
man, L. Norton, and M. Wigler, “Representa-
via the Lasso,” Journal of Royal Statist. Soc., ser. B,
tional oligonucleotide microarray analysis: a high-
vol. 58, pp. 267-288, 1996.
resolution method to detect genome copy number
[18] J. Fridlyand, A.M. Snijders, D. Pinkel, D.G. Albert-
variation,” Genome Research, vol. 13, no. 10, pp.
son, and A.N. Jain, “Hidden Markov models ap-
2291–2305, Oct. 2003.
proach to the analysis of array CGH data”, Journal
of Multivariate Analysis, vol. 90, no. 1, pp. 132-153, [29] F. Picard, S. Robin, M. Lavielle, C. Vaisse, and J.-J.
2004. Daudin, “A statistical approach for array CGH data
[19] J. Chen and Y.-P. Wang, “A statistical change point analysis,” BMC Bioinformatics, vol. 6, no. 1, pp.
model approach for the detection of DNA copy 27–37, 2005.
number variations in array CGH data”, IEEE/ACM [30] A. B. Olshen, E. S. Venkatraman, R. Lucito, M.
Trans. on Comput. Biology and Bioinform., vol. 6, Wigner, “Circular binary segmentation for the anal-
no. 4, pp. 529–541, 2009. ysis of array-based DNA copy number data,” Bio-
[20] S. H. Chung and R. A. Kennedy, “Forward- statistics, vol. 5, no. 4, pp. 557–572, Oct. 2004.
backward non-linear filtering technique for extract- [31] J. T. Simpson, R. E. Mclntyre, D. J. Adams, and R.
ing small biological signal from noise,” J Neuro- Durbin, “Copy number variant detection in inbred
science Meth, vol. 40, no. 1, pp. 71–86, Nov 1991. strains from short read sequence data,” Bioinformat-
[21] O. Vite-Chavez, R. Olivera-Reyna, O. Ibarra- ics, vol. 26, no. 4, pp. 565–567, Feb. 2010.
Manzano, Y. S. Shmaliy, and L. Morales-Mendoza,
[32] L. Wang, A. Abyzov, J. O. Korbel, M. Snyder,
“Time-variant forward-backward FIR denoising of
and M. Gerstein, “MSB: A mean-shift-based ap-
piecewise-smooth signals,” Int. J. Electron. Com-
proach for the analysis of structural variation in the
mun. (AEU), vol. 67, no. 5, pp. 406–413, May 2013.
genome,” Genomic Res., vol. 19, no. 1, pp. 106–
[22] J. Muñoz-Minjares, O. Ibarra-Manzano, and Y. S.
117, Jan 2009.
Shmaliy, “Maximum likelihood estimation of DNA
copy number variations in HR-CGH arrays data,” [33] V. Boeva, A. Zinovyev, K. Bleakley, J. P. Vert,
In Proc. 12th WSEAS Int. Conf. on Signal Process., I. Janoueix-Lerosey, O. Delattre, and E. Barillot,
Comput. Geometry and Artif. Vision (ISCGAV’12), “Control-free calling of copy number alterations in
Proc. 12th WSEAS Int. Conf. on Systems Theory deep-sequencing data using GC-content normaliza-
and Sc. Comput. (ISTASC’12), Istanbul (Turkey), tion,” Bioinformatics, vol. 27, no. 2, pp. 268–269,
pp. 45-50, 2012. Jan. 2011.
[23] J. Huang, W. Wei, J. Zhang, G. Liu, G. R. Bignell, [34] R. Tibshirani and P. Wang, “Spatial smoothing and
M. R. Stratton, P. A. Futreal, R. Wooster, K. W. hot spot detection for CGH data using the fused
Jones, and M. H. Shapero, “Whole genome DNA lasso,” Biostatistics, vol. 9, no. 1, pp. 18–29, Jan.
copy number changes identified by high density 2008.
oligonucleotide arrays,” Hum. Genomics, vol. 1, no. [35] R. Pique-Regi, J. Monso-Varona, A. Ortega, R. C.
4, pp. 287–299, May 2004. Seeger, T. J. Triche, and S. Asgharzadeh, “Sparse
[24] C. Xie and M. T. Tammi, “CNV-seq, a new method representation and Bayesian detection of genome
to detect copy number variantion using high- copy number alterations from microarray data,”
throughput sequencing,” BMC Bioinform., vol. 10, Bioinformatics, vol. 24, no. 3, pp. 309–318, Mar.
no. 80, pp. 1–9, Mar 2009. 2008.
Abstract—This paper exposes features a fuzzy inference system The problem suggests a solution that enables you to relate the
General focused on educational evaluation, constructed from a data
specific content with daily practice and perception of personal
base relational, with the objectives of, first, generate discussion about
fuzzy logic as a new tool to interpret the educational process through realization that knowledge gives to each person. Why is
information technologies and communication, and second, exposing proposed an evaluation system with fuzzy logic, that not only
the virtues and drawbacks of these data structures in the development focuses on themes, concepts and theories, they represent the
of models of complex phenomena. First is the introduction with the connection of multiple variables within the educational
theoretical bases is exposed, Later, the technical characteristics of the process, allowing study it, understand it to improve it
application with the successes and difficulties rose by this, and finally permanently, from the reading of each student.
presents the conclusions of the authors.
A. Education: A product and a right at
Keywords—Fuzzy Logic, Fuzzy Inference System, decision
making with Multicriteria, alternative assessment in education. the same time
This project is mainly geared towards the educational
I. Introduction evaluation for distance mode of studies in higher education,
because in this case the tutor, for sharing with the student less
Education is a system based on the particular properties of its than in face-to-face mode, requires greater predictive and
elements and non-linear and interdependent relationships that assertive, to update and improve permanently their contents
they have with non-reversible dynamics in time and that and methodologies, projecting to a more personal realization
remain without any outside control, presenting patterns of and knowledge to a higher yield in the working life of the
order at various scales [1, 2]. That is why education is a students.
complex phenomenon, where training and learning process
The document “Opens and distance learning. Considerations
cannot be weighted from a scale standardized for all contexts
on trends, policies and strategies" of the UNESCO [5], it
and experiences, since the only knowledge can be appropriate,
exposes education distance and open as a product provided by
understood and applied in all their dimensions, when each
suppliers, but at the time defines the mode of studies as a
person internalizing it from in their everyday practice.
possibility that all persons have access to the right to
Current education and its ways of assessment is based in the education, the academic to the student the broadest spectrum
classical world modern vision as machine, which of possibilities through knowledge purposes providing
dismembered the reality in its parts and reset it from a logical education to isolated populations or with difficulties to access
order, on the one hand, this has caused that the educational to traditional education methods, and offer academic programs
evaluation is developed under a same logical scheme oriented to the training for work [5]. The educational spectrum
regardless of the person who studies [3]. in Colombian society is characterized by difficulties of access
And on the other hand It has led to underestimate, and even at to higher education, economic dependence, segregation, and in
times ignored, the particularity of student and its context, and some cases violence (social, economic and political) [6], and
the possibilities it offers to exercise and develop studied that requires evaluation methods not only considered the
knowledge, putting the topics above its actual application, interpretation of a specific subject, but the way in which this
where the students are overwhelmed and concealed by subject can contribute to the personal, social and productive
concepts, diagrams and formulas that do not relate to its development of the person, and not just one of these aspects.
existence [4].
B. Fuzzy logic
Fuzzy logic is a knowledge originated from mathematical
statistics, giving an element a weighting of non-deterministic
M.A. Luis Gabriel Moreno Sandoval as the one used in the classical logic, because in this one, an
University “Manuela Beltran” element belongs or not to a set, but in the fuzzy logic that
Bogotá, Colombia
gabrielmoreno10@gmail.com
element has different ranges of belonging. Loty Admeh Sadeh
(1972), creator of fuzzy logic, for example, proposes the case
of a classical group called "Tall people", to which belong the
William David Peña Peña people with stature greater than or equal to 1.8 meters, where a
UniversityCorporation UNIMINUTO person with stature of 1,79 m not belonged to this, however, in
Cundinamarca, Colombia
wpenapena1@hotmail.com
fuzzy logic, by means of mathematical functions, pre-
Table Quest_App, short for Applied Questionnaire, stores the issues or more likely to relate their personal fulfillment with
responses of students in a standard from its relationship with knowledge (A. Figure 4.). And the membership function of the
the User Id (User Table), Id Questionnaire (Id_Quest), Id Type consequence (SAA) is trapezoidal, because it represents a
Variable (Id_TVar) structure, Id Variable (Id_Var), Id transition (in belonged = 1), a momentary stage of the
valuation variable (Id_Val_Var) and Rule Id (Id_Rule) from relationship between knowledge and perception of
the input data (Table Ruleset fields) (Figure 3). Table Ruleset embodiment through this (B. Figure 4.).
contains all relationships between the antecedent variables and TABLE I. RULES OF IMPLICATION FOR “SAA”.
consequences, it inherits all the attributes of the Rules Table,
TP
and also records the group to which each rule belongs, used in
the processes of involvement and aggregation explained later. 1 2 3 4 5
These relationships are called rules of Implication, and in this P
case are of the form IF A is a' and IF B is b' THEN C is c' 1 1 2 2 3 3
P
(Mamdani type) which are the most used, unlike the rules type F 2 1 2 3 3 4
Takagi-Sugeno: IF a is a' and IF B is b' THEN C is f(x) [12].
3 1 2 3 4 4
4 1 2 3 4 5
5 2 3 4 5 5
Note: Rules temporary transition, subject to review by experts.
D. Defuzzification
Finally, it is necessary to apply a mathematical procedure to
generate a non-diffuse value from the ordered pairs. The most
common and used in this application method is the centroid,
consistent in finding an average of the area of the figure (B.
Figure 5), through the operation [13]:
Figure 5. Fuzzification (A) and implication (B)Schemein the FIS by querying. y= (∑i yiμB (y)dy)/ (∑iμB (y)dy)(2)
B. Implication III. Conclusions
Held a consultation to assess the results of the Fuzzification 1) SQL language has the limitation not to be parameterizable
from a logical condition called involvement. In this case using dynamically according to the requirements of the user, in the
the involvement of Mamdani [13] (1). Fuzzification, the involvement, the aggregation or the
μRc (x, y)= min [μA(x), μB(y)](1) Defuzzification, each of the evaluated function segments or
each of the mathematical procedures should be performed as a
query apart, with which generated great expense of memory
on the machine. However, this application proposes as a [11] Date C.J. An Introduction to Database Systems. Pearson Education.
Mexico , 2001.
solution the moving of data from these complex queries
[12] F. Dernoncourt. Introduction to fuzzy logic. Technical and theoretical
joining new tables, releasing the burden of memory of the report.MIT, January 2013.
above procedures. [13] MSc. Supo H. R. Fuzzy Logic. Technical and theoretical report.Jorge
2) Referential integrity and standardization remain the basis of Basadre Grohmann National University. Tacna, Perú, 2003.
coherence, strength and operation of data bases. Hence the About Author (s):
need to keep in mind these conditions regardless of the type of • M.A. Luis Gabriel Moreno Sandoval: Systems Engineer, Master of
information structure, since the combination of paradigms in a Science in Information and Communication are the University
District "Francisco José de Caldas" (Bogotá, Colombia). Manager
single computer system, allows to take elements suitable of and participant in many academic and business initiatives in
each, allowing a deeper conceptual, with more ownership and information science and communication with agent-based
social knowledge, such as which intends to consider this applications, natural language processing, and other cutting-edge
technologies models.
alternative (a pioneer in Colombia) educational evaluation
• William David Peña Peña: Philosopher, researcher in Computer
system, from the own perception of the student (in remote Science. Participant important international events on human and
mode). That is why that as work to a near future arises the Sciences to technology (Argentina, Uruguay and Cuba). Currently
union's database with a JAVA program that supplements the teaching at the university corporation UNIMINUTO in the area of
procedures with enough memory load, and even provides a research and developer of multiple p'royectos for academic lasd
application of ICT.
GUI based, to make this program a full framework for Fuzzy
Systems.
3) The relational databases, have gradually tended to be
replaced by object-oriented database, however this application,
checks that the effectiveness of these technologies lies in the
ability of abstraction on a problem, coming to represent
complex phenomena such as those described by means of
fuzzy logic. Taking the most valuable elements of each
technology, as for example, some SQL queries of few lines of
instruction that can replace many lines of code in advanced
programming languages. Sometimes technology makes us see
both the future that we seem to wonder traveling on a plane,
forgetting the importance of having learned to walk.
References
(1)
Abstract—The aim of this study is to determine the movement
(2)
equation of the drivers spine and to simulate the spine’s movements
in the coronal plane by using the MathCAD and CATIA software. On (3)
this line was proposed a methodology to approach the interaction
between driver and the vehicle to allow accurate conclusions for the
driving activity.
I. INTRODUCTION
(7)
In this equation the unknowns are: ai, bi, ci, and di. Knowing
the coordinates of four points, the unknowns ai, bi, ci, and di
are determined with the Cramer method:
(8)
Due to the fact that the origin is one of the four points
results that di is equal to 0. So the system of equations (8) is
transformed into a system of three equations with three
unknowns:
(9)
Where , .
(10)
V. CONCLUSION
The amplitude values of the sinusoidal functions describing
the variation in time of angles between vertebrae gives an
insight into the degree of deformation of intervertebral discs.
According to the literature the maximum tilt in the coronal
plane of the lumbar vertebrae are 5 ° for L1-L2; 5 ° for L2-L3;
4.5 ° for L3-L4; 2.2 ° for L4-L5; 1 degrees for L1-S1.
REFERENCES
[1] Adams M. ; Bogduk N. ; Burton K. ; Dolan P. (2006). The
Biomechanics of Back Pain Second Edition, Churchill Livingstone
Elsevier.
[2] Borozan I. S.; Maniu I.; Kulcsar R. M.; „Ergonomic analysis on driving
an Automatic Gearbox equipped vehicle”, SACI 2012 IEEE 7th
International Symposium on Applied Computational Intelligence and
Informatics, May 24-26, 2012, Timisoara, Romania.
[3] Borozan I. S.; Kulcsar R. M.; “Vertebral column bioengineering
analysis at bending and torsion”, International Conference on Human-
Machine Systems, Cyborgs and Enhancing Devices HUMASCEND,
Iasi, Romania, June 14-17, 2012.
[4] Goran Devedžić, Saša Ćuković, Vanja Luković, Danijela Milošević, K.
Subburaj, Tanja Luković, "ScolioMedIS: web-oriented information
system for idiopathic scoliosis visualization and monitoring", Journal of
Computer Methods and Programs in Biomedicine, Vol.108, No.-, pp.
736-749, ISSN -, Doi 10.1016/j.cmpb.2012.04.008, 2012.
[5] Hilohi C., Untaru M., Soare I., Druţa Gh., „Metode si mijloace de
incercare a automobilelor”, Editura Tehnică București, 1982.
[6] Hinza B., Seidel H., „The significance of using anthropometric
parameters and postures of European drivers as a database for finite-
element models when calculating spinal forces during whole-body
vibration exposure”, International Journal of Industrial Ergonomics,
Elsevier, 2008.
[7] Kolich M., „A conceptual framework proposed to formalize the
scientific investigation of automobile seat comfort”, Applied
Ergonomics, Elsevier, 2008.
[8] Kulcsar R. M.; Madaras L.; “Ergonomical study regarding the effects of
the inertia and centrifugal forces on the driver”, MTM & Robotics 2012,
The Joint International Conference of the XI International Conference
on Mechanisms and Mechanical Transmissions (MTM) and the
International Conference on Robotics (Robotics’12), Clermont-Ferrand,
France, June 6-8, 2012 Applied Mechanics and Materials, Vol. 162,
Mechanisms, Mechanical Transmissions and Robotics, ISBN-13:978-3-
03785-395-5, pp. 84-91.
[9] Muksian R., Nash C.D.Jr. “A model for the response of seated humans
to sinusoidal displacements of the seat”, J. Biomechanics, vol.7, pp
209-215, Pergamon Press, 1974.
[10] Tae-Yun Koo1, Kee-Jun Park, „A Study on Driver’s Workload of
Telematics Using a Driving Simulator: A Comparison among
Information Modalities”, International journal of precision engineering
and manufacturing vol. 10, no. 3, pp. 59-63, 2009.
Authors Index
Adochiei, F.-C. 153 Grigorie, T. L. 153 Peña, W. D. P. 222
Argesanu, V. 172, 227 Hai, X. 148 Perminov, V. 55
Atangana, A. 202 He, J. 148 Phinikettos, I. 26
Bacalu, I. 68 He, Y. 148 Pop, I. 98
Bai, J. 148 Hitzer, E. 19 Pota, M. 107
Balas, M. J. 48 Husain, I. 79 Poulimenou, S. 120
Barta, J. 86 Isakov, A. A. 138 Poulos, M. 120
Bidabad, Be. 198 Jirina, M. 179 Puskás, A. 132
Bidabad, Bi. 198 Jula, M. 172, 227 Remizov, M. 43
Borozan, I. S. 172, 227 Kovarik, M. 142 Rotaru, C. 153
Bota, E. 172 Kucerova, A. 163 Rozehnalova, P. 163
Cabal-Aragon, J. 212 Kulcsar, R. M. 172, 227 Sandoval, M. A. L. G. M. 126, 222
Chira, A. 132 Labropulu, F. 98 Schwark, J. 79
Choudhary, J. P. 187 Langdon, C. R. 79 Senichenkov, Y. B. 138
Cravo, G. 103 Li, D. 98 Şerbănescu, C. 68
Ćuković, S. 172 Li, S. 148 Shmaliy, Y. S. 212
De, M. 187 Liagkouras, K. 158 Singh, D. 187
Demetriou, I. 26 Ligere, E. 61 Smith, D. H. 86
Dombi, J. 91 Maniu, I. 227 Sokolov, N. L. 115
Dzenite, I. 61 Mastorakis, N. 198 Stamou, S. 120
Edu, I.-R. 153 Metaxiotis, K. 158 Sumbatyan, M. 43
Esposito, M. 107 Montemanni, R. 86 Torres, E. 167
Fernandez, M. 33 Munoz-Minjares, J. 212 Vala, J. 163
Frost, S. A. 48 Nagel, A. 227 Villareal, E. 167
Gandy, A. 26 Obreja, R. 153 Wang, Y. 148
Gnitchogna, R. B. 202 Papavlasopoulos, S. 120