A Polynomial Model for the Level Variations of the Caspian

advertisement
Genetic Design of GMDH-type Neural Networks for Modelling of
Thermodynamically Pareto Optimized Turbojet Engines
a
a
K. Atashkari, aN. Nariman-zadeh , aA. Darvizeh, bX. Yao, aA. Jamali, aA. Pilechi
Department of Mechanical Engineering,Engineering Faculty, The University of Guilan
P.O. Box 3756, Rasht, IRAN
b
School of Computer Science, The University of Birmingham
Edgbaston, Birmingham B15 2TT, U.K.
Abstract:- Firstly, a multi-objective genetic algorithm (GAs) is used for Pareto based optimization of
thermodynamic cycle of ideal turbojet engines considering four important conflicting thermodynamic objectives,
namely, specific thrust (ST), specific fuel consumption (SFC), propulsive efficiency (p), and thermal efficiency
(t). This provides the best Pareto front of such four-objective optimization from the space of design variables,
which are Mach number and pressure ratio, to the space of the above-mentioned four thermo-mechanical
objective functions. Secondly, genetic algorithms (GA) with a new encoding scheme are used for optimal design
of both connectivity configuration and the values of coefficients, respectively, involved in GMDH-type neural
networks for the inverse modelling of the input-output data table obtained as the best Pareto front.
Key-Words:- Turbojet Engines, Pareto, Group Method of Data Handling (GMDH), Genetic Algorithms (GAs).
1
Introduction
In the optimization of complex real-world problems,
there are several objective functions or cost functions
(a vector of objectives) to be optimized (minimized
or maximized) simultaneously. These objectives
often conflict each other so that improving one of
them will deteriorate another objective function.
Therefore, there is no single optimal solution as the
best with the respect to all the objective functions.
Instead, there is a set of optimal solutions, well
known as Pareto optimal solutions or Pareto front
[1], which distinguishes significantly the inherent
natures between single-objective and multi-objective
optimization problems. The concept of Pareto front
or set of optimal solutions in the space of objective
functions in multi-objective optimization problems
(MOPs) stands for a set of solutions that are nondominated to each other but are superior to the rest
of solutions in search space. It must be noted that
this non-dominancy does exist in different levels,
forming different ranked Pareto fronts, although the
first Pareto front is the most important and will be
the ultimate solution. The inherent parallelism in
evolutionary algorithms makes them suitably eligible
for solving MOPs. Among these methods, the Vector
Evaluated Genetic Algorithm (VEGA) proposed by
Schaffer [2], Fonseca and Fleming’s Genetic
Algorithm (FFGA) [2], Non-dominated Sorting
Genetic Algorithm (NSGA) by Srinivas and Deb [1],
and Strength Pareto Evolutionary Algorithm (SPEA)
by Zitzler and Thiele [2], and the Pareto archived
evolution strategy (PEAS) by Knowles and Corne
[2] are the most important ones. A very good and
comprehensive survey of these methods has been
presented in [2]. Basically, both NSGA and FFGA as
Pareto-based approaches use the revolutionary nondominated sorting procedure originally proposed by
Goldberg [3]. Besides, the diversity issue and the
lack of elitism was also a motivation for
modification of that algorithm to NSGA-II [4] in
which a direct elitist mechanism, instead of sharing
mechanism, has been introduced to enhance the
population diversity. This modified algorithm has
been known as the state-of-the-art in evolutionary
MOPs.
System identification techniques are applied in many
fields in order to model and predict the behaviors of
unknown and/or very complex systems based on
given input-output data [5]. The Group Method of
Data Handling (GMDH) algorithm is a selforganizing approach by which gradually complicated
models are generated based on the evaluation of their
performances on a set of multi-input-single-output
data pairs (i=1, 2, …, M). The GMDH was firstly
developed by Ivakhnenko [6] as a multivariate
analysis method for complex systems modelling and
identification. The main idea of GMDH is to build an
analytical function in a feedforward network based
on a quadratic node transfer function [7] whose
coefficients are obtained using regression technique.
In fact, real GMDH algorithm in which model
coefficients are estimated by means of the least
squares method has been classified into complete
induction and incomplete induction, which represent
the combinatorial (COMBI) and multilayered
iterative algorithms (MIA), respectively [7]. In
recent years, the use of such self-organizing network
leads to successful application of the GMDH-type
algorithm in a broad range area in engineering,
science, and economics [7, 8].
DEFINITION OF PARETO DOMINANCE
A vector U  u1 , u2 ,..., uk   k is dominance to
vector V  v1 , v2 ,..., vk   k (denoted by U  V )
i 1,2,..., k , ui  vi 
j  1,2,..., k : u j  v j . In other words, there is at
if
In this paper, firstly, an optimal set of design
variables in turbojet engines, namely, the input Mach
number Mo and the pressure ratio of compressor c
are found using Pareto approach to multi-objective
optimization. In this study, four conflicting
objectives in ideal subsonic turbojet engines are
selected for optimization. These include thermal
efficiency (t) and propulsive efficiency (p)
together with specific fuel consumption (SFC) and
specific thrust (ST). Secondly, GAs are deployed in a
new approach to design the whole architecture of the
GMDH-type neural networks, i.e., the number of
neurons in each hidden layer and their connectivity
configuration, for modelling of the four-input-twooutput data table obtained in the first part of the
study. The connectivity configuration is not limited
to the adjacent layers, as have already proposed in
[8], so that general structure GMDH-type neural
networks, in which neurons in any layers can be
connected to each other, are evolved.
Multi-objective optimization which is also called
multicriteria optimization or vector optimization has
been defined as finding a vector of decision variables
satisfying constraints to give acceptable values to all
objective functions [2, 4]. In general, it can be
mathematically defined as:

only
if
least one u j which is smaller than v j whilst the rest
u ’s are either smaller or equal to corresponding v ’s.
DEFINITION OF PARETO OPTIMALITY
A point X *   (  is a feasible region in  n
satisfying equations (2) and (3)) is said to be Pareto
optimal (minimal) with respect to the all X   if
and only if F ( X * )  F ( X ) . Alternatively, it can be
readily restated as
i 1,2,..., k , X    { X *} f i ( X * )  f i ( X )
 j  1,2,..., k :
f j ( X * )  f j ( X ) . In other
words, the solution X * is said to be Pareto optimal
(minimal) if no other solution can be found to
dominate X * using the definition of Pareto
dominance.
DEFINITION OF PARETO SET
2 Multi-objective Optimization
*
find the vector X *  x1* , x*
2 ,..., xn
and
 to optimize
T
For a given MOP, a Pareto set Ƥ ‫٭‬is a set in the
decision variable space consisting of all the Pareto
optimal
vectors


Ƥ‫ ٭‬ { X   | ∄ X   : F ( X )  F ( X )} . In other
words, there is no other X  as a vector of decision
variables in  that dominates any X ∈Ƥ‫٭‬.
F ( X )  f1( X ), f 2 ( X ),..., f k ( X ) T ,
subject to m inequality constraints
(1)
DEFINITION OF PARETO FRONT
For a given MOP, the Pareto front ƤŦ fo tes a si ‫٭‬
vector of objective functions which are obtained
using the vectors of decision variables in the Pareto
gi ( X )  0 ,
(2)
set
Ƥ si
taht
,‫٭‬
ƤŦ‫ ٭‬ {F ( X )  ( f1 ( X ), f 2 ( X ),...., f k ( X )) : X  Ƥ‫٭‬

i  1 to m ,
and p equality constraints
h j ( X )  0 , j  1 to p ,

}. In other words, the Pareto front ƤŦ eht fo tes a si ‫٭‬
(3)
where X   is the vector of decision or design
variables, and F ( X )  k is the vector of objective
functions which each of them be either minimized or
maximized. However, without loss of generality, it is
assumed that all objective functions are to be
minimized. Such multi-objective minimization based
on Pareto approach can be conducted using some
definitions:
*
vectors of objective functions mapped from Ƥ.‫٭‬
n
3
Modelling Using GMDH-Type
Neural Networks
By means of GMDH algorithm a model can be
represented as set of neurons in which different pairs
of them in each layer are connected through a
quadratic polynomial and thus produce new neurons
in the next layer. Such representation can be used in
modelling to map inputs to outputs. The formal
definition of the identification problem is to find a
function fˆ so that can be approximately used instead
of actual one, f , in order to predict output ŷ for a
it is now possible to train a GMDH-type neural
network to predict the output values ŷ for any given
In the basic form of the GMDH algorithm, all the
possibilities of two independent variables out of total
n input variables are taken in order to construct the
regression polynomial in the form of equation (8)
that best fits the dependent observations ( y i , i=1, 2,
…, M) in a least-squares sense. Consequently,
 n  n(n  1)
  
neurons will be built up in the first
2
 2
hidden layer of the feedforward network from the
observations { ( y i , x , x ); (i=1, 2, …, M)} for
input vector X  ( x , x , x ,..., x ) , that is
different p, q  {1,2,..., n} [7]. In other words, it is
given input vector X  ( x , x , x ,..., x ) as close as
1
2
3
n
possible to its actual output y . Therefore, given M
observation of multi-input-single-output data pairs so
that
(i=1,2,…,M), (4)
y  f ( x , x , x ,..., x )
i
i1 i 2
i3
in
i
i1
i2
yˆ  fˆ ( x , x , x ,..., x )
i
i1 i 2 i 3
in
i3
in
(i=1,2,…,M).
(5)
The problem is now to determine a GMDH-type
neural network so that the square of difference
between the actual output and the predicted one is
minimised, that is
M
2
 [ fˆ ( x , x , x ,..., x )  y ]  min
i1 i 2 i 3
in
i
i 1
.
(6)
General connection between inputs and output
variables can be expressed by a complicated discrete
form of the Volterra functional series in the form of
n
n n
n n n
y  a 0   ai xi    aij xi x j     aijk xi x j x k ...
i 1
i 1 j 1
i 1 j 1 k 1
(7)
which is known as the Kolmogorov-Gabor
polynomial [7]. This full form of mathematical
description can be represented by a system of partial
quadratic polynomials consisting of only two
variables (neurons) in the form of
yˆ  G ( xi , x j )  a 0  a1 xi  a 2 x j  a3 xi x j  a 4 xi
2
 a5 x j
2
.
(8)
In this way, such partial quadratic description is
recursively used in a network of connected neurons
to build the general mathematical relation of inputs
and output variables given in equation (7). The
coefficient a i in equation (8) are calculated using
regression techniques [7] so that the difference
between actual output, y, and the calculated one, ŷ ,
for each pair of x i , x j as input variables is
minimized. Indeed, it can be seen that a tree of
polynomials is constructed using the quadratic form
given in equation (8) whose coefficients are obtained
in a least-squares sense. In this way, the coefficients
of each quadratic function G are obtained to
i
optimally fit the output in the whole set of inputoutput data pair, that is
M
2
 ( y  G ())
i
i
E  i 1
 min
M
.
(9)
ip
iq
now possible to construct
M
data triples
{ ( yi , xip , xiq ); (i=1, 2, …, M)} from observation
using such p, q  {1,2,..., n} in the form
 x1 p

 x2 p
 x Mp

x1q
x 2q
x Mq
y1 

y2  .
y M 
Using the quadratic sub-expression in the form of
equation (8) for each row of M data triples, the
following matrix equation can be readily obtained as
(10)
Aa  Y
where a is the vector of unknown coefficients of
the quadratic polynomial in equation (8)
a  {a 0 , a1 , a 2 , a 3 , a 4 , a 5 }
(11)
and
(12)
Y  { y1 , y 2 , y 3 ,..., y M }T
is the vector of output’s value from observation. It
can be readily seen that
1

A  1
1

x1 p
x2 p
x Mp
x1q
x 2q
x Mq
x1 p x1q
x 2 p x 2q
x Mp x Mq
x12p
x 22 p
2
x Mp
x12q 

x 22q  .
2 
x Mq

(13)
The least-squares technique from multiple-regression
analysis leads to the solution of the normal equations
in the form of
(14)
a  ( AT A) 1 AT Y
which determines the vector of the best coefficients
of the quadratic equation (8) for the whole set of M
data triples. It should be noted that this procedure is
repeated for each neuron of the next hidden layer
according to the connectivity topology of the
network.
4
Application
of
Genetic
Algorithm in the Topology Design of
GMDH-type Neural Networks
Evolutionary methods such as genetic algorithms
have been widely used in different aspects of design
in neural networks because of their unique
capabilities of finding a global optimum in highly
multi-modal and/or non-differentiable search space
[8-9]. Such stochastic methods are commonly used
in the training of neural networks in terms of
associated weights or coefficients which have
successfully performed better than traditional
gradient-based techniques [9]. The literature shows
that a wide range of evolutionary design approaches
either for architectures or for connection weights
separately, in addition to efforts for them
simultaneously [9]. In the most GMDH-type neural
network, neurons in each layer are only connected to
neuron in its adjacent layer as it was the case in
Methods I and II previously reported in [8]. Taking
this advantage, it was possible to present a simple
encoding scheme for the genotype of each individual
in the population as already proposed by authors [8].
the second hidden layer and used with abbc in the
same layer to make the output neuron abbcadad as
shown in the figure (1). It should be noted that such
repetition occurs whenever a neuron passes some
adjacent hidden layers and connects to another
neuron in the next 2nd, or 3rd,or 4th,or … following
hidden layer. In this encoding scheme, the number of
repetition of that neuron depends on the number of
passed hidden layers, ñ, and is calculated as 2 ñ . It is
easy to realize that a chromosome such as abab bcbc,
unlike chromosome abab acbc for example, is not a
valid one in GS-GMDH networks and has to be
simply re-written as abbc .
a●
●ab
●abbc
b●
●bc
c●
However, in order to make it more general, it is
necessary to remove such restriction. In such
representation, neurons in different layers including
the input layer can be connected to others far away,
not only in the very adjacent layers. The encoding
schemes in generalized GMDH neural networks
must demonstrate the ability of representing different
length and size of such neural networks. Moreover,
the ability of changing building blocks of
information using crossover and mutation operators
must be preserved in such coding representation. In
the next sections, the encoding scheme of the newly
developed generalized GMDH neural networks is
discussed.
4.1 The Genome Representation of
Generalized GMDH Neural Networks
In the generalized GMDH neural networks, neurons
connections can occur between different layers
which are not necessarily very adjacent ones, unlike
the conventional structure GMDH neural networks in
which such connections only occur between adjacent
layers. For example, a network structure which
depicted in figure (1) shows such connection of
neuron ad directly to the output layer. Consequently,
this generalisation of network's structure can
evidently extend the performance of generalized
GMDH neural networks in modelling of real-world
complex processes. Such generalization is
accomplished by repeating the name of the neuron
which directly passing the next layers. In figure (1),
neuron ad in the first hidden layer is connected to the
output layer by directly going through the second
hidden layer. Therefore, it is now very easy to notice
that the name of output neuron ( network's output)
includes ad twice as abbcadad. In other words, a
virtual neuron named adad has been constructed in
●ad
●
abbcadad
d●
Figure 1: A Generalized GMDH network structure of a
chromosome
4.2 Genetic
Generalized
Reproduction
Operators
for
GMDH
Network
The genetic operators of crossover and mutation can
now be implemented to produce two offsprings from
two parents. The natural roulette wheel selection
method is used for choosing two parents producing
two offsprings. The crossover operator for two
selected individuals is simply accomplished by
exchanging the tails of two chromosomes from a
randomly chosen point as shown in figure (2). It
should be noted, however, such a point could only be
n 1
chosen randomly from the set { 21 , 2 2 , ..., 2 l }
where nl is the number of hidden layers of the
chromosome with the smaller length. It is very
evident from figures (2) and (3) that the crossover
operation can certainly exchange the building blocks
information of such generalized GMDH neural
networks so that the two type of generalized GMDH
and conventional GMDH-type neural networks can
be converted to
Parents
Offsprings
a b b c a d a d
a b b c
a b d d
a b d d a d a d
Figure 2: Crossover operation for two individuals in
generalized GMDH networks
each other, as can be seen from figure (3). In
addition, such crossover operation can also produce
different length of chromosomes which in turn leads
to different size of either generalized GMDH or
conventional GMDH network structures. Similarly,
the mutation operation can contribute effectively to
a
ab
b
bc
c
ad
ab
b
bc
abbc
abbcadad
c
a
ab
ab
abdd
b
b
c
abbc
d
d
a
a
c
abdd
abddadad
d
ad
d
Figure 3: Crossover operation on two generalized GMDH
networks
The incorporation of genetic algorithm into the
design of such GMDH-type neural networks starts
by representing each network as a string of
concatenated sub-strings of alphabetical digits. The
fitness, (  ), of each entire string of symbolic digits
which represents a GMDH-type neural network to
model the Pareto optimized data pairs is evaluated in
the form
Φ=1/E
,
(15)
where E, is the mean square of error given by
equation (9), is minimized through the evolutionary
process by maximizing the fitness  . The
evolutionary process starts by randomly generating
an initial population of symbolic strings each as a
candidate solution. Then, using the aforementioned
genetic operations of roulette wheel selection,
crossover, and mutation, the entire populations of
symbolic strings to improve gradually. In this way,
GMDH-type neural network models with
progressively increasing fitness, , are produced
until no further significant improvement is
achievable.
5 Multi-Objective Thermodynamic
Optimization and Modelling of
Turbojet Engines
Turbojet engines use air as the working fluid and
produce thrust based on the variation of kinetic
energy of burnt gases after combustion. The study of
thermodynamic cycle of a turbojet engine involves
different thermo-mechanical aspects such as
developed specific thrust, thermal and propulsive
efficiencies, and specific fuel consumption. A
detailed description of the thermodynamic analysis
and equations of ideal turbojet engines are given in
[10]. Input parameters involved in such
thermodynamic analysis in an ideal turbojet engine
given in Appendix A are Mach number (M0), input
air temperature (T0, K), specific heat ratio (  ),
heating value of fuel (hpr, kj/kg), exit burner total
temperature (Tt4, K), and pressure ratio, c. Output
parameters involved in the thermodynamic analysis
in the ideal turbojet engine are, specific thrust, (ST,
N/kg/sec), fuel-to-air ratio (f), specific fuel
consumption (SFC, kg/sec/N), thermal efficiency
(t), and propulsive efficiency (p). However, in
multi-objective optimization study, some input
parameters are already known or assumed as, T0 =
216.6 K,  =1.4, hpr =48000 kj/kg, and Tt4 = 1666 K.
The input Mach number 0 < M0 ≤ 1 and the
compressor pressure ratio 1 ≤ c ≤ 40 are considered
as design variables to be optimally found based on
multi-objective optimization of 4 output parameters,
namely, ST, SFC, ηt, and ηp. Consequently, four
objectives, namely, SFC, ST, p, and t, are chosen
for multi-objective optimization in which ST, p, and
t are maximized whilst SFC is minimized
simultaneously. A population size of 200 has been
chosen with crossover probability Pc and mutation
probability Pm as 0.8 and 0.02, respectively. The
result of such multi-objective optimization in the
plane of p and t is shown in figure (4). It should be
noted that such result could also be shown in
different plane; but such plane would be of the most
Training Data
Testing Data
0.45
0.4
Propulsive efficiency
the diversity of the population. This operation is
simply accomplished by changing one or more
symbolic digits as genes in a chromosome to another
possible symbols, for example, abbcadad to
abbccdad. It is very evident that mutation operation
can also convert a generalized GMDH network to a
conventional GMDH network or vice versa. It should
be noted that such evolutionary operations are
acceptable provided a valid chromosome is
produced. Otherwise, these operations are simply
repeated until a valid chromosome is constructed.
0.35
0.3
0.25
0.2
0.15
0.1
0.05
0
0.6
0.62
0.64
0.66
0.68
0.7
0.72
Thermal efficiency
Figure 4: Non-dominated Pareto points in the plane of
efficiencies
important one in terms of modelling. Consequently,
the input-output data pairs used in the modelling
involve two different data tables which have been
just optimally found using multi-objective GAs for
thermodynamic Pareto optimisation of ideal turbojet
engines. The first table consists of four variables as
inputs, namely, SFC, ST, p, and t and one output
which is the Mach number. The second table consists
of the same four variables as inputs and another
output which is pressure ratio. It can be readily seen
that the results of a Pareto-based four-objective
optimization process for two design variables, Mach
number and pressure ratio, are inversely deployed for
a polynomial-type neural network modelling so that
the values of such design variables can now be
mathematically represented for the corresponding
combination values of four parameters, SFC, ST, p,
and t. In order to demonstrate the generalization
property of the evolved polynomial neural network,
data samples have been divided randomly for both
training and testing purposes, as shown in figure (4).
The unforeseen data samples during training process
or the test data have been merely used for testing
purpose. The structure of the evolved 1-hidden layer
GMDH-type neural network obtained for both design
variables is shown in figure (5).
(a)
(b)
Figure 5: Evolved GMDH-type neural networks for
modelling: (a) Mach No.
(b) Pressure ratio
The very good behaviours of such GMDH-type
network models are also depicted in scattered figure
(6) for both Mach number and pressure ratio,
respectively. It can be clearly seen that both models
have very good generalization ability in terms of
predicting unforeseen test data during training
process.
1.2
Computed & Predicted Mach No.
Computed & predicted pressure ratio
45
40
35
30
25
20
15
10
5
0
1
0.8
0.6
0.4
0.2
0
0
5
10
15
20
25
Actual pressure ratio
30
35
40
45
0
0.2
0.4
0.6
0.8
1
1.2
Actual Mach No.
(a)
(b)
Figure 6: Modelling behaviours of the evolved GMDH-type
neural networks in both training and testing data:
(a) Pressure ratio (b) Mach No.
Therefore, GMDH-type neural models shown in
figure (5) simply represent the optimal design
variables from a multi-objective optimisation point
of view for SFC, ST, p, and t.
6
Conclusion
Evolutionary methods for designing generalized
GMDH-type networks have been proposed and
successfully used for the modelling of input-output
data table obtained as the best Pareto front of
thermodynamic cycle of ideal turbojet engines. Such
Pareto front has been already found by a multiobjective optimization GA for the turbojet engine.
Such combination of modelling of non-dominated
Pareto optimized data pairs is very promising in
discovering useful
and interesting design
relationships.
References:
[1]Srinivas, N. and Deb, K., “ Multiobjective
optimization Using Nondominated Sorting in
Genetic Algorithms”, Evolutionary Computation,
Vol. 2, No. 3, pp 221-248, 1994.
[2]Coello Coello, C.A., “A comprehensive survey of
evolutionary based multiobjective optimization
techniques”, Knowledge and Information Systems:
An Int. Journal, (3), pp 269-308, 1999.
[3]Goldberg, D. E., “Genetic Algorithms in Search,
Optimization, and Machine Learning”, AddisonWesley, 1989.
[4]Deb, K., Agrawal, S., Pratap, A., Meyarivan, T.,
“A fast and elitist multi-objective genetic algorithm:
NSGA-II”, IEEE Trans. On Evolutionary
Computation 6(2):182-197, 2002.
[5] Astrom, K.J. and Eykhoff, P., “System
identification, a survey”, Automatica 7-123-62,
(1971).
[6] Ivakhnenko, A.G., “ Polynomial Theory of
Complex Systems”, IEEE Trans. Syst. Man &
Cybern, SMC-1, 364-378, (1971).
[7] Farlow, S.J., ed., “ Self-organizing Method in
Modeling: GMDH type algorithm”, Marcel Dekker
Inc., (1984).
[8] Nariman-Zadeh, N., Darvizeh, A., Ahmad-Zadeh,
R., “Hybrid Genetic Design of GMDH-Type Neural
Networks Using Singular Value Decomposition for
Modelling and Prediction of the Explosive Cutting
Process”, Proceedings of the I MECH E Part B
Journal of Engineering Manufacture, Vol. 217, pp.
779 – 790, (2003).
[9]Yao, X., “Evolving Artificial Neural Networks”,
Proceedings of IEEE, 87(9):1423-1447, Sept.,
(1999).
[10] Mattingly, J.P., “Elements of Gas Turbine
Propulsion”, Mc-Graw Hill, 1996
Download