Chapter 06: Basic Structure of Fuzzy Neural Networks

advertisement
Chapter 6
Basic Structure of Fuzzy Neural Networks
In this chapter we shall discuss the structure of fuzzy neural networks. We start
with general definitions of multifactorial functions. And we show that a fuzzy neuron can be formulated by means of standard multifactorial function. We also give
definitions of a fuzzy neural network based on fuzzy relationship and fuzzy neurons.
Finally, we describe a learning algorithm for a fuzzy neural network based on V and
A operations.
6.1
Definition of Fuzzy Neurons
Neural networks alone have demonstrated their ability to classify, recall, and associate information [l].In this chapter, we shall incorporate fuzziness to the networks.
The objective to include the fuzziness is to extend the capability of the neural networks to handle “vague” information than “crisp” information only. Previous work
has shown that fuzzy neural networks have achieved some level of success both fundamentally and practically [l-lo]. As indicated in reference [l],there are several
ways to classify fuzzy neural networks: (1) a fuzzy neuron with crisp signals used
to evaluate fuzzy weights, (2) a fuzzy neuron with fuzzy signals which is combined
with fuzzy weights, and (3) a fuzzy neuron described by fuzzy logic equations.
In this chapter, we shall discuss a fuzzy neural network where both inputs and
outputs can be either a crisp value or a fuzzy set. To do this we shall first introduce
multifactorial function [ll, 121. We have pointed out from Chapter 4 that one of
the basic functions of neurons is that the input to a neuron is synthesized first, then
activated, where the basic operators to be used as synthesizing are “ ” and “ . ”
denoted by (
. ) and called synthetic operators. However, there are divers styles
of synthetic operators such as (V,A), (V,.), (+, A), etc. More general synthetic
operators will be multifactorial functions, so we now briefly introduce the concept
of multifactorial functions.
In [0, lIm,a natural partial ordering “5” is defined as follows:
+
+,
( V X , Y E [O,
11”) (.5 Y * (Zj 5 yj,
j = 1 , 2 , . . . ,rn))
A multifactorial function is actually a projective mapping from an rn-ary space to a
© 2001 by CRC Press LLC
one-ary space, denoted by M,. On many occasions, it is possible to transform the
spaces into closed unit intervals:
that we call standard multifactorial functions. For what follows, we will focus our
discussion on standard multifactorial functions.
The standard multifactorial functions can be classified into two groups:
Group 1. Additive Standard Multifactorial (ASM) functions
The functions in this group satisfy the condition: b'(xl,x2,. . * , xm) E [0,l]",
which means that the synthesized value should not be greater than the largest of
component states and should not be less than the smallest of the component states.
The following is its normal definition.
is called an m-ary Additive
Definition 1 A mapping Mm : [0,1]" + [0,1]
Standard Multifactorial function (ASMm-func) if it satisfies the following axioms:
m
(m.2)
m
A xj 5 M m ( X ) 5 j=1
V xj;
j=1
(m.3) Mm(xl,x2,.. . , z m ) is a continuous function of each variable xj.
n
The set of all ASMm-func is denoted by M , = {MmlMm is a m - dimensional
ASM, - func}. Clearly, when m = 1, an ASMm-func is an identity mapping from
[0,1] to [0,1]. Also (m.2) implies M m ( a ,. . . , a ) = a.
Example 1 The follwing mappings are examples of ASMm-funcs from [0, lImto
[O, 11:
A :x
m
I--+
A ( X ) :=
A xj,
j=1
v
x HV ( X ):= v xj,
m
:
j=1
cx
:
C(X):= cwjxj,
m
+-+
j=1
© 2001 by CRC Press LLC
(6.3)
m
where wj E [0,1], and
C wj = 1.
j=1
m
where wj E [0,1], and
V
w j = 1.
j=1
where wj E [0,1] and
Cj”=1wj = 1.
where wj E [0,1], and
v wj = 1.
j=1
m
(6.11)
(6.12)
m
where p
> 0, wj E [0,1] and C
wj =
1.
j=1
Group 2. Non-additive Standard Multifactorial (NASM) functions
This group of functions does not satisfy axiom (m.2). That is, the synthesized
value can exceed the boundaries of axiom (m.2), i.e.,
© 2001 by CRC Press LLC
For example, a department is led and managed by three people; each of them has
a strong leading ability. But for some reason, they cannot work smoothly among
themselves. Hence, the collective leading ability (a multifactorial score) falls below
the individual’s, i.e.,
M3(z1,22,x3)5
v
xj,
j=1
where xj is the leading ability of the individual i, i = 1 , 2 , 3 , and M3(z1,22,~3)
is
the multifactorial leading ability indictor for the group of three.
On the other hand, it is possible for the three management people to work together
exceedingly well. This implies that the combined leadership score can be higher than
any one of the three individual’s, i.e.,
It has the same meaning in the Chinese old saying: “Three cobblers with their wits
combined can exceed Chulceh Liang, the master minded”.
Definition 2 A mapping M, : [0, 11, e [0,1] is called an m-ary Non-Additive
Standard Multifactorial function, denoted by NASM,-func, if it satisfies axioms
(m.l), (m.3),and the following axiom:
The set of all NASM,-funcs
Example 2 The mapping
NASM,-func:
is denoted by M A .
n
:
[0, 11,
+ [0,1] defined
as the following is a
m
(51,.
.*
,x,)
H r]:(~l,.
. . ,x,)
:=
r]: xj.
(6.14)
j=1
Next, we shall use the these definitions to define fuzzy neurons.
Definition 3
A fuzzy neuron is regarded as a mapping F N :
where M , E M,, 8 E [0,1] and cp is a mapping or an activation function, cp : R
! +
[0,1] with cp(u)= 0 when u 5 0; and !R is the field of all real numbers. And a neural
network formed by fuzzy neurons is called a fuzzy neural network.
Figure 1 illustrates the working mechanism of a fuzzy neuron.
© 2001 by CRC Press LLC
Figure 1 Illustration of a fuzzy neuron
Example 3
The following mappings from [O,lIn to [0,1] are all fuzzy neurons:
(6.16)
where wi E [0,1] and
Cy=lwi = 1.
(6.17)
where wi E [0,1] and
Vy=1 W i =
1
(6.18)
where wi E [0,1] and
Vy=l wi = 1.
(6.19)
where wi E [0,1] and
CyZlwi = 1.
(6.20)
where wi E [0,1] and
Vy=l wi 5 Vy=l xi.
(6.21)
where wi E [0,1] and
ATZl wi 5
xi .
Example 4 In Equation (6.17), if we let (V i)(wi= l ) ,8 = 0 and cp = id,where
id is an identity function, i.e., (V x ) ( i d ( x )= x ) , we have a special fuzzy neuron:
n
Y= Vxi,
i=l
© 2001 by CRC Press LLC
(6.22)
In the same way, from (6.21) we have another special fuzzy neuron:
n
y = Axa.
(6.23)
i= 1
6.2
Fuzzy Neural Networks
In this section, we shall discuss fuzzy neural networks. We first use the concept
of fuzzy relationship followed by the definition of fuzzy neurons. We also discuss a
learning algorithm for a fuzzy neural network.
6.2.1
Neural Network Representation of Fuzzy Relation Equations
We consider a typical kind of fuzzy relation equation:
XoR=B
(6.24)
is the matrix of
where X = ( x l , x 2 ,* . . , x n ) is the input vector, R = ( r i j ) n ,
coefficients, and B = ( b l , b 2 , . . . , b,) is the constant matrix. Commonly, the operator
‘(0”
can be defined as follows:
(6.25)
At first, it is easy to realize that the equation can be represented by a network
shown as Figure 2 , where the activation functions of the neurons f l , f 2 , . . . , f , are
all taken as identity functions and the threshold values are zero.
z 2 A network representing fuzzy relation equations
Equation (6.25) can be solved using a a fuzzy S learning algorithm descirbed in
Section 6.3. However, if the operator is not V and A, then Equation (6.24) is difficult
© 2001 by CRC Press LLC
to solve. In this way, we consider to re-structure the network as s hown Figure 3
where the activation function of the neuron f is also taken as an identity function
and the threshold value is zero. We can interpret the network as: given a group of
training samples:
(6.26)
{ ( ( r l j ,r2j, . . . ,r n j ) ,b j ) I j = 172, . . . ,m }
Figure 3 Another network representing fuzzy relation equations
In this way, we can solve the problem by
find the weight vectors (x1,x2,..-,xn).
using an adequate learning algorithm.
In Equation (6.25), if operator “A” is replaced by operator “.”
, i.e.:
n
V (xi
*
rij)
= bj, j = 1,2,*
* *
,m
(6.27)
i=l
then Equation (6.24) is a generalized fuzzy relation equation.
If synthetic operator (V, .) is replaced by (@, .), where “@” is so-called bounded
sum, i.e.,
(6.28)
then Equation (6.24) is “almost” the same as a usual system of linear equations.
Especially, if ‘@” is replaced by “+”, i.e.,
n
rijxi =
b j , j = 1 , 2 , .. . , m
(6.29)
i=l
then it is a system of linear equations. Of course, rij and bj must not be in [0,1] (it
has already exceeded the definition of fuzzy relation equations). In other words, a
system of linear equations can be also represented by a network.
6.2.2
A Fuzzy Neural Network Based on F N ( V ,A)
Obviously, there are different types of operations for neurons existing in a fuzzy
neural network. For example, from Definition 4 we have different ASM,-func mappings, Mm, that generate different results for a neuron. A commonly used fuzzy
© 2001 by CRC Press LLC
neural
erator
network
is to take A operator
on an input and weights followed by a V opon all inputs.
Here we consider that such a fuzzy neural network,
shown in
A) f rom Definition
4. The network
is known to have fuzzy
Figure 4 7 based on FN(V,
associative
memories
ability.
Wll
W12
X1-0
.-
x2-m
x,-m
tIzi!z!
WTl77l
A fuzzy neural network
Clearly,
based on the definition,
network
is as follows:
Yl
=
Y2
= (W12
. . .
m-
Y
Rewriting
Equation
(Wll
-
l
-
Y2
l
-
Ym
L2
-h
Figure 4
Yl
base on FN(V,
the relation
between
A)
input
A Xl)
v
(w21
A x2)
v
. . . v
(W,l
A x,)
A Xl)
v
(w22
A x2)
v
. . v
(w,2
A x,)
(Wlm
(6.30)
A 51)
V (W2m
as a matrix
A X2)
form,
V ’ ’ ’ V
(Wnm
Y = (yr, ~2,. . . , ym),
(6.31)
and
I 1
Wll
W12
W21
W22
”
. . .
Wnl
For given
(6.30)
we have
X = (51, ~2,. . . ,x,)
w=
of this
A 2,)
Y =xow,
where
and output
Wn2
“’
‘.’
Wlm
’
W2m
Wnm
a set of samples:
{(as, Wls
where a, = (a,~,a,~;~~,a,,),
a weight matrix W by means
= L2,. . s,P>,
(6.32)
b, = (bsr,bS:!,...,b,,),
s = 1,2;..,p,
we can obtain
of the following
system of fuzzy relation equations:
a weight matrix W by means of the following system of fuzzy relation equations:
al o W = br
a2 o W = b2
. ..
ap o W = b,
© 2001 by CRC Press LLC
.
If we collect a, and b,, respectively, we have
and
then Equation (6.33) can be expressed by a single matrix equation as follows:
AoW=B.
(6.34)
that is:
Equation (6.21) is a fuzzy relation equation and it is not difficult to solve. We
shall discuss next a fuzzy learning algorithm.
6.3
A Fuzzy 6 Learning Algorithm
We now briefly describe procedures for the fuzzy 6 learning algorithm [13].
Step 1 Randomize
wij
initial values w;'.
(i = 1 , 2 , . . . , n,j = 1,2, . . . ,m ).
w$3 = wi"j
'
Often we can assign ( w s = I ) , (V i , j ) .
Step 2
Collect a pair of sample (a,,b,). Let s = 1.
Step 3
Calculate the outputs incurred b y a,. Let k = 1,
v
n
b:j =
i=l
© 2001 by CRC Press LLC
(wij A
a s i ) , j = 1 , 2 , . . . , m.
(6.36)
Step 4
Adjust weights. Let
6s.l . = b sg
' . - b .s 3 , j = l , 2, . . . m.
7
Update weights, i.e., calculate ( k
Wij(k
where 0
< r] 5 1 is the
Step 5
+ 1) =
+ 1)th weights based on kth weights:
i
wij
( 4- r ] & j ,
wij ( t )A as2
E
(6.37)
learning rate.
Looping. Go to Step 3 until the following condition holds:
(V i j ) ( W i j ( k ) - Wij(k
where
> bsj
otherwise,
Wij@>,
+ 1) <:
> 0 is a small number for stopping the algorithm.
Step 6 Repeat a new input. Let s
=s
(6.38)
E),
Set k = k
+ 1.
+ 1 and go to Step 2 until s = p .
We give the following example with four inputs
Example 5
Given samples a,, b,, s = 1,2,3,4:
a1
= (0.3,0.4,0.5,0.6), b l = (0.6,0.4,0.5),
(0.7,0.2,1.0,0.1),
a3 = (0.4,0.3,0.9,0.8),
a4 = (0.2,0.1,0.2,0.3),
a2 =
SowehaveA=
and
When 0.5
[
0.3 0.4 0.5 0.6
0.7 0.2 1.0 0.1 I , B =
0.4 0.3 0.9 0.8
0.2 0.1 0.2 0.3
(0.7,0.7,0.7),
= (0.8,0.4,0.5),
= (0.3,0.3,0.3).
b2 =
b3
[ ]
b4
0.6 0.4 0.5
0.7 0.7 0.7
0.8 0.4 0.5
0.3 0.3 0.3
0.3 0.4 0.5 0.6
w11 w12 w13
0.6 0.4 0.5
0.7 0.2 1.0 0.1
w21 w22 w23
0.7 0.7 0.7
< r] 5 1 and E
= 0.0001,
(6.39)
at k = 80 we have stable W as follows:
1.0 1.0 1.0
1.0 1.0 1.0
0.7 0.4 0.5
1.0 0.4 0.5
We run several tests and find out that most values in W are the same, except
w33, and w43. The following table details the difference.
© 2001 by CRC Press LLC
~ 3 1 ,
Table 1 Results for Different Tests
10.5163 ~0.700001~0.500001~0.400001~
10.4181 ~0.700001~0.500001~0.420001~
I
6.4
,
I
I
The Convergence of Fuzzy 6 Learning Rule
In this section, we shall prove that fuzzy 6 learning is convergent
Theorem 1 Let { W ( k ) l k = 1 , 2 , . . .} be the weight matrix sequence in fuzzy 6
learning rule. Then W ( k ) must be convergent.
Proof From Expression (6.33), we have the following two cases:
Case 1: If w i j ( k ) A a,i > b,j, then
n
bbj =
V ( w i j ( k )A US^) > b s j .
i=l
Therefore
6,j =
As
> 0, we
b:j
- b,j
> 0.
know that
Case 2: If w i j ( k ) A u,i 5 b,j, then
WZj(k
+ 1) = W Z j ( k ) .
Hence, based on the two cases, we always have
which means that the sequence {W(k)}is a monotonous decrease sequence.
Besides, { W ( k ) } is bounded, because
0
c W(k) c I,
where 0 is a null matrix and I is an unit matrix. Clearly, {W(k)}must be
convergent.
Q.E.D.
© 2001 by CRC Press LLC
6.5
Conclusions
In this chapter, we introduced the basic structure of a fuzzy neuron and fuzzy
neural networks. First, we described what a fuzzy neuron is by means of multifactorial functions. Then the definition of fuzzy neural networks was given by using
fuzzy neurons. We also described a fuzzy 6 learning algorithm to solve the weights
of the F N ( V ,A) type of fuzzy neural network. An example was also given. At last,
we proved that the fuzzy 6 learning algorithm must be convergent.
© 2001 by CRC Press LLC
References
1. C. T . Lin and C. S. G. Lee, Neural Fuzzy Systems, Prentice-Hall, Englewood
Cliffs, 1996.
2. B. Kosko, Fuzzy cognitive, Journal of Man-machine Studies, Vol. 24, pp.
65-75, 1986.
3. B. Kosko, Neural Networks and Fuzzy Systems, Prentice-Hall, Englewood
Cliffs, 1990.
4. S. G. Raniuk and L. 0. Hall, Fuzznet: towards a fuzzy connectionist expert
system development tool, Proceedings of IJCNN-90- WASH-DC, pp. 483-486,
1989.
5. G. A. Carpenter, S. Grossberg, and D. B. Rosen, Fuzzy ART : Fast stable
learning and categorization of analog patterns by an adaptive resonance system, Neural Networks, Vol. 4, pp. 759-771, 1991.
6. G. A. Carpenter, S. Grossberg, and D. B. Rosen, Fuzzy ART: An adaptive
resonance algorithm for rapid stable classification of analog patterns, International Joint Conference on Neural Networks, IEEE Service Center, Piscataway,
N J , 1991.
7. G. A. Carpenter, S. Grossberg, and D. B. Rosen, A neural networks realization
of Fuzzy ART, Technical Report CAS/CNS-91-021, Boston University, Boston.
8. T . Yamakawa and A. Tomoda, A fuzzy neuron and its application to pattern
recognition, 3rd IFSA Congress, pp. 330-338, 1989.
9. T. Furura, A. Kokubu, and T . Sakamoto, NFS: Neuro-fuzzy inference system,
37th meeting of IPSJ 3J-4, pp. 1386-1387, 1988.
10. H. Takagi, Fusion technology of fuzzy theory and neural networks-survey and
future directions, Proceedings of the International Conference on Fuzzy Logic
and Neural Networks, Japan, 1990.
11. H. X. Li, Multifactorial functions in fuzzy sets theory, Fuzzy Sets a n d Systems,
Vol. 35, pp. 69-84, 1990.
12. H. X. Li, Multifactorial fuzzy sets and multifactorial degree of nearness, Fuzzy
Sets and Systems, Vol. 19, No. 3, pp. 291-298, 1986.
13. X. Li, Fuzzy Neural Networks and Its Applications, Guizhou Scientific and
Technologic Press, Guizhou, China, 1994.
© 2001 by CRC Press LLC
Download