Institute for Advanced Management Systems Research Department of Information Technologies

advertisement
Institute for Advanced Management Systems Research
Department of Information Technologies
Faculty of Technology, Åbo Akademi University
Basics of Soft Computing - Tutorial
Robert Fullér
Directory
• Table of Contents
• Begin Article
c 2013
April 11, 2013
rfuller@abo.fi
Table of Contents
1. Probability, measure theory, and fuzzy sets
2. Fuzzy sets
3. Fuzzy numbers
4. Material implication
5. Fuzzy implications
6. The theory of approximate reasoning
7. Simplified fuzzy reasoning schemes
Table of Contents (cont.)
3
8. Tsukamoto’s and Sugeno’s fuzzy reasoning scheme
9. Bellman and Zadeh’s principle to fuzzy decision making
10. The continuous case
11. Aggregation operators
12. OWA Operators
Toc
JJ
II
J
I
Back
J Doc
Doc I
4
1. Probability, measure theory, and fuzzy sets
Question 1. If there a salami sandwich in the refrigerator?
The answer is 0.5.
If probability, then there either is or isn’t, with probability one half:
FIFTY/FIFTY; 50 percent YES and 50 percent NO.
If measure, then there is half a salami sandwich there. Somebody has
already eaten the other half!
If fuzzy, then there is something there, but it isn’t really a salami sandwich.
Perhaps it is some other kind of sandwich, or salami without the bread
or bread without the salami.
Toc
JJ
II
J
I
Back
J Doc
Doc I
5
2. Fuzzy sets
Fuzzy sets were introduced by Zadeh in 1965 to represent/manipulate
data and information possessing nonstatistical uncertainties.
nonstatistical uncertainties
It was specifically designed to mathematically represent uncertainty
and vagueness and to provide formalized tools for dealing with the imprecision intrinsic to many problems.
Fuzzy logic provides an inference morphology that enables approximate human reasoning capabilities to be applied to knowledge-based
systems. The theory of fuzzy logic provides a mathematical strength
to capture the uncertainties associated with human cognitive processes,
such as thinking and reasoning.
approximate reasoning
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 2: Fuzzy sets
6
Some of the essential characteristics of fuzzy logic relate to the following
(Zadeh, 1992):
• In fuzzy logic, exact reasoning is viewed as a limiting
case of approximate reasoning.
• In fuzzy logic, everything is a matter of degree.
• In fuzzy logic, knowledge is interpreted a collection of
elastic or, equivalently, fuzzy constraint on a collection
of variables.
• Inference is viewed as a process of propagation of elastic constraints.
• Any logical system can be fuzzified.
There are two main characteristics of fuzzy systems that give them better performance for specific applications.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 2: Fuzzy sets
7
• Fuzzy systems are suitable for uncertain or approximate reasoning, especially for the system with a mathematical model that is
difficult to derive.
• Fuzzy logic allows decision making with estimated values under
incomplete or uncertain information.
In classical set theory, a subset A of a set X can be defined by its characteristic function χA as a mapping from the elements of X to the elements of the set {0, 1},
χA : X → {0, 1}.
The value zero is used to represent non-membership, and the value one
is used to represent membership. The truth or falsity of the statement
”x is in A”
is determined by χA (x).
The statement is true if χA (x) = 1, and the statement is false if it is 0.
JJ
II
J
I
J Doc Doc I
Toc
Back
Section 2: Fuzzy sets
8
Figure 1: Characteristic function of A = [a, b].
Toc
JJ
χA (t) =
(
II
J
1
0
if a ≤ t ≤ b
otherwise
I
Back
J Doc
Doc I
Section 2: Fuzzy sets
9
Definition 2.1. (Zadeh, 1965) Let X be a nonempty set. A fuzzy set A in X
is characterized by its membership function
µA : X → [0, 1]
and µA (x) is interpreted as the degree of membership of element x in fuzzy
set A for each x ∈ X. Frequently we will write simply A(x) instead.
The value zero is used to represent complete non-membership, the value
one is used to represent complete membership, and values in between
are used to represent intermediate degrees of membership.
degree of satisfaction to a property
The degree to which the statement ”x is A” is true is determined by
A(x).
degree of truth
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 3: Fuzzy numbers
10
3. Fuzzy numbers
A fuzzy set A of the real line R is defined by its membership function
(denoted also by A) A : R → [0, 1]. If x ∈ R then A(x) is interpreted as
the degree of membership of x in A.
Example 3.1. Assume someone wants to buy a cheap car. Cheap car can be
represented as a fuzzy set on a universe of prices, and depends on his purse.
For instance, from the Figure ’cheap car’ is roughly interpreted as follows:
• Below 3000$ cars are considered as cheap, and prices make no real
difference to buyer’s eyes.
• Between 3000$ and 4500$, a variation in the price induces a weak
preference in favor of the cheapest car.
• Between 4500$ and 6000$, a small variation in the price induces a
clear preference in favor of the cheapest car.
• Beyond 6000$ the costs are too high (out of consideration).
JJ
II
J
I
J Doc Doc I
Toc
Back
Section 3: Fuzzy numbers
11
Figure 2: Membership function for ”cheap car”.
Example 3.2. The membership function of the fuzzy set of real numbers
”close to 1”, is can be defined as
A(t) = exp(−β(t − 1)2 )
where β is a positive real number.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 3: Fuzzy numbers
12
Figure 3: A membership function for ”x is close to 1”.
Definition 3.1. A fuzzy set A is called triangular fuzzy number with peak
(or center) a, left width α > 0 and right width β > 0 if its membership
function has the following form


 1 − a − t if a − α ≤ t ≤ a



α

t
−
a
A(t) =
if a ≤ t ≤ a + β
1−



β



0
otherwise
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 3: Fuzzy numbers
13
and we use the notation A = (a, α, β). A triangular fuzzy number with
center a may be seen as a fuzzy quantity
”x is close to a” or ”x is approximately equal to a”.
4
1. the
Fuzzyproperty
Sets and Fuzzy
x = a satisfies
”aLogic
is close to a” with degree one
[A] = ”a
[a1 (γ),
a2 (γ)].
x = a − α satisfies the property
−α
is close to a” with degree zero
γ
The support of A is the open interval (a1 (0), a2 (0)).
A is not athe
fuzzyproperty
number then
an γ ∈to
[0,a”
1] such
thatdegree
[A]γ
x = a + β Ifsatisfies
”athere
+ βexists
is close
with
zero
is not a convex subset of R.
1
a-!
a
a+"
Fig. 1.2. Triangular fuzzy number.
Figure 4: A triangular fuzzy number.
Definition 1.1.4 A fuzzy set A is called triangular fuzzy number with peak
width α > 0 and right width β > 0 if its membership
JJ a, left II
J
I
J Doc Doc I
Toc (or center)
Back
The support
of A is (a − α, b + β).
Section 3: Fuzzy
numbers
14
1
a-!
a
b
b+"
Fig. 1.3. Trapezoidal fuzzy number.
Figure 5: Trapezoidal fuzzy number.
A trapezoidal fuzzy number may be seen as a fuzzy quantity
”x is approximately in the interval [a, b]”.
DefinitionDefinition
3.2. A 1.1.6
fuzzyAny
setfuzzy
A isnumber
called
fuzzy
A ∈trapezoidal
F can be described
as number if its mem % form
&
bership function has the following
a−t

 

if t ∈ [a − α, a]
L


α t
a
−





 1
−b] α ≤ t ≤ a
−

1
if if
t ∈a[a,


% α &
A(t)
=



t − b)

 1

≤b
if if
t ∈a[b,≤
b +t β]

R
β


A(t) =




0 t − b otherwise


1−
if a ≤ t ≤ b + β


β

where [a, b] is the peak
or
core
of A,



L : [0,01] → [0, 1], R : otherwise
[0, 1] → [0, 1]
are continuous and non-increasing shape functions with L(0) = R(0) = 1 and
J intervalI
JitDoc
Toc R(1) JJ
= L(1) = 0. II
We call this fuzzy
of LR-typeBack
and refer to
by
Doc I
Section 3: Fuzzy numbers
15
and we use the notation A = (a, b, α, β). A trapezoidal fuzzy number may
be seen as a fuzzy quantity ”x is approximately in the interval [a, b]”.
Figure 6: Membership functions for monthly ”small salary” and ”big
salary”.
If x0 is the amount of the salary then x0 belongs to fuzzy set
Toc
JJ
II
A1 = small
J
I
Back
J Doc
Doc I
Section 3: Fuzzy numbers
16
with degree of membership
A1 (x0 ) =
and to


1−
 0
x0 − 2000
4000
if 2000 ≤ x0 ≤ 6000
otherwise
A2 = big
with degree of membership


1


6000 − x0
A2 (x0 ) =
1−

4000

 0
Toc
JJ
II
J
if x0 ≥ 6000
if 2000 ≤ x0 ≤ 6000
otherwise
I
Back
J Doc
Doc I
Section 4: Material implication
17
4. Material implication
Let p = ’x is in A’ and q = ’y is in B’ are crisp propositions, where A and
B are crisp sets for the moment. The full interpretation of the material
implication p → q is that: the degree of truth of p → q quantifies to
what extend q is at least as true as p, i.e.
1 if τ (p) ≤ τ (q)
τ (p → q) =
0 otherwise
τ (p)
τ (q)
1
0
0
1
1
1
0
0
τ (p → q)
1
1
1
0
Table 1: Truth table for the material implication.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 5: Fuzzy implications
18
5. Fuzzy implications
Consider the implication statement
if pressure is high then volume is small
Figure 7: Membership function for ”big pressure”.
The membership function of the fuzzy set A, big pressure, can be interpreted as
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 5: Fuzzy implications
19
• 1 is in the fuzzy set big pressure with grade of membership 0
• 4 is in the fuzzy set big pressure with grade of membership 0.75
• x is in the fuzzy set big pressure with grade of membership 1, x ≥ 5

1



if u ≥ 5
5−u
A(u) =
1−
if 1 ≤ u ≤ 5

4


0
otherwise
The membership function of the fuzzy set B, small volume, can be interpreted as

1
if v ≤ 1



v−1
B(v) =
1−
if 1 ≤ v ≤ 5

4


0
otherwise
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 5: Fuzzy implications
20
Figure 8: Membership function for ”small volume”.
• 5 is in the fuzzy set small volume with grade of membership 0
• 2 is in the fuzzy set small volume with grade of membership 0.75
• x is in the fuzzy set small volume with grade of membership 1,
x≤1
If p is a proposition of the form ’x is A’ where A is a fuzzy set, for
example, big pressure and q is a proposition of the form ’y is B’ for
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 5: Fuzzy implications
21
example, small volume then we define the implication p → q as
A(x) → B(y)
For example,
x is big pressure → y is small volume ≡ A(x) → B(y)
Remembering the full interpretation of the material implication
1 if τ (p) ≤ τ (q)
p→q=
0 otherwise
We can use the definition
A(x) → B(y) =
1
0
if A(x) ≤ B(y)
otherwise
4 is big pressure → 1 is small volume = 0.75 → 1 = 1
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 6: The theory of approximate reasoning
22
6. The theory of approximate reasoning
In 1979 Zadeh introduced the theory of approximate reasoning. This
theory provides a powerful framework for reasoning in the face of imprecise and uncertain information.
How to make inferences in fuzzy environment?
<1 :
if
observation
pressure is BIG then
pressure is 4
conclusion
<1 :
volume is ?
if
observation
pressure is BIG then
JJ
volume is SMALL
pressure is 4
conclusion
Toc
volume is SMALL
volume is 2
II
J
I
Back
J Doc
Doc I
23
Figure 9: BIG(4) = SMALL(2) = 0.75.
7. Simplified fuzzy reasoning schemes
Suppose that we have the following rule base
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 7: Simplified fuzzy reasoning schemes
<1 :
also
<2 :
<n :
fact:
24
if
x is A1 then
y is z1
if
x is A2 then
............
x is An then
y is z2
if
y is zn
x is x0
y is z0
action:
where A1 , . . . , An are fuzzy sets.
Suppose further that our data base consists of a single fact x0 . The
problem is to derive z0 from the initial content of the data base, x0 , and
from the fuzzy rule base < = {<1 , . . . , <n }.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 7: Simplified fuzzy reasoning schemes
25
<1 :
if
salary is small then
loan is z1
<2 :
if
salary is big then
loan is z2
also
salary is x0
fact:
loan is z0
action:
A deterministic rule base can be formed as follows
<1 :
if
2000 ≤ s ≤ 6000 then
loan is max 1000
<3 :
if
s ≥ 6000 then
loan is max 2000
<4 :
if
s ≤ 2000 then
no loan at all
The data base contains the actual salary, and then one of the rules is
applied to obtain the maximal loan can be obtained by the applicant.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 7: Simplified fuzzy reasoning schemes
26
Figure 10: Discrete causal link between ”salary” and ”loan”.
In fuzzy logic everything is a matter of degree.
If x is the amount of the salary then x belongs to fuzzy set
• A1 = small with degree of membership 0 ≤ A1 (x) ≤ 1
• A2 = big with degree of membership 0 ≤ A2 (x) ≤ 1
JJ
II
J
I
J Doc
Toc
Back
Doc I
Section 7: Simplified fuzzy reasoning schemes
27
Figure 11: Membership functions for ”small” and ”big”.
In fuzzy rule-based systems each rule fires.
The degree of match of the input to a rule (which is the firing strength)
is the membership degree of the input in the fuzzy set characterizing
the antecedent part of the rule.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 7: Simplified fuzzy reasoning schemes
28
The overall system output is the weighted average of the individual rule
outputs, where the weight of a rule is its firing strength with respect to
the input.
To illustrate this principle we consider a very simple example mentioned above
<1 :
if
salary is small then
loan is z1
<2 :
if
salary is big then
loan is z2
also
fact:
salary is x0
loan is z0
action:
Then our reasoning system is the following
• input to the system is x0
JJ
II
J
Toc
I
Back
J Doc
Doc I
Section 7: Simplified fuzzy reasoning schemes
29
Figure 12: Example of simplified fuzzy reasoning.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 7: Simplified fuzzy reasoning schemes
30
• the firing level of the first rule is α1 = A1 (x0 )
• the firing level of the second rule is α2 = A2 (x0 )
• the overall system output is computed as the weighted average of
the individual rule outputs
α1 z1 + α2 z2
z0 =
α1 + α2
that is
A1 (x0 )z1 + A2 (x0 )z2
z0 =
A1 (x0 ) + A2 (x0 )
A1 (x0 ) =
1 − (x0 − 2000)/4000
0

 1
1 − (6000 − x0 )/4000
A2 (x0 ) =

0
Toc
JJ
II
J
I
if 2000 ≤ x0 ≤ 6000
otherwise
if x0 ≥ 6000
if 2000 ≤ x0 ≤ 6000
otherwise
Back
J Doc
Doc I
Section 7: Simplified fuzzy reasoning schemes
31
Figure 13: Example of simplified fuzzy reasoning.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 7: Simplified fuzzy reasoning schemes
32
It is easy to see that the relationship
A1 (x0 ) + A2 (x0 ) = 1
holds for all x0 ≥ 2000.
It means that our system output can be written in the form.
z0 = α1 z1 + α2 z2 = A1 (x0 )z1 + A2 (x0 )z2
that is,
z0 =
6000 − x0
x0 − 2000
z1 + 1 −
z2
1−
4000
4000
if 2000 ≤ x0 ≤ 6000.
And z0 = 1 if x0 ≥ 6000. And z0 = 0 if x0 ≤ 2000.
The (linear) input/oputput relationship is illustrated in figure 14.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 7: Simplified fuzzy reasoning schemes
33
Figure 14: Input/output function derived from fuzzy rules.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 8: Tsukamoto’s and Sugeno’s fuzzy reasoning scheme
34
8. Tsukamoto’s and Sugeno’s fuzzy reasoning scheme
Tsukamoto’s reasoning scheme
<1 :
also
<2 :
fact :
if x is A1 and y is B1 then
z is C1
if x is A2 and y is B2 then
z is C2
x is x̄0 and y is ȳ0
cons. :
z is z0
Sugeno and Takagi use the following architecture
<1 :
if x is A1 and y is B1 then
z1 = a1 x + b1 y
<2 :
if x is A2 and y is B2 then
z2 = a2 x + b2 y
also
fact :
x is x̄0 and y is ȳ0
cons. :
Toc
JJ
z0
II
J
I
Back
J Doc
Doc I
Section 8: Tsukamoto’s and Sugeno’s fuzzy reasoning scheme
35
Figure 15: An illustration of Tsukamoto’s inference mechanism. The firing
level of the first rule: α1 = min{A1 (x0 ), B1 (y0 )} = min{0.7, 0.3} =
0.3, The firing level of the second rule: α2 = min{A2 (x0 ), B2 (y0 )} =
min{0.6, 0.8} = 0.6, The crisp inference: z0 = (8 × 0.3 + 4 × 0.6)/(0.3 +
0.6) = 6.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 8: Tsukamoto’s and Sugeno’s fuzzy reasoning scheme
36
Figure 16: Example of Sugeno’s inference mechanism. The overall system
output is computed as the firing-level-weighted average of the individual
rule outputs: z0 = (5 × 0.2 + 4 × 0.6)/(0.2 + 0.6) = 4.25.
Toc
JJ
II
J
I
Back
J Doc
Doc I
37
9. Bellman and Zadeh’s principle to fuzzy decision making
A classical MADM problem can be expressed in a matrix format. The
decision matrix is an m × n matrix whose element xij indicates the
performance rating of the i-th alternative, ai , with respect to the j-th
attribute, Xj ,


a1
x11 x12 . . . x1n



a2 
 x21 x22 . . . x2n 


..  ..
..
.. 

. 
.
.
.


am
xm1 xm2 . . . xmn
The classical maximin method is defined as: choose ak such that
sk = max si = max min xij .
i=1,...,m
i
j
where si the security level of ai , i.e. ai guarantees the decision maker a
return of at least si .
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 9: Bellman and Zadeh’s principle to fuzzy decision making
38
Example 9.1.
M athematics
English
History
John
5
8
10
M ary
6
6
7
Kate
10
7
5
Mary is selected if we use the maximin method, because her minimal performance is the maximal.
The overall performance of an alternative is determined by its weakest
performance:
A chain is only as strong as its weakest
link.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 9: Bellman and Zadeh’s principle to fuzzy decision making
39
The classical maximax return criterion is: choose ak such that
ok = max oi = max max xij .
i=1,...,m
i
j
where
oi = max xij .
j=1,...,n
is the best performance of ai .
Example 9.2.
M ath
English
History
John
5
8
10
M ary
6
6
7
Kate
10
7
5
John or Kate are selected if we use the maximax method, because their
maximal performances provide the global maximum.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 9: Bellman and Zadeh’s principle to fuzzy decision making
40
In the fuzzy case the values of the decision matrix are given as degrees
of ”how an alternative satisfies a certain attribute”. For each attribute
Xj we are given a fuzzy set µj measuring the degrees of satisfaction to
the j-th attribute.
(
1 − (8 − t)/8 if t ≤ 8
µM ath (t) =
1
otherwise
Figure 17: Membership function for attribute ’Math’.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 9: Bellman and Zadeh’s principle to fuzzy decision making
µEnglish (t) =
(
1 − (7 − t)/7
1
41
if t ≤ 7
otherwise
Figure 18: Membership function for attribute ’English’.
µHistory (t) =
Toc
JJ
II
(
1 − (6 − t)/6
if t ≤ 6
J
Back
1
I
otherwise
J Doc
Doc I
Section 9: Bellman and Zadeh’s principle to fuzzy decision making
42
Figure 19: Membership function for attribute ’History’.
John
M ary
Kate
Toc
JJ
II
M ath
English
History
5/8
1
1
6/8
6/7
1
1
1
5/6
J
I
Back
J Doc
Doc I
Section 9: Bellman and Zadeh’s principle to fuzzy decision making
43
Bellman and Zadeh’s principle to fuzzy decision making chooses the
”best compromise” alternative using the maximin method:
score(John) = min{5/8, 1, 1} = 5/8
score(Mary) = min{6/8, 6/7, 1} = 6/8
score(Kate) = min{1, 1, 5/6} = 5/6
Since
score(Kate) = max{5/8, 6/8, 5/6}.
Kate is chosen as the best student.
Toc
JJ
II
J
I
Back
J Doc
Doc I
44
10. The continuous case
If we have a continuous problem, which means that we have infinite
number of alternatives then suppose that
C1 , C2 , . . . , Cn
are the membership functions corresponding to attributes X1 , X2 , . . . ,
Xn , respectively.
Let x be an object such that for any criterion Cj , Cj (x) ∈ [0, 1] indicates
the degree to which this criterion is satisfied by x.
If we want to find out the degree to which x satisfies
’all the criteria’
denoting this by D(x), we get following
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 10: The continuous case
45
D(x) = min{C1 (x), . . . , Cn (x)}.
In this case we are essentially requiring x to satisfy
C1 and C2 and . . . and Cn .
The best alternative, denoted by x∗ , is determined from the relationship
D(x∗ ) = max D(x)
x∈X
This method is called Bellman and Zadeh’s principle to fuzzy decision
making.
The first attribute called ”x should be close to 3” is represented by the
fuzzy set C1 defined by
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 10: The continuous case
46
Figure 20: The fuzzy attributes.
C1 (x) =
(
1 − |3 − x|/2
0
if |3 − x| ≤ 2
otherwise
The second attribute called ”x should be close to 5” is represented by
the fuzzy set C2 defined by
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 10: The continuous case
C2 (x) =
47
(
1 − |5 − x|/2
0
if |5 − x| ≤ 2
otherwise
The fuzzy decision D is defined by the intersection of C1 and C2 and
the membership function of D is
D(x) =
(
1/2(1 − |4 − x|) if |4 − x| ≤ 1
0
otherwise
The optimal solution is
x∗ = 4
since
D(4) = 1/2 = max D(x).
x
Toc
JJ
II
J
I
Back
J Doc
Doc I
48
11. Aggregation operators
In a decision process the idea of trade-offs corresponds to viewing the
global evaluation of an action as lying between the worst and the best
local ratings. This occurs in the presence of conflicting goals, when a
compensation between the corresponding performances is allowed. Averaging operators realize trade-offs between objectives, by allowing a
positive compensation between ratings. An averaging (or mean) operator M is a function M : [0, 1] × [0, 1] → [0, 1], satisfying the following
properties
M (x, x) = x, ∀x ∈ [0, 1], (idempotency)
M (x, y) = M (y, x), ∀x, y ∈ [0, 1], (commutativity)
M (0, 0) = 0, M (1, 1) = 1, (extremal conditions)
M (x, y) ≤ M (x0 , y 0 ) if x ≤ x0 and y ≤ y 0 (monotonicity)
M is continuous (continuity).
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 11: Aggregation operators
49
Lemma 1. If M is an averaging operator then
min{x, y} ≤ M (x, y) ≤ max{x, y},
for all x, y ∈ [0, 1].
Proof. From idempotency and monotonicity of M it follows that
min{x, y} = M (min{x, y}, min{x, y})
≤ M (x, y)
and
M (x, y) ≤ M (max{x, y}, max{x, y})
= max{x, y}.
Which ends the proof.
The most often used mean operators.
2xy
x+y
JJ
II
harmonic mean:
Toc
J
I
Back
J Doc
Doc I
Section 11: Aggregation operators
√
geometric mean:
arithmetic mean:
xy
x+y
2
dual of geometric mean: 1 −
dual of harmonic mean:
median
50
p
(1 − x)(1 − y)
x + y − 2xy
2−x−y


 y
α
med (x, y, α) =


x
if x ≤ y ≤ α
if x ≤ α ≤ y
if α ≤ x ≤ y
generalized p-mean:
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 11: Aggregation operators
51
xp + y p
2
1/p
, p ≥ 1.
The process of information aggregation appears in many applications
related to the development of intelligent systems.
One sees aggregation in neural networks, fuzzy logic controllers, vision
systems, expert systems and multi-criteria decision aids. In
• R.R.Yager, Ordered weighted averaging aggregation operators in
multi-criteria decision making, IEEE Trans. on Systems, Man and
Cybernetics, 18(1988) 183-190.
Toc
JJ
II
J
I
Back
J Doc
Doc I
52
12. OWA Operators
Yager introduced a new aggregation technique based on the ordered
weighted averaging (OWA) operators.
Definition 12.1. An OWA operator of dimension n is a mapping
F : Rn → R,
that has an associated n vector
W = (w1 , w2 , . . . , wn )T ,
such as wi ∈ [0, 1], 1 ≤ i ≤ n,
n
X
wi = 1.
i=1
Furthermore
F (a1 , . . . , an ) =
n
X
wj bj
j=1
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 12: OWA Operators
53
where bj is the j-th largest element of the bag
< a1 , . . . , an > .
Example 12.1. Assume
W = (0.4, 0.3, 0.2, 0.1)T
then
F (0.7, 1, 0.2, 0.6) = 0.4 × 1
+0.3 × 0.7 + 0.2 × 0.6 + 0.1 × 0.2 = 0.75.
A fundamental aspect of this operator is the re-ordering step, in particular an aggregate ai is not associated with a particular weight wi
but rather a weight is associated with a particular ordered position of
aggregate.
When we view the OWA weights as a column vector we shall find it
convenient to refer to the weights with the low indices as weights at the
top and those with the higher indices with weights at the bottom.
JJ
II
J
I
J Doc Doc I
Toc
Back
Section 12: OWA Operators
54
It is noted that different OWA operators are distinguished by their
weighting function.
Yager pointed out three important special cases of OWA aggregations:
• F ∗ : In this case W = W ∗ = (1, 0 . . . , 0)T and
F ∗ (a1 , . . . , an ) = max{a1 , . . . , an },
• F∗ : In this case W = W∗ = (0, 0 . . . , 1)T and
F∗ (a1 , . . . , an ) = min{a1 , . . . , an },
• FA : In this case
W = WA = (1/n, . . . , 1/n)T
and
n
FA (a1 , . . . , an ) =
Toc
JJ
II
J
I
1X
ai .
n i=1
Back
J Doc
Doc I
Section 12: OWA Operators
55
A number of important properties can be associated with the OWA operators. We shall now discuss some of these.
For any OWA operator F
F∗ (a1 , . . . , an ) ≤ F (a1 , . . . , an ) ≤ F ∗ (a1 , . . . , an ).
Thus the upper an lower star OWA operator are its boundaries.
From the above it becomes clear that for any F
max{a1 , . . . , an } ≤ F (a1 , . . . , an ) ≤ max{a1 , . . . , an }.
The OWA operator can be seen to be commutative.
Let {a1 , . . . , an } be a bag of aggregates and let
{d1 , . . . , dn }
be any permutation of the ai . Then for any OWA operator
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 12: OWA Operators
56
F (a1 , . . . , an ) = F (d1 , . . . , dn ).
A second characteristic associated with these operators is monotonicity.
Assume ai and ci are a collection of aggregates, i = 1, . . . , n such that
for each i, ai ≥ ci .
Then
F (a1 , . . . , an ) ≥ F (c1 , c2 , . . . , cn )
where F is some fixed weight OWA operator.
Another characteristic associated with these operators is idempotency.
If ai = a for all i then for any OWA operator
F (a1 , . . . , an ) = F (a, . . . , a) = a.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 12: OWA Operators
57
From the above we can see the OWA operators have the basic properties
associated with an averaging operator.
Example 12.2. A window type OWA operator takes the average of the m
arguments about the center. For this class of operators we have

if i < k
 0
1/m if k ≤ i < k + m
wi =

0
if i ≥ k + m
In order to classify OWA operators in regard to their location between
and and or, a measure of orness, associated with any vector W is introduce by Yager as follows
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 12: OWA Operators
58
n
orness(W ) =
1
n−1
1 X
(n − i)wi
n − 1 i=1
orness(W ) =
× (n − 1)w1 + · · · + wn−1
orness(W ) =
w1 +
n−2
1
× w2 + · · · +
× wn−1
n−1
n−1
It is easy to see that for any W the orness(W ) is always in the unit
interval.
Furthermore, note that the nearer W is to an or, the closer its measure
is to one; while the nearer it is to an and, the closer is to zero.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 12: OWA Operators
59
Lemma 2. Let us consider the the vectors
W ∗ = (1, 0 . . . , 0)T ,
W∗ = (0, 0 . . . , 1)T ,
WA = (1/n, . . . , 1/n)T .
Then
orness(W ∗ ) = 1,
orness(W∗ ) = 0
orness(WA ) = 0.5
Proof.
orness(W ∗ ) =
1
(n − 1)w1 + · · · + wn−1 =
n−1
1
(n − 1) + · · · + 0 = 1.
n−1
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 12: OWA Operators
60
orness(W∗ ) =
=
1
(n − 1)w1 + · · · + wn−1
n−1
1
0 + · · · + 0 = 0.
n−1
orness(WA ) =
1
n−1
1
+ ··· +
=
n−1
n
n
n(n − 1)
= 0.5.
2n(n − 1)
A measure of andness is defined as
andness(W ) = 1 − orness(W )
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 12: OWA Operators
61
Generally, an OWA opeartor with much of nonzero weights near the
top will be an orlike operator,
orness(W ) ≥ 0.5
and when much of the weights are nonzero near the bottom, the OWA
operator will be andlike operator
andness(W ) ≥ 0.5.
Example 12.3. Let W = (0.8, 0.2, 0.0)T . Then
1
orness(W ) = (2 × 0.8 + 0.2)
2
= 0.8 + 1/2 × 0.2 = 0.9
and
andness(W ) = 1 − orness(W )
= 1 − 0.9 = 0.1.
This means that the OWA operator, defined by
F (a1 , a2 , a3 ) = 0.8b1 + 0.2b2 + 0.0b3
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 12: OWA Operators
62
= 0.8b1 + 0.2b2
where bj is the j-th largest element of the bag
< a1 , a2 , a3 >,
is an orlike aggregation.
Suppose we have n appliccants for a Ph.D. program. Each application is
evaluated by experts, who provides ratings on each of the criteria from
the set
• 3
(high)
• 2
(medium)
• 1
(low)
Compensative connectives have the property that a higher degree of
satisfaction of one of the criteria can compensate for a lower degree of
satisfaction of another criteria.
JJ
II
J
I
J Doc Doc I
Toc
Back
Section 12: OWA Operators
63
Oring the criteria means full compensation and Anding the criteria means
no compensation.
We illustrate the effect of compensation rate on the overall rating:
Let us have the following ratings
(3, 2, 1)
If w = (w1 , w2 , w3 ) is an OWA weight then
1
orness(w1 , w2 , w3 ) = w1 + w2 .
2
Min operator: the overalling rating is
min{3, 2, 1} = 1
in this case there is no compensation, because
orness(0, 0, 1) = 0.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 12: OWA Operators
64
Max operator: the overalling rating is
max{3, 2, 1} = 3
in this case there is full compensation, because
orness(1, 0, 0) = 1.
Mean operator: the overalling rating is
1
(3 + 2 + 1) = 2
3
in this case the measure of compensation is 0.5, because
orness(1/3, 1/3, 1/3) = 1/3 + 1/2 × 1/3 = 1/2.
An andlike operator: the overalling rating is
0.2 × 3 + 0.1 × 2 + 0.7 × 1 = 1.5
Toc
JJ
II
J
I
Back
J Doc
Doc I
Section 12: OWA Operators
65
in this case the measure of compensation is 0.25, because
orness(0.2, 0.1, 0.7) = 0.2 + 1/2 × 0.1 = 0.25
An orlike operator: the overalling rating is
0.6 × 3 + 0.3 × 2 + 0.1 × 1 = 2.5
in this case the measure of compensation is 0.75, because
orness(0.6, 0.3, 0.1) = 0.6 + 1/2 × 0.3 = 0.75
Yager defined the measure of dispersion (or entropy) of an OWA vector
by
disp(W ) = −
X
wi ln wi .
i
We can see when using the OWA operator as an averaging operator
Disp(W ) measures the degree to which we use all the aggregates equally.
Toc
JJ
II
J
I
Back
J Doc
Doc I
Download