Some illustrations of possibilistic correlation Robert Full´er J´ozsef Mezei 1

advertisement
Some illustrations of possibilistic correlation1
Robert Fullér
IAMSR, Åbo Akademi University,
Joukahaisenkatu 3-5 A, FIN-20520 Turku
e-mail: rfuller@abo.fi
József Mezei
Turku Centre for Computer Science,
Joukahaisenkatu 3-5 B, FIN-20520 Turku
e-mail: jmezei@abo.fi
Péter Várlaki
Budapest University of Technology and Economics,
Bertalan L. u. 2, H-1111 Budapest, Hungary
and
Széchenyi István University,
Egyetem tér 1, H-9026 Győr, Hungary
e-mail: varlaki@kme.bme.hu
Abstract: In this paper we will show some examples for possibilistic correlation.
In particular, we will show (i) how the possibilistic correlation coefficient of two
linear marginal possibility distributions changes from zero to -1/2, and from -1/2
to -3/5 by taking out bigger and bigger parts from the level sets of a their joint
possibility distribution; (ii) how to compute the autocorrelation coefficient of fuzzy
time series with linear fuzzy data.
1
Introduction
A fuzzy number A is a fuzzy set R with a normal, fuzzy convex and continuous membership function of bounded support. The family of fuzzy numbers
is denoted by F. Fuzzy numbers can be considered as possibility distributions.
1
Appeared in: Robert Fullér, József Mezei and Péter Várlaki, Proceedings of the Tenth International Symposium of Hungarian Researchers on Computational Intelligence and Informatic (CINTI
2009), [ISBN 978-963-7154-96-6], November 12-14, 2009, Budapest, Hungary, pp. 647-658.
A fuzzy set C in R2 is said to be a joint possibility distribution of fuzzy numbers A, B ∈ F, if it satisfies the relationships max{x | C(x, y)} = B(y) and
max{y | C(x, y)} = A(x) for all x, y ∈ R. Furthermore, A and B are called
the marginal possibility distributions of C. Let A ∈ F be fuzzy number with a γlevel set denoted by [A]γ = [a1 (γ), a2 (γ)], γ ∈ [0, 1] and let Uγ denote a uniform
probability distribution on [A]γ , γ ∈ [0, 1].
In possibility theory we can use the principle of expected value of functions on
fuzzy sets to define variance, covariance and correlation of possibility distributions.
Namely, we equip each level set of a possibility distribution (represented by a fuzzy
number) with a uniform probability distribution, then apply their standard probabilistic calculation, and then define measures on possibility distributions by integrating these weighted probabilistic notions over the set of all membership grades.
These weights (or importances) can be given by weighting functions. A function
f : [0, 1] → R is said to be a weighting function if f is non-negative,
monotone
R1
increasing and satisfies the following normalization condition 0 f (γ)dγ = 1.
Different weighting functions can give different (case-dependent) importances to
level-sets of possibility distributions.
In 2009 we introduced a new definition of possibilistic correlation coefficient (see
[2]) that improves the earlier definition introduced by Carlsson, Fullér and Majlender in 2005 (see [1]).
Definition ([2]). The f -weighted possibilistic correlation coefficient of A, B ∈ F
(with respect to their joint distribution C) is defined by
Z 1
ρf (A, B) =
ρ(Xγ , Yγ )f (γ)dγ
(1)
0
where
cov(Xγ , Yγ )
p
ρ(Xγ , Yγ ) = p
var(Xγ ) var(Yγ )
and, where Xγ and Yγ are random variables whose joint distribution is uniform on
[C]γ for all γ ∈ [0, 1], and cov(Xγ , Yγ ) denotes their probabilistic covariance.
In other words, the f -weighted possibilistic correlation coefficient is nothing else,
but the f -weighted average of the probabilistic correlation coefficients ρ(Xγ , Yγ )
for all γ ∈ [0, 1].
Consider the case, when A(x) = B(x) = (1 − x) · χ[0,1] (x), for x ∈ R, that
is [A]γ = [B]γ = [0, 1 − γ], for γ ∈ [0, 1]. Suppose that their joint possibility
distribution is given by F (x, y) = (1 − x − y) · χT (x, y), where
T = (x, y) ∈ R2 | x ≥ 0, y ≥ 0, x + y ≤ 1 .
Figure 1: Illustration of joint possibility distribution F .
Then we have [F ]γ = (x, y) ∈ R2 | x ≥ 0, y ≥ 0, x + y ≤ 1 − γ .
This situation is depicted on Fig. 1, where we have shifted the fuzzy sets to get a
better view of the situation. In this case the f -weighted possibilistic correlation of
A and B is computed as (see [2] for details),
Z 1
1
1
− f (γ)dγ = − .
ρf (A, B) =
2
2
0
Consider now the case when A(x) = B(x) = x · χ[0,1] (x) for x ∈ R, that is
[A]γ = [B]γ = [γ, 1], for γ ∈ [0, 1]. Suppose that their joint possibility distribution
is given by
W (x, y) = max{x + y − 1, 0}.
Then we get
Z
ρf (A, B) = −
0
1
1
1
f (γ)dγ = − .
2
2
We note here that W is nothing else but the Lukasiewitz t-norm, or in the statistical
literature, W is generally referred to as the lower Fréchet-Hoeffding bound for
copulas.
2
A transition from zero to -1/2
Suppose that a family of joint possibility distribution of A and B (where A(x) =
B(x) = (1 − x) · χ[0,1] (x), for x ∈ R) is defined by
n−1
y+x≤1
n
n
−
1
Cn (x, y) =
x+y ≤1
1 − n−1

n x − y, if 0 ≤ y ≤ 1, y ≤ x,

n

0,
otherwise



 1−x−
n−1
n y,
if 0 ≤ x ≤ 1, x ≤ y,
In the following, for simplicity, we well write C instead of Cn . A γ-level set of C
is computed by
[
n
n−1
(1 − γ), 0 ≤ y ≤ 1 − γ −
x
[C]γ = (x, y) ∈ R2 | 0 ≤ x ≤
2n − 1
n
n
n−1
2
(x, y) ∈ R |
(1 − γ) ≤ x ≤ 1 − γ, 0 ≤
y ≤1−γ−x .
2n − 1
n
The density function of a uniform distribution on [C]γ can be written as


γ
 R 1
 2n − 1 , if (x, y) ∈ [C]γ
, if (x, y) ∈ [C]
f (x, y) =
=
n(1 − γ)2
γ dxdy
 [C]
 0
otherwise
0
otherwise
We can calculate the marginal density functions:

n
(2n − 1)(1 − γ − x)


,
if
(1 − γ) ≤ x ≤ 1 − γ

2

2n − 1
 (n − 1)(1 − γ)
(2n − 1)(1 − γ − n−1
n
f1 (x) =
n x)
, if 0 ≤ x ≤
(1 − γ)


2

n(1 − γ)
2n − 1


0
otherwise
and,

(2n − 1)(1 − γ − y)
n


,
if
(1 − γ) ≤ y ≤ 1 − γ

2

2n − 1
 (n − 1)(1 − γ)
(2n − 1)(1 − γ − n−1
n
f2 (y) =
n y)
, if 0 ≤ y ≤
(1 − γ)


2

n(1
−
γ)
2n
−1


0
otherwise
We can calculate the probabilistic expected values of the random variables Xγ and
Yγ , whose joint distribution is uniform on [C]γ for all γ ∈ [0, 1] as,
Z n(1−γ)
2n−1
2n − 1
n−1
M (Xγ ) =
x(1 − γ −
x)dx
2
n(1 − γ) 0
n
Z 1−γ
2n − 1
(1 − γ)(4n − 1)
+
x(1 − γ − x)dx =
2
n(1−γ)
(n − 1)(1 − γ)
6(2n − 1)
2n−1
and, M (Yγ ) =
(1 − γ)(4n − 1)
.
6(2n − 1)
1−γ
, and for n → ∞ we
(We can easily see that for n = 1 we have M (Xγ ) =
2
1−γ
find M (Xγ ) →
.) We calculate the variations of Xγ and Yγ as,
3
M (Xγ2 )
Z n(1−γ)
2n−1
2n − 1
n−1
=
x)dx
x2 (1 − γ −
2
n(1 − γ) 0
n
Z 1−γ
2n − 1
+
x2 (1 − γ − x)dx
(n − 1)(1 − γ)2 n(1−γ)
2n−1
(1 − γ)2 ((2n − 1)3 + 8n3 − 6n2 + n)
=
.
12(2n − 1)3
(We can easily see that for n = 1 we get M (Xγ2 ) =
find M (Xγ2 ) →
(1 − γ)2
.) Furthermore,
6
(1 − γ)2
, and for n → ∞ we
3
(1 − γ)2 ((2n − 1)3 + 8n3 − 6n2 + n)
12(2n − 1)3
(1 − γ)2 (4n − 1)2
(1 − γ)2 (2(2n − 1)2 + n)
=
.
−
36(2n − 1)2
36(2n − 1)2
var(Xγ ) = M (Xγ2 ) − M (Xγ )2 =
And similarly we obtain
var(Yγ ) =
(1 − γ)2 (2(2n − 1)2 + n)
.
36(2n − 1)2
(We can easily see that for n = 1 we get var(Xγ ) =
(1 − γ)2
, and for n → ∞ we
12
find var(Xγ ) →
(1 − γ)2
.) And,
18
cov(Xγ , Yγ ) = M (Xγ Yγ ) − M (Xγ )M (Yγ )
=
(1 − γ)2 n(4n − 1) (1 − γ)2 (1 − n)(4n − 1)
−
.
12(2n − 1)2
36(2n − 1)2
(We can easily see that for n = 1 we have cov(Xγ , Yγ ) = 0, and for n → ∞ we
(1 − γ)2
find cov(Xγ , Yγ ) → −
.) We can calculate the probabilisctic correlation
36
of the random variables,
cov(Xγ , Yγ )
(1 − n)(4n − 1)
p
=
ρ(Xγ , Yγ ) = p
.
2(2n − 1)2 + n
var(Xγ ) var(Yγ )
(We can easily see that for n = 1 we have ρ(Xγ , Yγ ) = 0, and for n → ∞ we find
1
ρ(Xγ , Yγ ) → − .) And finally the f -weighted possibilistic correlation of A and
2
B is computed as,
Z 1
(1 − n)(4n − 1)
ρf (A, B) =
ρ(Xγ , Yγ )f (γ)dγ =
.
2(2n − 1)2 + n
0
1
We obtain, that ρf (A, B) = 0 for n = 1 and if n → ∞ then ρf (A, B) → − .
2
3
A transition from -1/2 to -3/5
Suppose that the joint possibility distribution of A and B (where A(x) = B(x) =
(1 − x) · χ[0,1] (x), for x ∈ R) is defined by
Cn (x, y) = (1 − x − y) · χTn (x, y),
where
[
1
Tn = (x, y) ∈ R | x ≥ 0, y ≥ 0, x + y ≤ 1,
x≥y
n−1
(x, y) ∈ R2 | x ≥ 0, y ≥ 0, x + y ≤ 1, (n − 1)x ≤ y .
2
In the following, for simplicity, we well write C instead of Cn . A γ-level set of C
is computed by
[
1
γ
2
[C] = (x, y) ∈ R | 0 ≤ x ≤ (1 − γ), (n − 1)x ≤ y ≤ 1 − γ − x
n
1
(x, y) ∈ R | 0 ≤ y ≤ (1 − γ), (n − 1)y ≤ x ≤ 1 − γ − y .
n
2
The density function of a uniform distribution on [C]γ can be written as

(
n
 R 1
, if (x, y) ∈ [C]γ
, if (x, y) ∈ [C]γ
2
dxdy
(1
−
γ)
=
f (x, y) =
γ
 [C]
0
otherwise
0
otherwise
We can calculate the marginal density functions:

x
)
n(1 − γ − nx + n−1
1−γ



,
if 0 ≤ x ≤

2

(1 − γ)
n



nx
(1 − γ)
(n − 1)(1 − γ)

,
if
≤x≤
f1 (x) =
(1 − γ)2 (n − 1)
n
n


(n
−
1)(1
−
γ)
n(1
−
γ
−
x)


,
if
≤x≤1−γ



(1 − γ)2
n


0
otherwise
and,

y
)
n(1 − γ − ny + n−1
1−γ



,
if 0 ≤ y ≤

2

(1
−
γ)
n



ny
(1 − γ)
(n − 1)(1 − γ)

,
if
≤y≤
f2 (y) =
(1 − γ)2 (n − 1)
n
n


(n
−
1)(1
−
γ)
n(1
−
γ
−
y)


,
if
≤y ≤1−γ



(1 − γ)2
n


0
otherwise
We can calculate the probabilistic expected values of the random variables Xγ and
Yγ , whose joint distribution is uniform on [C]γ for all γ ∈ [0, 1] as
n
M (Xγ ) =
(1 − γ)2
Z
n
+
(1 − γ)2
Z
=
1−γ
.
3
1−γ
n
x(1 − γ − nx +
0
(n−1)(1−γ)
n
1−γ
n
x
)dx
n−1
x2
n
dx +
n−1
(1 − γ)2
Z
1−γ)
(n−1)(1−γ)
n
x(1 − γ − x)dx
1−γ
. We calculate the variations of Xγ and Yγ as,
3
Z 1−γ
n
x
n
2
x2 (1 − γ − nx +
)dx
M (Xγ ) =
2
(1 − γ) 0
n−1
Z (n−1)(1−γ)
n
x3
n
+
dx
(1 − γ)2 1−γ
n−1
n
Z 1−γ)
n
+
x2 (1 − γ − x)dx
(1 − γ)2 (n−1)(1−γ)
That is, M (Yγ ) =
n
=
(1 −
γ)2 (3n2
− 3n + 2)
.
12n2
and,
(1 − γ)2 (3n2 − 3n + 2) (1 − γ)2
−
12n2
9
2
2
(1 − γ) (5n − 9n + 6)
=
.
36n2
And, similarly, we obtain
var(Xγ ) = M (Xγ2 ) − M (Xγ )2 =
var(Yγ ) =
(1 − γ)2 (5n2 − 9n + 6)
.
36n2
From
(1 − γ)2 (3n − 2) (1 − γ)2
−
,
12n2
9
we can calculate the probabilisctic correlation of the reandom variables:
cov(Xγ , Yγ ) = M (Xγ Yγ ) − M (Xγ )M (Yγ ) =
cov(Xγ , Yγ )
−3n2 + 7n − 6
p
ρ(Xγ , Yγ ) = p
=−
.
5n2 − 9n + 6
var(Xγ ) var(Yγ )
And finally the f -weighted possibilistic correlation of A and B:
Z 1
−3n2 + 7n − 6
ρf (A, B) =
ρ(Xγ , Yγ )f (γ)dγ = −
5n2 − 9n + 6
0
We obtain, that for n = 2
1
ρf (A, B) = − ,
2
and if n → ∞, then
3
ρf (A, B) → − .
5
We note that in this extremal case the joint possibility distribution is nothing else
but the marginal distributions themselves, that is, C∞ (x, y) = 0, for any interior
point (x, y) of the unit square.
4
A trapezoidal case
Consider the case, when




x,
if 0 ≤ x ≤ 1
1,
if 1 ≤ x ≤ 2
A(x) = B(x) =
3 − x, if 2 ≤ x ≤ 3



0,
otherwise
for x ∈ R, that is [A]γ = [B]γ = [γ, 3 − γ], for γ ∈ [0, 1]. Suppose that their joint
possibility distribution is given by:

y, if 0 ≤ x ≤ 3, 0 ≤ y ≤ 1, x ≤ y, x ≤ 3 − y



1,
if 1 ≤ x ≤ 2, 1 ≤ y ≤ 2, x ≤ y
C(x, y) =
x, if 0 ≤ x ≤ 1, 0 ≤ y ≤ 3, y ≤ x, x ≤ 3 − y



0,
otherwise
Then [C]γ = (x, y) ∈ R2 | γ ≤ x ≤ 3 − γ, γ ≤ y ≤ 3 − x .
The density function of a uniform distribution on [F ]γ can be written as


2

 R 1
, if (x, y) ∈ [C]γ
, if (x, y) ∈ [C]γ
2
dxdy
=
f (x, y) =
(3
−
2γ)
γ

 [C]
0
otherwise
0
otherwise
The marginal functions are obtained as

 2(3 − γ − x)
, if γ ≤ x ≤ 3 − γ
f1 (x) =
(3 − 2γ)2

0
otherwise
and,

 2(3 − γ − y)
, if γ ≤ y ≤ 3 − γ
f2 (y) =
(3 − 2γ)2

0
otherwise
We can calculate the probabilistic expected values of the random variables Xγ and
Yγ , whose joint distribution is uniform on [C]γ for all γ ∈ [0, 1] :
Z 3−γ
2
γ+3
x(3 − γ − x)dx =
M (Xγ ) =
2
(3 − 2γ) γ
3
and,
M (Yγ ) =
2
(3 − 2γ)2
Z
3−γ
y(3 − γ − y)dx =
γ
γ+3
3
We calculate the variations of Xγ and Yγ from the formula
var(X) = M (X 2 ) − M (X)2
as
M (Xγ2 ) =
2
(3 − 2γ)2
Z
3−γ
x2 (3 − γ − x)dx =
γ
2γ 2 + 9
6
and,
var(Xγ ) = M (Xγ2 ) − M (Xγ )2 =
2γ 2 + 9 (γ + 3)2
(3 − 2γ)2
−
=
.
6
9
18
And similarly we obtain
var(Yγ ) =
(3 − 2γ)2
.
18
Using the relationship,
cov(Xγ , Yγ ) = M (Xγ Yγ ) − M (Xγ )M (Yγ ) = −
(3 − 2γ)2
,
36
we can calculate the probabilisctic correlation of the random variables:
cov(Xγ , Yγ )
1
p
ρ(Xγ , Yγ ) = p
=− .
2
var(Xγ ) var(Yγ )
And finally the f -weighted possibilistic correlation of A and B is,
Z 1
1
1
ρf (A, B) = −
f (γ)dγ = − .
2
0 2
5
Time Series With Fuzzy Data
A time series with fuzzy data is referred to as fuzzy time series (see [3]). Consider
a fuzzy time series indexed by t ∈ (0, 1],
(
x
1 − , if 0 ≤ x ≤ t
1, if x = 0
At (x) =
and A0 (x) =
t
0
otherwise
0
otherwise
It is easy to see that in this case,
[At ]γ = [0, t(1 − γ)], γ ∈ [0, 1].
If we have t1 , t2 ∈ [0, 1], then the joint possibility distribution of the corresponding
fuzzy numbers is given by:
x
y
C(x, y) = 1 − −
· χT (x, y),
t1 t2
where
x
y
T = (x, y) ∈ R | x ≥ 0, y ≥ 0, +
≤1 .
t1 t2
y
x
γ
2
≤1−γ .
Then [C] = (x, y) ∈ R | x ≥ 0, y ≥ 0, +
t1 t2
The density function of a uniform distribution on [C]γ can be written as

 R 1
, if (x, y) ∈ [C]γ
dxdy
f (x, y) =
γ
 [C]
0
otherwise
That is,
2


2
, if (x, y) ∈ [C]γ
f (x, y) =
t1 t2 (1 − γ)2
 0
otherwise
The marginal functions are obtained as

x
 2(1 − γ − t1 )
, if 0 ≤ x ≤ t1 (1 − γ)
f1 (x) =
t1 (1 − γ)2

0
otherwise
and,

y
 2(1 − γ − t2 )
, if 0 ≤ y ≤ t2 (1 − γ)
f2 (y) =
t2 (1 − γ)2

0
otherwise
We can calculate the probabilistic expected values of the random variables Xγ and
Yγ , whose joint distribution is uniform on [C]γ for all γ ∈ [0, 1] :
2
t1 (1 − γ)2
Z
2
M (Yγ ) =
t2 (1 − γ)2
Z
M (Xγ ) =
t1 (1−γ)
x(1 − γ −
x
t1 (1 − γ)
)dx =
t1
3
y(1 − γ −
y
t2 (1 − γ)
)dx =
.
t2
3
0
and
0
t2 (1−γ)
We calculate now the variations of Xγ and Yγ as,
M (Xγ2 )
2
=
t1 (1 − γ)2
t1 (1−γ)
Z
x2 (1 − γ −
0
x
t2 (1 − γ)2
)dx = 1
t1
6
and,
var(Xγ ) = M (Xγ2 ) − M (Xγ )2 =
t2 (1 − γ)2
t21 (1 − γ)2 t21 (1 − γ)2
−
= 1
.
6
9
18
And, in a similar way, we obtain,
var(Yγ ) =
t22 (1 − γ)2
.
18
From,
t1 t2 (1 − γ)2
,
36
we can calculate the probabilisctic correlation of the reandom variables,
cov(Xγ , Yγ ) = −
cov(Xγ , Yγ )
1
p
=− .
ρ(Xγ , Yγ ) = p
2
var(Xγ ) var(Yγ )
The f -weighted possibilistic correlation of At1 and At2 ,
Z
ρf (At1 , At2 ) =
0
1
1
1
− f (γ)dγ = − .
2
2
So, the autocorrelation function of this fuzzy time series is constant. Namely,
1
R(t1 , t2 ) = − ,
2
for all t1 , t2 ∈ [0, 1].
References
[1] C. Carlsson, R. Fullér and P. Majlender, On possibilistic correlation, Fuzzy
Sets and Systems, 155(2005) 425-445.
[2] R. Fullér, J. Mezei and P. Várlaki, An improved index of interactivity for fuzzy numbers, Fuzzy Sets and Systems (to appear). doi:
10.1016/j.fss.2010.06.001
[3] B. Möller, M. Beer and U. Reuter, Theoretical Basics of Fuzzy Randomness
- Application to Time Series with Fuzzy Data, In: Safety and Reliability of
Engineering Systems and Structures - Proceedings of the 9th Int. Conference
on Structural Safety and Reliability, edited by G. Augusti and G.I. Schueller
and M. Ciampoli. Millpress, Rotterdam, pages 1701-1707.
Download