Document 10948667

advertisement
Hindawi Publishing Corporation
Journal of Probability and Statistics
Volume 2013, Article ID 208950, 8 pages
http://dx.doi.org/10.1155/2013/208950
Research Article
Note on Qualitative Robustness of Multivariate
Sample Mean and Median
Evgueni Gordienko, Andrey Novikov, and J. Ruiz de Chávez
Department of Mathematics, Autonomous Metropolitan University, Iztapalapa, San Rafael Atlixco 186,
Col. Vicentina, C.P. 09340, Mexico City, DF, Mexico
Correspondence should be addressed to J. Ruiz de Chávez; jrch@xanum.uam.mx
Received 4 October 2012; Revised 11 December 2012; Accepted 12 December 2012
Academic Editor: Shein-chung Chow
Copyright © 2013 Evgueni Gordienko et al. This is an open access article distributed under the Creative Commons Attribution
License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly
cited.
It is known that the robustness properties of estimators depend on the choice of a metric in the space of distributions. We introduce
a version of Hampel’s qualitative robustness that takes into account the √𝑛-asymptotic normality of estimators in π‘…π‘˜ , and examine
such robustness of two standard location estimators in Rπ‘˜ . For this purpose, we use certain combination of the Kantorovich and
Zolotarev metrics rather than the usual Prokhorov type metric. This choice of the metric is explained by an intention to expose a
(theoretical) situation where the robustness properties of sample mean and 𝐿 1 -sample median are in reverse to the usual ones. Using
the mentioned probability metrics we show the qualitative robustness of the sample multivariate mean and prove the inequality
which provides a quantitative measure of robustness. On the other hand, we show that 𝐿 1 -sample median could not be “qualitatively
robust” with respect to the same distance between the distributions.
1. Introduction
The following Hampel’s definition (originally given for the
one-dimensional case) of qualitative robustness [1, 2] deals
with πœ‹-balls in the space of distributions rather than with
standard “contamination neighborhoods” (see for the latter,
e.g., [3, 4]).
The sequence 𝑄𝑛 , 𝑛 ≥ 1, of estimators is qualitatively
robust at the distribution L if for every πœ– > 0 there exists
Μƒ < 𝛿 entails
𝛿 > 0 such that πœ‹(L, L)
Μƒ1 , . . . , 𝑋
̃𝑛 )) < πœ–.
supπœ‹ (𝑄𝑛 (𝑋1 , . . . , 𝑋𝑛 ) , 𝑄𝑛 (𝑋
𝑛≥1
(1)
Here, and throughout, πœ‹ denotes the Prokhorov metric,
Μƒ1 , 𝑋
Μƒ2 , . . ., are i.i.d. random vectors disand 𝑋1 , 𝑋2 , . . .; 𝑋
Μƒ
tributed, respectively, as L and L.
For a metric πœ‡ on the space of distributions and random
vectors 𝑋, π‘Œ we will write πœ‡(𝑋, π‘Œ) (as in (1)) having in mind
the πœ‡-distance between the distributions of 𝑋 and π‘Œ.
By all means, the use of the Prokhorov metric is only an
option. For instance, in [5] other probability metrics in the
definition of qualitative robustness were used. (See also [6–
9] for using different probability metrics or pseudo-metrics
related to the estimation of robustness).
As noted in [1, 2] in R1 , sample means are not qualitatively
robust at any L, while sample medians are qualitatively
robust at any L having a unique median. (See also [10] for
lack of qualitative robustness of sample means in certain
Banach spaces).
Moreover, in [11] it was shown that for symmetric distributions the median is, in certain sense, the “most robust”
estimator of a center location when using the pseudo-metric
corresponding to neighborhoods of contamination type. (See
more with this respect in [8]).
At the same time, it is known from the literature that
under different circumstances, in particular, using distinct
probability metrics (as, e.g., in (1) or in other definitions) the
robustness properties of estimators can change considerably
(see, for instance, the discussion in [12]).
The first aim of the present paper is to consider a modified
version of Hampel’s definition of qualitative robustness taking
into account the √𝑛-asymptotic normality of the sequence of
2
Journal of Probability and Statistics
estimators. The formal definition is given in the next section
but, basically, we replace (1) with the following condition:
Μƒ1 , . . . , 𝑋
̃𝑛 ) + 𝛾) ≤ πœ€,
sup√𝑛𝑙 (𝑄𝑛 (𝑋1 , . . . , 𝑋𝑛 ) , 𝑄𝑛 (𝑋
𝑛≥1
(2)
where 𝛾 is some constant, and 𝑙 is a probability metric
(different from the Prokhorov metric in our case).
The second goal of the paper is to present an example of
two probability metrics 𝑙 and πœ‡ (on the space of distributions
in Rπ‘˜ ) for which the following holds:
(i) when 𝑄𝑛 = 𝑇𝑛 , 𝑛 ≥ 1, are multivariate sample means,
the left-hand side of inequality (2) (with some 𝛾) is
Μƒ
bounded by const ⋅πœ‡(L, L);
(ii) when 𝑄𝑛 = 𝑀𝑛 , 𝑛 ≥ 1, are sample medians, we give
an example of symmetric smooth distributions L and
Μƒ 𝛼 , 𝛼 > 0, in R, such that πœ‡(L, L
Μƒ 𝛼 ) → 0 as 𝛼 → 0,
L
while there is a positive constant π‘Ÿ such that the lefthand side of inequality (2) (with 𝛾 = 0) is greater than
π‘Ÿ for all sufficiently small 𝛼 > 0. Therefore, sample
medians are not qualitative robust (in our sense) with
respect to these metrics.
The metrics 𝑙 and πœ‡ are the following (the complete definitions
are given in Section 2). 𝑙 is the Kantorovich metric (see, e.g.,
[13]) and πœ‡ is certain combination of 𝑙 and of the Zolotarev
metric of order 2 (see [14]).
We should stress that this choice is not determined by
any advantages for statistical applications. Moreover, the
closeness of distributions in the Kantorovich metric implies
the closeness in the Prokhorov metric, but also reduces the
probability of “large-valued outliers” (but not the rounding
errors). Therefore, our selection of metric is not quite consistent with the standard approach to qualitative robustness
where the Prokhorov metric (or its invariant versions) is used.
Nevertheless, our choice allows to unveil the possible
unusual robustness properties of sample means and medians
and to assess the certain quantitative robustness of the
multivariate sample mean (with respect to the considered
metrics!). The obtained “robustness inequality” does not
work in the “gross error model” but (jointly with inequalities
(23)) it could be useful for quantitative assessment of the
robustness of sample means, under perturbation of data of
“rounding” type.
2. Basic Definitions
Let (𝐻, | ⋅ |) be a separable Hilbert space with the norm | ⋅
| generated by an inner product ⟨⋅, ⋅⟩, and let B(𝐻) be the
Borel 𝜎-algebra of subsets of 𝐻. Let also 𝑋; 𝑋1 , . . . , 𝑋𝑛 , . . . and
̃𝑋
Μƒ1 , . . . , 𝑋
̃𝑛 , . . . be two sequences of i.i.d. random vectors in
𝑋;
Μƒ
𝐻 with their respective distributions denoted by L and L.
Under the assumption
𝐸 |𝑋| < ∞,
󡄨 ̃󡄨󡄨
󡄨󡄨 < ∞,
𝐸 󡄨󡄨󡄨󡄨𝑋
󡄨
(3)
Μƒ are defined as the corresponding
the means 𝐸𝑋 and 𝐸𝑋
Bochner integrals.
The definition of the median of 𝑋 is less standard (consult,
e.g., [15] for different definitions in 𝐻 = Rπ‘˜ ). We will use
the definition of median given in general setting in [16]
(sometimes called L1 -median):
𝑀 := arg min 𝐸 (|𝑋 − π‘₯| − |π‘₯|) ,
(4)
󡄨󡄨
Μƒ
Μƒ := arg min 𝐸 (󡄨󡄨󡄨󡄨𝑋
󡄨
𝑀
󡄨 − π‘₯󡄨󡄨 − |π‘₯|) .
(5)
π‘₯∈𝐻
π‘₯∈𝐻
In [16] it was shown that 𝑀 exists, and it is unique unless
L is concentrated on a one-dimensional subspace of 𝐻. In
the last case the set of minimizers in (4) is {𝑑π‘₯0 : 𝑑 ∈ [π‘Ž, 𝑏]}
for some π‘₯0 ∈ 𝐻, and one can set 𝑀 := ((π‘Ž + 𝑏)/2)π‘₯0 .
In what follows we denote (𝑛 = 1, 2, . . .):
Μƒ1 + ⋅ ⋅ ⋅ + 𝑋
̃𝑛 ,
𝑆̃𝑛 := 𝑋
𝑆𝑛 := 𝑋1 + ⋅ ⋅ ⋅ + 𝑋𝑛 ,
𝑇𝑛 =
𝑆𝑛
,
𝑛
Μƒ
̃𝑛 := 𝑆𝑛
𝑇
𝑛
(6)
(sample means) .
On the other hand, let (𝑛 = 1, 2, . . .),
𝑀𝑛 = 𝑀𝑛 (𝑋1 , . . . , 𝑋𝑛 ) ,
̃𝑛 = 𝑀
̃𝑛 (𝑋
Μƒ1 , . . . , 𝑋
̃𝑛 ) (7)
𝑀
be sample medians defined by (4) and (5) replacing L and
Μƒ by the corresponding empirical distributions L𝑛 and L
̃𝑛
L
Μƒ
Μƒ
(obtained from 𝑋1 , . . . , 𝑋𝑛 and 𝑋1 , . . . , 𝑋𝑛 , resp.). Robustness
(in terms of πœ–-contamination neighborhoods) and asymptotic
normality of 𝑀𝑛 , 𝑛 ≥ 1, were proved, for instance, in [17]
(see also [18]). The qualitative robustness properties of sample
means {𝑇𝑛 } and sample medians {𝑀𝑛 } when the Prokhorov
metric is used were discussed in Introduction.
Let us first see what happens with qualitative robustness
of {𝑇𝑛 } and {𝑀𝑛 } if in the above definition we replace πœ‹ with
the Kantorovich metric:
Μƒ := sup σ΅„¨σ΅„¨σ΅„¨σ΅„¨πΈπœ™ (𝑋) − πΈπœ™ (𝑋)
Μƒ 󡄨󡄨󡄨󡄨 ,
Μƒ ≡ 𝑙 (L, L)
𝑙 (𝑋, 𝑋)
(8)
󡄨
󡄨
πœ™∈Lip
where
Lip := {πœ™ : 𝐻 󳨃󳨀→ R : πœ™ is bounded and
󡄨
󡄨 󡄨
󡄨󡄨
σ΅„¨σ΅„¨πœ™ (π‘₯) − πœ™ (𝑦)󡄨󡄨󡄨 ≤ 󡄨󡄨󡄨π‘₯ − 𝑦󡄨󡄨󡄨 ; π‘₯, 𝑦 ∈ 𝐻} .
(9)
Μƒ < ∞, and it is well known
Under condition (3) 𝑙(𝑋, 𝑋)
(see, e.g., [13]) that 𝑙(π‘Œπ‘› , π‘Œ) → 0 if and only if πœ‹(π‘Œπ‘› , π‘Œ) → 0
and 𝐸|π‘Œπ‘› | → 𝐸|π‘Œ|.
For 𝐻 = R,
∞
Μƒ =∫
𝑙 (𝑋, 𝑋)
−∞
󡄨
󡄨󡄨
󡄨󡄨𝐹𝑋 (π‘₯) − 𝐹𝑋̃ (π‘₯)󡄨󡄨󡄨 𝑑π‘₯.
(10)
Now, applying the regularity properties of 𝑙 (see, e.g.,
[13]):
𝑙(
𝑆𝑛 𝑆̃𝑛
1 𝑛
Μƒ ,
̃𝑖 ) = 𝑙 (L, L)
, ) ≤ ∑𝑙 (𝑋𝑖 , 𝑋
𝑛 𝑛
𝑛 𝑖=1
(11)
we see that the sequence of sample means {𝑇𝑛 , 𝑛 ≥ 1} is
“qualitatively robust” using 𝑙 instead of the Prokhorov metric.
Journal of Probability and Statistics
3
It seems straightforward (using the approach similar to
one given in [10, 19] and results in [16]) to show that the
sequence of sample medians {𝑀𝑛 , 𝑛 ≥ 1} is also “qualitatively
robust” with respect to 𝑙. However, this is out of the scope of
this paper.
3. 𝑙−πœ‡−√𝑛-Robustness and Main Results
π‘˜
Now and in what follows we suppose that 𝐻 = R with the
Euclidean norm |⋅|. The results presented in this section are an
extension to the multidimensional case of the similar findings
for 𝐻 = R, published in the hardly accessible proceedings
[20]. Moreover, we improve the results of [20] even in the onedimensional case.
In order to simplify calculations in the proof of Theorem 5
below we will use in the definitions of Kantorovich’s and
Zolotarev’s metrics (see below) the following norm β€–π‘₯β€– :=
∑π‘˜π‘–=1 |π‘₯𝑖 | in the space Rπ‘˜ .
Thus, in what follows the Kantorovich metric 𝑙 is defined
by relationships (8), (9), where in (9) in place of the norm | ⋅ |
the norm || ⋅ || is used.
Let πœ‡ be some fixed simple probability metric on the set
of all probability distributions on (Rπ‘˜ , B(Rπ‘˜ )).
We will consider a sequence 𝑄𝑛 , 𝑛 ≥ 1, of estimators
of some parameter πœƒ ∈ Rπ‘˜ of the distribution L (πœƒΜƒ of the
Μƒ resp.).
distribution L,
Definition 1. We say that a sequence 𝑄𝑛 , 𝑛 ≥ 1, is (𝑙 − πœ‡ −
Μƒ ∈
√𝑛)-robust at L if there is some fixed vector 𝛾 = 𝛾(L, L)
π‘˜
R such that for every πœ– > 0 there exists 𝛿 > 0 such that
Μƒ ≤ 𝛿 entails
πœ‡(L, L)
Μƒ1 , . . . , 𝑋
̃𝑛 ) + 𝛾) ≤ πœ–.
sup√𝑛𝑙 (𝑄𝑛 (𝑋1 , . . . , 𝑋𝑛 ) , 𝑄𝑛 (𝑋
(12)
𝑛≥1
Remark 2. Taking into account that 𝑙(π‘Žπ‘‹ + 𝑏, π‘Žπ‘Œ + 𝑏) =
π‘Žπ‘™(𝑋, π‘Œ), π‘Ž ≥ 0, 𝑏 ∈ Rπ‘˜ , we see that (12) can be related
with the √𝑛-asymptotic normality of estimators 𝑄𝑛 , 𝑛 ≥ 1.
The “scaling parameter” 𝛾 (in (12)) is necessary to ensure
the equality of means of the corresponding limit normal
Μƒ in
distributions (when they exist). Only in case 𝐸𝑋 = 𝐸𝑋
(12) 𝛾 = 0.
Zolotarev’s probability metric of order 2 𝜁2 is defined as
follows (see, e.g., [14, 21]):
Μƒ := sup σ΅„¨σ΅„¨σ΅„¨σ΅„¨πΈπœ™ (𝑋) − πΈπœ™ (𝑋)
Μƒ 󡄨󡄨󡄨󡄨 ,
Μƒ ≡ 𝜁2 (L, L)
𝜁2 (𝑋, 𝑋)
󡄨
󡄨
πœ™∈D
2
(13)
where
Μƒ On the other hand, if there exist second moments,
𝐸𝑋 =ΜΈ 𝐸𝑋.
Μƒ2 < ∞, then 𝜁2 (𝑋, 𝑋)
Μƒ < ∞
Μƒ and 𝐸𝑋2 < ∞, 𝐸𝑋
𝐸𝑋 = 𝐸𝑋
(see, e.g., [14]). The function πœ™(π‘₯) := (1/2)|π‘₯|2 ∈ D2 in (14).
Therefore, by (13),
󡄨󡄨 2
Μƒ2 󡄨󡄨󡄨󡄨 ≤ 𝜁2 (𝑋, 𝑋)
Μƒ .
󡄨󡄨𝐸𝑋 − 𝐸𝑋
󡄨
󡄨
Μƒ < ∞ then 𝐸𝑋
Μƒ2 < ∞.
Thus, if 𝐸𝑋2 < ∞ and 𝜁2 (𝑋, 𝑋)
2
Μƒ <
Μƒ
Consequently, if 𝐸𝑋 = 𝐸𝑋 and 𝐸|𝑋| < ∞, then 𝜁2 (𝑋, 𝑋)
2
Μƒ < ∞.
∞ if and only if 𝐸|𝑋|
We now define the metric πœ‡ to work with:
Μƒ := max {2𝑙 (𝑋, 𝑋)
Μƒ , inf 𝜁2 (𝑋, 𝑋
Μƒ + 𝑏)} .
πœ‡ (𝑋, 𝑋)
𝑏∈Rπ‘˜
π‘₯, 𝑦 ∈ Rπ‘˜ } ,
(14)
and π·πœ™ = (𝐷1 πœ™, . . . , π·π‘˜ πœ™) is the gradient of πœ™.
Μƒ can take infinite value.
Remark 3. The distance 𝜁2 (𝑋, 𝑋)
Μƒ = ∞ if
Particularly, from (13), (14) we see that 𝜁2 (𝑋, 𝑋)
(16)
3.1. (𝑙−πœ‡−√𝑛)-Robustness of Sample Means. To prove the
inequality in Theorem 5 below we need to impose the
following restriction on the distribution L.
Assumption 4. (i) 𝐸|𝑋|π‘˜+2 < ∞, and the covariance matrix
M of 𝑋 is positive definite.
The distribution L of 𝑋 has a density 𝑓 such that for some
𝑠 ≥ 1:
(ii) the density 𝑓𝑠 of 𝑆𝑠 := 𝑋1 + ⋅ ⋅ ⋅ + 𝑋𝑠 is bounded and
differentiable;
(iii) the gradient 𝐷𝑓𝑠 is bounded and |𝐷𝑓𝑠 | belongs to
L1 (Rπ‘˜ );
(iv) for some 𝛼 > 0
∫
|π‘₯|>𝛼𝑛
󡄨󡄨 󡄨󡄨
−1/2
󡄨󡄨𝐷𝑓𝑠 󡄨󡄨 𝑑π‘₯ = 𝑂 (𝑛 ) ,
as 𝑛 󳨀→ ∞.
(17)
Note that in view of Remark 3 under the condition (i)
Μƒ < ∞ if and only if
πœ‡(𝑋, 𝑋)
󡄨 ̃󡄨󡄨2
󡄨󡄨 < ∞,
𝐸󡄨󡄨󡄨󡄨𝑋
󡄨
(18)
(see (15)).
Theorem 5. Under Assumption 4 and supposing (18) it holds
Μƒ ,
̃𝑛 + 𝛾) ≤ π‘πœ‡ (L, L)
sup√𝑛𝑙 (𝑇𝑛 , 𝑇
𝑛≥1
(19)
Μƒ
where 𝛾 = 𝐸𝑋 − 𝐸𝑋,
𝑐 = max {(10𝑠 − 1)1/2 , 5.4 π‘‘π‘˜} ,
(20)
σ΅„©
σ΅„©
𝑑 = sup max󡄩󡄩󡄩𝐷𝑖 𝑔𝑛 σ΅„©σ΅„©σ΅„©L1 (Rπ‘˜ ) < ∞,
(21)
𝑛≥𝑠 1≤𝑖≤π‘˜
σ΅„©
σ΅„© σ΅„©
σ΅„©
D2 := {πœ™ : Rπ‘˜ 󳨃󳨀→ R : σ΅„©σ΅„©σ΅„©π·πœ™ (π‘₯) − π·πœ™ (𝑦)σ΅„©σ΅„©σ΅„© ≤ σ΅„©σ΅„©σ΅„©π‘₯ − 𝑦󡄩󡄩󡄩 ;
(15)
and 𝑔𝑛 is the density of (1/√𝑛)(𝑋1 + ⋅ ⋅ ⋅ + 𝑋𝑛 ), 𝑛 ≥ 1.
Remark 6. (i) The constant 𝑐 in (20), (21) is entirely determined by the distribution L of 𝑋. For various particular
densities of 𝑋 the constant 𝑑 in (21) can be bounded by
means of computer calculations. For this one can use the fact
(true under wide conditions) that the sequence 𝐷𝑖 𝑔𝑛 , 𝑛 ≥ 𝑠,
converges in L1-norm to the corresponding partial derivative
4
Journal of Probability and Statistics
of the limit normal density with covariance matrix M (and
zero mean since ||𝐷𝑖 𝑔𝑛 ||L1 (Rπ‘˜ ) is invariant under translations).
For example, let π‘˜ = 2 and 𝑋 = (𝑋󸀠 , 𝑋󸀠󸀠 ), where 𝑋󸀠 and
σΈ€ σΈ€ 
𝑋 are independent random variables; 𝑋󸀠 has the gamma
density with 𝛼 = 2.1 and arbitrary πœ†σΈ€  > 0, while 𝑋󸀠󸀠 has the
gamma density with 𝛼 = 3 and arbitrary πœ†σΈ€ σΈ€  . Simple computer
calculations show that in (21) 𝑑 < max{0.7065πœ†σΈ€  , 0.5414πœ†σΈ€ σΈ€  },
and since we can take 𝑠 = 1, we obtain in (20) that
𝑐 < max {3, 10.8 max {0.7065πœ†σΈ€  , 0.5414πœ†σΈ€ σΈ€  }} .
(22)
For instance, 𝑐 < 7.6302 for πœ†σΈ€  = πœ†σΈ€ σΈ€  = 1. (For these values of
πœ†σΈ€  , πœ†σΈ€ σΈ€  we can take 𝑠 = 3 in (20) and obtain 𝑐 < 6.3829.)
Μƒ < ∞
(ii) Since under the above assumption πœ‡(L, L)
entails (18), inequality (19) ensures (𝑙 − πœ‡ − √𝑛)-robustness
of the sequence of sample means 𝑇𝑛 , 𝑛 ≥ 1.
(iii) For π‘˜ = 1, in [20] an example is given showing that
in general the sequence of sample means 𝑇𝑛 , 𝑛 ≥ 1 is not
Μƒ It is
(𝑙 − 𝑙 − √𝑛)-robust (even if (18) holds and 𝐸𝑋 = 𝐸𝑋).
also almost evident that the sample means 𝑇𝑛 , 𝑛 ≥ 1, are not
(𝑙 − πœ‡ − √𝑛)-robust, for example, if πœ‡ is the total variation
metric 𝑉 (or, if πœ‡ = max(𝑙, 𝑉)). The appearance of Zolotarev’s
metric on the right-hand side of (19) is related to closeness of
corresponding limit normal distributions.
Corollary 7. Suppose for a moment that πœƒ = πœƒΜƒ and that one
evaluates the quality of estimators by mean of absolute errors:
̃𝑛 − πœƒ||. Then from (19) it follows
𝛿𝑛 := 𝐸||𝑇𝑛 − πœƒ||, 𝛿̃𝑛 := 𝐸||𝑇
that
󡄨
󡄨󡄨
Μƒ ,
󡄨󡄨𝛿𝑛 − 𝛿̃𝑛 󡄨󡄨󡄨 ≤ 𝑛−1/2 const πœ‡ (L, L)
󡄨
󡄨
𝑛 ≥ 1.
(23)
(The simple proof is similar to the one given in [20]).
3.2. About (𝑙−πœ‡−√𝑛)-Robustness of Sample Medians. Let
again πœ‡ be the metric defined in (16). We show that the
sequence of sample medians 𝑀𝑛 , 𝑛 ≥ 1, in general, is not
Μƒ have strictly positive,
(𝑙 − πœ‡ − √𝑛)-robust even when 𝑋 and 𝑋
bounded, smooth densities symmetric with respect to the
̃𝑛 ,
origin, and the sequences of sample medians 𝑀𝑛 , 𝑛 ≥ 1, 𝑀
𝑛 ≥ 1, are √𝑛-asymptotically normal. We consider a modified
version of the corresponding example from [20].
Μƒπœ–
Example 8. Let 𝐻 = R, 𝑋 ∼ 𝑁(0, 1), and for πœ– ∈ (0, 1) let 𝑋
be a random variable with the density:
π‘“π‘‹Μƒπœ– (π‘₯) :=
2
π‘₯2
1
(πœ– + 3
) 𝑒−π‘₯ /2 ,
2
𝑐 (πœ–)
πœ– +π‘₯
π‘₯ ∈ R,
(24)
where 𝑐(πœ–) is a normalizing constant. By symmetry of the
density, we get
Μƒ = 𝐸𝑋 = 𝐸𝑋
Μƒ = 0,
𝑀=𝑀
(25)
and also it is clear that 𝑓𝑋, 𝑓𝑋̃ ∈ C∞ (R), and
sup
π‘“π‘‹Μƒπœ– (π‘₯) < ∞.
πœ–∈(0,1), π‘₯∈R
(26)
First of all let us show that for this example the left-hand
side of (12) is infinite for any 𝛾 =ΜΈ 0. It is well known (see,
̃𝑛 → 𝑀
Μƒ = 0 as 𝑛 → ∞
e.g., [16]) that 𝑀𝑛 → 𝑀 = 0, 𝑀
with probability 1. Also, from the results of [16] we can obtain
that |𝑀𝑛 | ≤ (2/𝑛) ∑𝑛𝑖=1 |𝑋𝑖 |. Using this inequality it is easy to
show that the sequence |𝑀𝑛 |, 𝑛 ≥ 1, is uniformly integrable.
̃𝑛 | → 0) as 𝑛 → ∞, and
Therefore, 𝐸|𝑀𝑛 | → 0 (and also 𝐸|𝑀
Μƒ
for this 𝑙(𝑀𝑛 , 𝑀𝑛 + 𝛾) → 𝑙(0, 𝛾) = 𝛾.
Let now 𝛾 = 0 in (12). By the well-known asymptotic
normality (see [22, page 307]),
̃𝑛 , πœ‚πœ– ) 󳨀→ 0,
πœ‹ (√𝑛𝑀
πœ‹ (√𝑛𝑀𝑛 , πœ‚) 󳨀→ 0,
(27)
where
πœ‚ ∼ 𝑁 (0,
1
πœ‹
) = 𝑁 (0, √ ) ,
2𝑓𝑋 (0)
2
𝑐 (πœ–)
) = 𝑁 (0,
πœ‚πœ– ∼ 𝑁 (0,
).
2π‘“π‘‹Μƒπœ– (0)
πœ–
1
(28)
From (27) and (28) it follows that there is π‘Ÿ > 0 such
that for all small enough πœ– and all large enough 𝑛 we get
̃𝑛 ) ≥ π‘Ÿ. Consequently, for all such πœ– and 𝑛,
πœ‹(√𝑛𝑀𝑛 , √𝑛𝑀
̃𝑛 ) = 𝑙 (√𝑛𝑀𝑛 , √𝑛𝑀
̃𝑛 ) ≥ π‘Ÿ2 > 0,
√𝑛𝑙 (𝑀𝑛 , 𝑀
(29)
since 𝑙 ≥ πœ‹2 (see [13, page 86]).
We have obtained that the left-hand side of (12) is positive
Μƒπœ– ) → 0
for all small enough πœ– > 0. On the other hand, πœ‡(𝑋, 𝑋
as πœ– → 0. It follows from the inequality
𝜁2 (𝑋, π‘Œ) ≤
1 ∞ 2 󡄨󡄨
󡄨
∫ π‘₯ 󡄨𝑓 (π‘₯) − π‘“π‘Œ (π‘₯)󡄨󡄨󡄨 𝑑π‘₯
2 −∞ 󡄨 𝑋
(30)
(valued when 𝐸𝑋 = πΈπ‘Œ; see [21, page 376]) and from (10).
Remark 9. The densities as in (24) represent the following somewhat strange type of “contamination.” Since
maxπ‘₯ 𝑓𝑋 (π‘₯) = 𝑓𝑋(0) sample points from 𝑓𝑋 tend to concentrate around the origin. But π‘“π‘‹Μƒπœ– (0) → 0 as πœ– → 0, and
therefore sample points from π‘“π‘‹Μƒπœ– frequently in some extent
are separated from 0.
A natural question is “how to choose the metric πœ‡Μƒ to
ensure (𝑙 − πœ‡Μƒ − √𝑛)-robustness of the sequence of sample
medians 𝑀𝑛 , 𝑛 ≥ 1?” Our conjecture is (for π‘˜ = 1, e.g.)
Μƒ = max{𝑙(𝑋, 𝑋),
Μƒ esssupπ‘₯∈R |𝑓𝑋 (π‘₯) − 𝑓̃ (π‘₯)|}
Μƒ 𝑋)
to try πœ‡(𝑋,
𝑋
(supposing the existence of densities).
Μƒ then under certain conditions the closeness
If 𝑀 = 𝑀
in πœ‡Μƒ guarantees the closeness of normal densities which are
̃𝑛 , 𝑛 ≥ 1},
limiting for {√𝑛𝑀𝑛 , 𝑛 ≥ 1} and for {√𝑛𝑀
respectively. To attempt proving (𝑙 − πœ‡Μƒ − √𝑛)-robustness of
𝑀𝑛 , 𝑛 ≥ 1 (as in (12)) one can show Hampel’s qualitative
robustness of 𝑀𝑛 , 𝑛 ≥ 1 with respect to the metric 𝑙 and
̃𝑛 ) = 𝑙(√𝑛𝑀𝑛 , √𝑛𝑀
̃𝑛 ). A
then use the property √𝑛𝑙(𝑀𝑛 , 𝑀
not clear point of this plan is finding conditions under which
𝐸|√𝑛(𝑀𝑛 − 𝑀)| → 0 as 𝑛 → ∞.
Journal of Probability and Statistics
5
Example 10. Let us give another (very simple) example of the
sequence of estimators which is (𝑙 − πœ‡ − √𝑛)-robust with
πœ‡ := 𝑙 (on the class of distributions described below). For
𝐻 = R we consider the class π‘Š of all random variables 𝑋
with bounded supports Supp(𝑋) = [0, πœƒπ‘‹ ], having density
𝑓𝑋 such that 𝑓𝑋 (π‘₯) ≥ 𝛽 > 0, π‘₯ ∈ [0, πœƒπ‘‹ ]. We suppose that
πœƒπ‘‹ ≤ πœƒ∗ < ∞, 𝑋 ∈ π‘Š, and 𝛽 is the same for all 𝑋 ∈ π‘Š.
Assume that parameter πœƒ = πœƒπ‘‹ is unknown, and the sequence
of estimators 𝑄𝑛 := max1≤𝑖≤𝑛 𝑋𝑖 is used to estimate it.
Μƒ
Denoting πœƒ = πœƒπ‘‹ , πœƒΜƒ = πœƒΜƒπ‘‹ and choosing in (12) 𝛾 = πœƒ − πœƒ,
we get
̃𝑖 )
̃𝑛 + 𝛾) = 𝑙 (πœƒ − max𝑋𝑖 , πœƒΜƒ − max𝑋
𝑙 (𝑄𝑛 , 𝑄
1≤𝑖≤𝑛
1≤𝑖≤𝑛
󡄨󡄨
󡄨
̃𝑖 )󡄨󡄨󡄨󡄨
≤ 𝐸 󡄨󡄨󡄨󡄨(πœƒ − max𝑋𝑖 ) − (πœƒΜƒ − max𝑋
󡄨󡄨
1≤𝑖≤𝑛
1≤𝑖≤𝑛
󡄨
(31)
because of the metric 𝑙 is minimal for the compound metric
𝐸|𝑋 − π‘Œ| (see, e.g., [13]).
By elementary calculations we bound the right-hand side
of (31) by
∞
̃𝑖 ) = ∫ 󡄨󡄨󡄨󡄨𝐹𝑛 (π‘₯) − 𝐹̃𝑛 (π‘₯)󡄨󡄨󡄨󡄨 𝑑π‘₯
𝑙 (max𝑋𝑖 , max𝑋
𝑋
󡄨
󡄨 𝑋
1≤𝑖≤𝑛
1≤𝑖≤𝑛
0
∞
󡄨
󡄨
Μƒ .
≤ 𝑛∫ 󡄨󡄨󡄨𝐹𝑋 (π‘₯) − 𝐹𝑋̃ (π‘₯)󡄨󡄨󡄨 𝑑π‘₯ = 𝑛𝑙 (𝑋, 𝑋)
0
(33)
Μƒ Then
On the other hand, let, for example, πœƒ ≤ πœƒ.
πœƒΜƒ
Μƒ = ∫ 󡄨󡄨󡄨𝐹𝑋 (π‘₯) − 𝐹̃ (π‘₯)󡄨󡄨󡄨 𝑑π‘₯ + ∫ (1 − 𝐹̃ (π‘₯)) 𝑑π‘₯.
𝑙 (𝑋, 𝑋)
𝑋
𝑋
󡄨
󡄨
πœƒ
0
(34)
Μƒ
But for π‘₯ ∈ [πœƒ, πœƒ],
πœƒΜƒ
1 − 𝐹𝑋̃ (π‘₯) = ∫ 𝑓𝑋̃ (𝑦) 𝑑𝑦 ≥ 𝛽 (πœƒΜƒ − π‘₯) .
Μƒ
𝑆 − 𝑛𝐸𝑋 𝑆̃𝑛 − 𝑛𝐸𝑋
= 𝑙( 𝑛
,
),
𝑛
𝑛
(35)
From this relationship it follows that
Μƒ 󳨀→ 0.
πœƒΜƒ 󳨀→ πœƒ as 𝑙 (𝑋, 𝑋)
(36)
because of 𝑙(𝑋 + 𝑏, π‘Œ + 𝑏) = 𝑙(𝑋, π‘Œ), 𝑏 ∈ Rπ‘˜ . Also it is easy to
see that 𝜁2 (𝑋 + 𝑏, π‘Œ + 𝑏) = 𝜁2 (𝑋, π‘Œ).
On the other hand, by definition of πœ‡ in (16)
Μƒ − 𝐸𝑋)
Μƒ
= 𝜁2 (𝑋 − 𝐸𝑋, 𝑋
(A.2)
(see Remark 3). Also
Μƒ − 𝐸𝑋)
Μƒ = 𝑙 (𝑋, 𝑋
Μƒ + 𝐸𝑋 − 𝐸𝑋)
Μƒ
𝑙 (𝑋 − 𝐸𝑋, 𝑋
Μƒ ,
̃󡄨󡄨󡄨󡄨 ≤ 2𝑙 (𝑋, 𝑋)
Μƒ + 𝐸 󡄨󡄨󡄨󡄨𝑋 − 𝑋
≤ 𝑙 (𝑋, 𝑋)
󡄨
󡄨
(A.3)
because the metric 𝑙 is minimal for the metric 𝐸| ⋅ |.
The above arguments show that to establish inequality
(19) it suffices to prove the version of (19) with 𝛾 = 0, 𝐸𝑋 =
Μƒ = 0, and the metric πœ‡σΈ€  := max{𝑙, 𝜁2 } on the right-hand
𝐸𝑋
side of (19).
The proof is based on the two following lemmas.
Let 𝑔 : Rπ‘˜ 󳨃→ R be a differentiable function. We will
write 𝐷𝑔 := (𝐷1 𝑔, . . . , π·π‘˜ 𝑔), where 𝐷𝑖 𝑔(π‘₯) := (πœ•/πœ•π‘₯𝑖 )𝑔(π‘₯),
𝑖 = 1, 2, . . . , π‘˜.
Lemma A.1. Let 𝑋, π‘Œ, and πœ‰ be random vectors in Rπ‘˜ such
that
(a) πœ‰ is independent of 𝑋 and π‘Œ;
(c) πœ‰ has a bounded differentiable density π‘“πœ‰ (with respect
to the Lebesgue measure) such that 𝐷𝑖 π‘“πœ‰ ∈ L1 (Rπ‘˜ ), 𝑖 =
1, 2, . . . , π‘˜.
Then
σ΅„©
σ΅„©
𝑙 (𝑋 + πœ‰, π‘Œ + πœ‰) ≤ π‘˜ max󡄩󡄩󡄩󡄩𝐷𝑖 π‘“πœ‰ σ΅„©σ΅„©σ΅„©σ΅„©L (Rπ‘˜ ) 𝜁2 (𝑋, π‘Œ) .
1≤𝑖≤π‘˜
Finally, applying (32), we can select 𝑁 in such a way that
̃𝑛 + 𝛾) ≤ πœ– .
sup√𝑛𝑙 (𝑄𝑛 , 𝑄
2
𝑛≥𝑁
(A.1)
(b) 𝐸|𝑋|2 < ∞, 𝐸|π‘Œ|2 < ∞; 𝐸𝑋 = πΈπ‘Œ;
π‘₯
Μƒ ∈ π‘Š,
for 𝑋, 𝑋
Μƒ
̃𝑛 + 𝛾) = 𝑙 ( 𝑆𝑛 , 𝑆𝑛 + 𝐸𝑋 − 𝐸𝑋)
Μƒ
𝑙 (𝑇𝑛 , 𝑇
𝑛 𝑛
(32)
Now for each 𝑛 fixed, by induction we obtain (see (10))
πœƒ
To start with the proof of Theorem 5 of the previous section,
first of all we note that in (19)
𝑏∈Rπ‘˜
1≤𝑖≤𝑛
2
.
𝛽 (𝑛 + 1)
A. The Proofs
Μƒ + 𝑏) = 𝜁2 (𝑋, 𝑋
Μƒ + 𝐸𝑋 − 𝐸𝑋)
Μƒ
inf 𝜁2 (𝑋, 𝑋
̃𝑖 ) ,
≤ 𝐸 (πœƒ − max𝑋𝑖 ) + 𝐸 (πœƒΜƒ − max𝑋
1≤𝑖≤𝑛
Appendix
(37)
For 𝑛 < 𝑁 we can use the inequality 𝑙(𝑋, π‘Œ + 𝑏) ≤ 𝑙(𝑋, π‘Œ) + |𝑏|
and (33), (36). Thus, choosing small enough 𝛿 > 0 we can
Μƒ ∈ π‘Š with πœ‡(𝑋, 𝑋)
Μƒ ≡
ensure inequality (12) for all 𝑋, 𝑋
Μƒ
𝑙(𝑋, 𝑋) ≤ 𝛿. In this way we proved (𝑙 − 𝑙 − √𝑛)-robustness
of 𝑄𝑛 on the set π‘Š.
1
(A.4)
Proof. From the definition (8), (9) of the metric 𝑙 (with the
norm ||⋅|| instead of |⋅| !) it follows (the proof is simple) that in
(8) the class of functions Lip (given in (9)) can be replaced by
the class of all bounded differentiable functions 𝑔 : Rπ‘˜ 󳨃→ R
such that
󡄨󡄨 πœ•π‘” (π‘₯) 󡄨󡄨
󡄨
󡄨󡄨
sup 󡄨󡄨󡄨
󡄨󡄨 ≤ 1, 𝑖 = 1, 2, . . . , π‘˜.
(A.5)
󡄨
󡄨
π‘₯∈Rπ‘˜ 󡄨 πœ•π‘₯𝑖 󡄨
6
Journal of Probability and Statistics
Let us fix any such function 𝑔 and arbitrary 𝑖, 𝑗 ∈
{1, 2, . . . , π‘˜}. Then (by the Fubini theorem),
Indeed, since
𝑣𝑖 (π‘₯𝑖 ) :=
𝐸𝑔 (𝑋 + πœ‰) − 𝐸𝑔 (π‘Œ + πœ‰)
Rπ‘˜
× ∫ [𝑑𝐹𝑋 (𝑑) − π‘‘πΉπ‘Œ (𝑑)] ∫ 𝑔 (π‘₯) π‘“πœ‰ (π‘₯ − 𝑑) 𝑑π‘₯.
Rπ‘˜
Rπ‘˜
σΈ€ 
πœ™π‘  (√𝑠𝑑) = ∫
Rπ‘˜−1
(A.6)
For each π‘₯ ∈ R fixed let
β„Ž (𝑑) := ∫ 𝑔 (π‘₯) π‘“πœ‰ (π‘₯ − 𝑑) 𝑑π‘₯
(A.7)
= ∫ 𝑔 (𝑦 + 𝑑) π‘“πœ‰ (𝑦) 𝑑𝑦,
Rπ‘˜
= ∫
𝑑∈R .
πœ•π‘”
πœ•β„Ž
β„Ž (𝑑) = ∫
𝑔 (𝑦 + 𝑑) π‘“πœ‰ (𝑦) 𝑑𝑦,
π‘˜
πœ•π‘‘j
R πœ•π‘‘π‘—
πœ•π‘“πœ‰
πœ•π‘”
πœ•2 β„Ž
β„Ž (𝑑) = − ∫
(π‘₯)
(π‘₯ − 𝑑) 𝑑π‘₯.
π‘˜
πœ•π‘‘π‘– πœ•π‘‘π‘—
πœ•π‘‘π‘–
R πœ•π‘₯𝑗
In view of (A.5) we obtain
󡄨󡄨
󡄨󡄨 2
󡄨󡄨 πœ•π‘“
󡄨󡄨
󡄨󡄨
󡄨󡄨 πœ• β„Ž
σ΅„©
󡄨󡄨 πœ‰
󡄨󡄨
σ΅„©
󡄨
󡄨󡄨
󡄨
󡄨󡄨 𝑑π‘₯ = 󡄩󡄩󡄩𝐷𝑖 π‘“πœ‰ σ΅„©σ΅„©σ΅„© π‘˜ .
≤
−
𝑑)
β„Ž
∫
(π‘₯
󡄨󡄨
󡄨󡄨 πœ•π‘‘ πœ•π‘‘ (𝑑)󡄨󡄨󡄨
σ΅„©
󡄨
σ΅„©L1 (R )
π‘˜
πœ•π‘‘
󡄨
󡄨
R 󡄨
󡄨󡄨
󡄨󡄨 𝑖 𝑗
𝑖
󡄨
𝑒
𝑔𝑛 (π‘₯) =
1
𝑑π‘₯
∫ 𝑣 (π‘₯ ) 𝑒𝑖𝑑𝑖 π‘₯𝑖 𝑑π‘₯𝑖 ,
𝑖𝑑𝑖 R 𝑖 𝑖
1
(2πœ‹)π‘˜
∫ πœ™π‘› (𝑑) 𝑒−π‘–βŸ¨π‘‘,π‘₯⟩ 𝑑𝑑,
(A.15)
Rπ‘˜
π‘˜
(A.8)
𝑔𝑛 (π‘₯) = ∑ 𝑛−π‘š/2 π‘ƒπ‘š (π‘₯) πœ™0,M (π‘₯) + π‘œ (
π‘š=0
(A.9)
(A.14)
σΈ€ 
and differentiate under the integral sign in (A.15).
The condition πœ™π‘› ∈ L1 (Rπ‘˜ ) (𝑛 ≥ 𝑙) and Assumption (i) in
Section 2 ensure the hypothesis of in [25, Theorem 19.2, Ch.
4]. By this theorem,
1
π‘›π‘˜/2
),
(A.16)
π‘₯ ∈ Rπ‘˜ (as 𝑛 → ∞), where π‘ƒπ‘š , π‘š = 0, . . . , π‘˜, are certain
polynomials and πœ™0,M is the normal density with zero mean
and the covariance matrix M (see Assumption (i)).
Let π‘žπ‘› denote the Fourier transform of 𝐷𝑗 β„Žπ‘› , where
π‘˜
For every π‘₯, 𝑦 ∈ Rπ‘˜ we have
β„Žπ‘› (π‘₯) := 𝑔𝑛 (π‘₯) − ∑ 𝑛−π‘š/2 π‘ƒπ‘š (π‘₯) πœ™0,M (π‘₯) .
󡄨
󡄨
σ΅„©
σ΅„©σ΅„©
σ΅„©σ΅„©π·β„Ž (π‘₯) − π·β„Ž (𝑦)σ΅„©σ΅„©σ΅„© = ∑ 󡄨󡄨󡄨󡄨𝐷𝑗 β„Ž (π‘₯) − 𝐷𝑗 β„Ž (𝑦)󡄨󡄨󡄨󡄨 ,
(A.10)
𝑗=1
πœ•2 β„Ž
󡄨󡄨
󡄨 π‘˜
σ΅„©
σ΅„©
󡄨󡄨𝐷𝑗 β„Ž (π‘₯) − 𝐷𝑗 β„Ž (𝑦)󡄨󡄨󡄨 ≤ ∑ sup
(π‘₯) σ΅„©σ΅„©σ΅„©π‘₯𝑖 − 𝑦𝑖 σ΅„©σ΅„©σ΅„© . (A.11)
󡄨
󡄨
π‘˜ πœ•π‘₯𝑖 πœ•π‘₯𝑗
𝑖=1 π‘₯∈R
Comparing (A.6)–(A.11) and taking into account the definition of 𝜁2 in (13), (14), we obtain inequality (A.4).
Lemma A.2. Under the assumption of the previous section the
constant 𝑑 in (21) is finite.
Remark A.3. A similar assertion was proved in [24] for || ⋅
||L1 -norm of second derivatives of the densities 𝑔𝑛 (under
sightly different conditions). For this reason we give only a
sketch of the proof of Lemma A.2 indicating only differences
in comparison to the proof of Lemma 4.1 in [24] (where the
omitted details can be seen).
Proof. As before, let 𝑔𝑛 (𝑛 ≥ 𝑠) be the density of 𝑆𝑛 /√𝑛, and
πœ™π‘› its characteristic function. Let also 𝑓𝑛 denote the density
of 𝑆𝑛 . There is 𝑙 ≥ 𝑠 such that |𝑑|πœ™π‘™ (𝑑) ∈ L1 (Rπ‘˜ ) and therefore
for 𝑛 ≥ 𝑙.
(A.17)
π‘š=0
π‘˜
|𝑑| πœ™π‘› (𝑑) ∈ L1 (Rπ‘˜ ) ,
R
and (A.12) follows for large enough 𝑙.
In view of (A.12) we can write down the inverse Fourier
transform for 𝑛 ≥ 𝑙:
π‘˜
Because of boundedness of πœ•π‘”/πœ•π‘‘π‘– and integrability of 𝐷𝑖 π‘“πœ‰
we can differentiate in (A.7) under the integral sign (see [23,
Appendix A]). Thus,
σΈ€ 
π‘’π‘–βŸ¨π‘‘ ,π‘₯ ⟩ 𝑑π‘₯σΈ€  ∫ 𝑓𝑠 (. . . , π‘₯𝑖 , . . .) 𝑒𝑖𝑑𝑖 π‘₯𝑖 𝑑π‘₯
π‘–βŸ¨π‘‘σΈ€  ,π‘₯σΈ€  ⟩
Rπ‘˜−1
Rπ‘˜
(A.13)
(fixing other variables), we can integrate by parts:
= ∫ 𝑔 (π‘₯) 𝑑π‘₯∫ π‘“πœ‰ (π‘₯ − 𝑑) [𝑑𝐹𝑋 (𝑑) − π‘‘πΉπ‘Œ (𝑑)]
Rπ‘˜
πœ•
𝑓 (. . . , π‘₯𝑖 , . . .) ∈ L1 (R)
πœ•π‘₯𝑖 𝑠
(A.12)
Using (A.15), (A.16), and in [25, Lemma 7.2, Ch. 2], we can
obtain that for each fixed 𝑗
π‘˜
1
π‘žπ‘› (𝑑) = 𝑑𝑗 [πœ™π‘› (𝑑) − ∑ 𝑛−π‘š/2 π‘ƒΜ‚π‘š (𝑑) exp (− βŸ¨π‘‘, Mπ‘‘βŸ©)] ,
2
π‘š=0
(A.18)
with certain polynomials π‘ƒΜ‚π‘š .
By arguments similar to those given in [24], it follows
from (A.18) that there exist constants 𝛽 > 0, 𝑐 < ∞ such
that
𝑐
󡄨
󡄨󡄨
σ΅„¨σ΅„¨π‘žπ‘› (𝑑)󡄨󡄨󡄨 𝑑𝑑 ≤ π‘˜/2
𝑛
|𝑑|≤𝛽√𝑛
∫
(𝑛 ≥ 𝑙) .
(A.19)
On the other hand, to prove that
∫
|𝑑|>𝛽√𝑛
1
󡄨
󡄨󡄨
σ΅„¨σ΅„¨π‘žπ‘› (𝑑)󡄨󡄨󡄨 = 𝑂 ( π‘˜/2 ) ,
𝑛
as 𝑛 󳨀→ ∞,
(A.20)
it is sufficient to show that
∫
|𝑑|>𝛾√𝑛
1
󡄨󡄨 󡄨󡄨 󡄨󡄨
󡄨󡄨𝑑𝑗 󡄨󡄨 σ΅„¨σ΅„¨πœ™π‘› (𝑑)󡄨󡄨󡄨󡄨 𝑑𝑑 = 𝑂 (
),
󡄨 󡄨
π‘›π‘˜/2
as 𝑛 󳨀→ ∞. (A.21)
Journal of Probability and Statistics
7
But the last equality follows from (A.12) and the fact that
󡄨󡄨
𝑑 󡄨󡄨󡄨󡄨
󡄨
sup σ΅„¨σ΅„¨σ΅„¨πœ™1 (
)󡄨 ≤ πœ… < 1.
󡄨
√𝑛 󡄨󡄨󡄨
|𝑑|>𝛾√𝑛 󡄨
(A.22)
Expressing 𝐷𝑗 β„Žπ‘› in terms of π‘žπ‘› (as an inverse Fourier
transform), and using (A.16)–(A.20) we can establish that
there exist a constant 𝑐 and polynomials π‘ƒΜƒπ‘š (π‘₯), π‘š = 0, . . . , π‘˜
such that for each 𝑗 ∈ {1, . . . , π‘˜},
󡄨󡄨
󡄨󡄨
π‘˜
󡄨󡄨
󡄨󡄨
𝑐
sup 󡄨󡄨󡄨𝐷𝑗 𝑔𝑛 (π‘₯) − ∑ 𝑛−π‘š/2 π‘ƒΜƒπ‘š (π‘₯) πœ™0,M (π‘₯)󡄨󡄨󡄨 ≤ π‘˜/2 , (A.23)
󡄨
󡄨
󡄨󡄨 𝑛
π‘₯∈Rπ‘˜ 󡄨󡄨
π‘š=0
𝑛 = 𝑙, 𝑙 + 1, . . .
The next step is to find for each 𝑗 ∈ {1, . . . , π‘˜} an upper
bound for ||𝐷𝑗 𝑔𝑛 ||L1 (Rπ‘˜ ) which does not depend on 𝑛 ≥ 𝑠.
For 𝑛 > 𝑠, 𝑔𝑛 (π‘₯) = π‘›π‘˜/2 𝑓𝑛 (√𝑛π‘₯) and 𝑓𝑛 (π‘₯) = ∫Rπ‘˜ 𝑓𝑠 (π‘₯ −
𝑑)𝑓𝑛−𝑠(𝑑) 𝑑𝑑. Thus, using Assumptions (ii), (iii), and the corresponding theorems in [23, Appendix A], we get
𝐷𝑗 𝑓𝑛 (π‘₯) = ∫ 𝐷𝑗 𝑓𝑠 (π‘₯ − 𝑑) 𝑓𝑛−𝑠 (𝑑) 𝑑𝑑.
Rπ‘˜
(A.24)
We have
σ΅„©
󡄨󡄨
󡄨
󡄨󡄨
󡄨
σ΅„©σ΅„©
󡄨󡄨𝐷𝑗 𝑔𝑛 󡄨󡄨󡄨 𝑑π‘₯ + ∫
󡄨󡄨𝐷𝑗 𝑔𝑛 󡄨󡄨󡄨 𝑑π‘₯,
󡄩󡄩𝐷𝑗 𝑔𝑛 σ΅„©σ΅„©σ΅„© π‘˜ ≤ ∫
σ΅„©L1 (R )
󡄨
󡄨
󡄨
󡄨
σ΅„©
|π‘₯|≤2√𝑛𝛼
|π‘₯|>2√𝑛𝛼
(A.25)
where 𝛼 is the constant from Assumption (iv). The first
summands on the right-hand side of (A.25) are uniformly
bounded in 𝑛 ≥ 𝑙 due to (A.23). To bound the second terms
in (A.25) we write (see (A.24))
∫
|π‘₯|>2√𝑛𝛼
󡄨󡄨
󡄨
󡄨󡄨𝐷𝑗 𝑔𝑛 󡄨󡄨󡄨 𝑑π‘₯
󡄨
󡄨
󡄨
󡄨󡄨
󡄨󡄨𝐷𝑗 𝑓𝑛 (𝑧)󡄨󡄨󡄨 𝑑𝑧
󡄨
󡄨
|𝑧|>2𝛼𝑛
= 𝑛1/2 ∫
󡄨
󡄨
𝑑𝑧 ∫ 󡄨󡄨󡄨󡄨𝐷𝑗 𝑓𝑠 (𝑧 − 𝑑)󡄨󡄨󡄨󡄨 𝑓𝑛−𝑠 (𝑑) 𝑑𝑑 =: 𝑛1/2 𝐼𝑛 .
π‘˜
|𝑧|>2𝛼𝑛
R
(A.26)
≤ 𝑛1/2 ∫
Now
𝐼𝑛 = ∫
|𝑧|>2𝛼𝑛
𝑑𝑧 ∫
|𝑑|≤𝛼𝑛
󡄨
󡄨󡄨
󡄨󡄨𝐷𝑗 𝑓𝑠 (𝑧 − 𝑑)󡄨󡄨󡄨 𝑓𝑛−𝑠 (𝑑) 𝑑𝑑
󡄨
󡄨
󡄨
󡄨󡄨
󡄨󡄨𝐷𝑗 𝑓𝑠 (𝑧 − 𝑑)󡄨󡄨󡄨 𝑓𝑛−𝑠 (𝑑) 𝑑𝑑.
𝑑𝑧 ∫
+∫
󡄨
󡄨
|𝑧|>2𝛼𝑛
|𝑑|>𝛼𝑛
(A.27)
From |𝑧| > 2𝛼𝑛 and |𝑑| ≤ 𝛼𝑛 it follows that |𝑧 − 𝑑| > 𝛼𝑛.
Thus, the first terms on the right-hand side of (A.27) are
less than 𝑐󸀠 /𝑛1/2 due to Assumption (iv). Applying the Fubini
theorem and using the fact of integrability of 𝐷𝑗 𝑓𝑠 we see
that the second summand in (A.27) is bounded by const
𝑃(| ∑𝑛−𝑠
𝑖=1 𝑋𝑖 | > 𝛼𝑛), which is 𝑂(1/𝑛) by the Chebyshev and
Rosenthal inequalities.
Exploiting Lemmas A.1 and A.2, the rest of the proof
of the Theorem 5 in Section 3 is carried out exactly as the
proof in one-dimensional case given in [20]. The “ideality
properties” of the metric 𝜁2 ,
𝑛
𝑛
𝑛
𝑖=1
𝑖=1
𝑖=1
̃𝑖 ) ≤ π‘Ž2 ∑𝜁2 (𝑋𝑖 , 𝑋
̃𝑖 ) ,
𝜁2 (π‘Ž∑𝑋𝑖 , π‘Ž∑𝑋
(A.28)
used there hold true for random vectors (see, e.g., [14]).
The general version of inequality (3.15) in [20] which relates
the Kantorovich and the total variation metrics is proved
in [13, page 89]. Note that the proof presented in [20]
uses the so-called convolution approach (see, e.g., [26]) and
induction arguments. This method has been widely used to
estimate rates of convergence in multidimensional Central
Limit Theorems (see, e.g., [13, 14, 21, 26]).
Acknowledgments
The authors thank the National System of Investigators (SNI)
of CONACYT, Mexico for the partial support of this work.
The authors are grateful to the anonymous referee for his
valuable suggestions on the improvement of an earlier version
of this paper.
References
[1] F. R. Hampel, “A general qualitative definition of robustness,”
Annals of Mathematical Statistics, vol. 42, pp. 1887–1896, 1971.
[2] F. R. Hampel, E. M. Ronchetti, P. J. Rousseeuw, and W. A. Stahel,
Robust Statistics: The Approach Based on Inuence Functions,
Wiley, New York, NY, USA, 1986.
[3] P. J. Huber, Robust Statistics, Wiley, New York, NY, USA, 1981.
[4] J. Jurečková and P. K. Sen, Robust Statistical Procedures. Asymptotics and Interrelations, Wiley, New York, NY, USA, 1996.
[5] X. He and G. Wang, “Qualitative robustness of S∗ -estimators
of multivariate location and dispersion1,” Statistica Neerlandica,
vol. 51, no. 3, pp. 257–268, 1997.
[6] L. Davies, “An efficient Fréchet differentiable high breakdown
multivariate location and dispersion estimator,” Journal of
Multivariate Analysis, vol. 40, no. 2, pp. 311–327, 1992.
[7] D. L. Donoho and R. C. Liu, “The “automatic” robustness of
minimum distance functionals,” Annals of Statistics, vol. 16, pp.
552–586, 1988.
[8] X. He and D. G. Simpson, “Lower bounds for contamination
bies: globally minimax versus locally linear estimation,” The
Annals of Statistics, vol. 21, pp. 314–337, 1993.
[9] A. Y. Kharin and P. A. Shlyk, “Robust multivariate Bayesian
forecasting under functional distortions in the πœ’2 -metric,”
Journal of Statistical Planning and Inference, vol. 139, no. 11, pp.
3842–3846, 2009.
[10] A. Cuevas, “Qualitative robustness in abstract inference,” Journal of Statistical Planning and Inference, vol. 18, no. 3, pp. 277–
289, 1988.
[11] P. J. Huber, “Robust estimation of a location parameter,” Annals
of Mathematical Statistics, vol. 35, pp. 73–101, 1964.
[12] R. Zielinski, “Robustness of sample mean and sample median
under restrictions on outliers,” Zastosowania MatematykiApplicationes Mathematicae, vol. 19, pp. 239–240, 1987.
8
[13] S. T. Rachev and L. R. Rüshendorf, Mass Transportation Problems, vol. II: Applications, Springer, New York, NY, USA, 1998.
[14] V. Zolotarev, “Probability metrics,” Theory of Probability and Its
Applications, vol. 28, pp. 264–287, 1983.
[15] C. G. Small, “A survey of multidimensional medians,” International Statistical Review, vol. 58, pp. 263–277, 1990.
[16] J. H. B. Kemperman, “The median of a finite measure on Banach
space,” in Statistical Data Analysis Based on the L1 -Norm and
Related Methods, Y. Dodge, Ed., pp. 217–230, Elsevier Science
Publishers, North Holland, 1987.
[17] D. Gervini, “Robust functional estimation using the median and
spherical principal components,” Biometrika, vol. 95, no. 3, pp.
587–600, 2008.
[18] P. Chandhuri, “Multivariate location estimation using extension
of R-estimates through U-statistics type approach,” Annals of
Statistics, vol. 30, pp. 897–916, 1992.
[19] I. Mizera, “Qualitative robustness and weak continuity: the
extreme function? Nonparametrics and robustness in modern
statistical inference and time series analysis,” in A Festschrift
in Honor of Professor Jana Jurečková Institute of Mathematical
Statistics Collections, J. Antoch, M. Hušková, and P. K. Sen,
Eds., vol. 7, pp. 169–181, Institute of Mathematical Statistics,
Beachwood, Ohio, USA, 2010.
[20] E. Gordienko and A. Novikov, “Probability metrics and robustness: is the sample median more robust than the sample mean?”
in Proceedings of the Joint Session of 7th Prague Symposium on
Asymptotic Statistics and 15th Prague Conference on Information
Theory, Statistics Decision Functions, pp. 374–386, Charles
University in Prague, Prague, Czech Republic, 2006.
[21] S. T. Rachev, Probability Metrics and the Stability of Stochastic
Models, Wiley, Chichester, UK, 2009.
[22] A. W. Van der Vaart, Asymptotic Statistics, Cambridge University Press, Cambridge, UK, 1998.
[23] R. M. Dudley, Uniform Central Limit Theorem, Cambridge
University Press, Cambridge, UK, 1999.
[24] E. Gordienko, “Comparing the distributions of sums of independent random vectors,” Kybernetika, vol. 41, no. 4, pp. 519–
529, 2005.
[25] R. N. Bhattacharya and R. Ranga Rao, Normal Approximation
and Asymptotic Expansions, Wiley, New York, NY, USA, 1976.
[26] V. V. Senatov, “Uniform estimates of the rate of convergence
in the multidimensional central limit theorem,” Theory of
Probability and Its Applications, vol. 25, pp. 745–759, 1980.
Journal of Probability and Statistics
Advances in
Operations Research
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Advances in
Decision Sciences
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Mathematical Problems
in Engineering
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Journal of
Algebra
Hindawi Publishing Corporation
http://www.hindawi.com
Probability and Statistics
Volume 2014
The Scientific
World Journal
Hindawi Publishing Corporation
http://www.hindawi.com
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
International Journal of
Differential Equations
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Volume 2014
Submit your manuscripts at
http://www.hindawi.com
International Journal of
Advances in
Combinatorics
Hindawi Publishing Corporation
http://www.hindawi.com
Mathematical Physics
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Journal of
Complex Analysis
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
International
Journal of
Mathematics and
Mathematical
Sciences
Journal of
Hindawi Publishing Corporation
http://www.hindawi.com
Stochastic Analysis
Abstract and
Applied Analysis
Hindawi Publishing Corporation
http://www.hindawi.com
Hindawi Publishing Corporation
http://www.hindawi.com
International Journal of
Mathematics
Volume 2014
Volume 2014
Discrete Dynamics in
Nature and Society
Volume 2014
Volume 2014
Journal of
Journal of
Discrete Mathematics
Journal of
Volume 2014
Hindawi Publishing Corporation
http://www.hindawi.com
Applied Mathematics
Journal of
Function Spaces
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Optimization
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Download