Hindawi Publishing Corporation Abstract and Applied Analysis Volume 2013, Article ID 148490, 6 pages http://dx.doi.org/10.1155/2013/148490 Research Article ERM Scheme for Quantile Regression Dao-Hong Xiang Department of Mathematics, Zhejiang Normal University, Jinhua, Zhejiang 321004, China Correspondence should be addressed to Dao-Hong Xiang; daohongxiang@gmail.com Received 30 November 2012; Accepted 21 February 2013 Academic Editor: Ding-Xuan Zhou Copyright © 2013 Dao-Hong Xiang. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. This paper considers the ERM scheme for quantile regression. We conduct error analysis for this learning algorithm by means of a variance-expectation bound when a noise condition is satisfied for the underlying probability measure. The learning rates are derived by applying concentration techniques involving the β2 -empirical covering numbers. 1. Introduction In this paper, we study empirical risk minimization scheme (ERM) for quantile regression. Let π be a compact metric space (input space) and π = R. Let π be a fixed but unknown probability distribution on π := π × π which describes the noise of sampling. The conditional quantile regression aims at producing functions to estimate quantile regression functions. With a prespecified quantile parameter π ∈ (0, 1), a quantile regression function ππ,π is defined by its value ππ,π (π₯) to be a π-quantile of π(⋅ | π₯), that is, a value π‘ ∈ π satisfying π ((−∞, π‘] | π₯) ≥ π, π ([π‘, ∞) | π₯) ≥ 1 − π, π₯ ∈ π, (1) where π(⋅ | π₯) is the conditional distribution of π at π₯ ∈ π. We consider a learning algorithm generated by ERM scheme associated with pinball loss and hypothesis space H. The pinball loss πΏ π : R → [0, ∞) is defined by πΏ π (π) = { (π − 1) π, ππ, if π ≤ 0, if π ≥ 0. (2) The hypothesis space H is a compact subset of πΆ(π). So there exists some π > 0 such that βπβπΆ(π) ≤ π for any π ∈ H. We assume without loss of generality βπβπΆ(π) ≤ 1 for any π ∈ H. The ERM scheme for quantile regression is defined with π a sample z = {(π₯π , π¦π )}π π=1 ∈ π drawn independently from π as follows: 1 π ∑πΏ π (π¦π − π (π₯π )) . π∈H π π=1 πz = arg min (3) A family of kernel based learning algorithms for quantile regression has been widely studied in a large literature [1–4] and references therein. The form of the algorithms is a regularized scheme in a reproducing kernel Hilbert space HπΎ (RKHS, see [5] for details) associated with a Mercer kernel πΎ. Given a sample z the kernel based regularized scheme for quantile regression is defined by πz,π = arg min { π∈HπΎ 1 π σ΅© σ΅©2 ∑πΏ π (π¦π − π (π₯π )) + πσ΅©σ΅©σ΅©πσ΅©σ΅©σ΅©πΎ } . π π=1 (4) In [1, 3, 4], error analysis for general HπΎ has been done. Learning with varying Gaussian kernel was studied in [2]. ERM scheme (3) is very different from kernel based regularized scheme (4). The output function πz produced by the ERM scheme has a uniform bound, under our assumption, βπz βπΆ(π) ≤ 1. However, we cannot expect it for πz,π . It is easy to see that πβπz,π β2πΎ ≤ ∑π π=1 |π¦π | by choosing π = 0. It happens often that βπz,π βπΎ → ∞ as π → 0. The lack of a uniform bound for πz,π has a serious negative impact on the learning rates. So in the literature of kernel based regularized schemes for quantile regression, values of the output function πz,π are always projected onto the interval [−1, 1], and error analysis is conducted for the projected function, not πz,π itself. In this paper, we aim at establishing convergence and learning rates for the error βπz − ππ,π βπΏππ in the space πΏπππ . π Here π > 0 depends on the pair (π, π) which will be decided in Section 2 and ππ is the marginal distribution of π on π. In the rest of this paper, we assume π = [−1, 1] which in turn leads to values of the target function ππ,π lie in the same interval. 2 Abstract and Applied Analysis 2. Noise Condition and Main Results There has been a large literature in learning theory (see [6] and references therein) devoted to the least square regression. It aims at learning the regression function ππ (π₯) = ∫π π¦ππ(π¦ | π₯). The identity Els (π) − Els (ππ ) = βπ − ππ β2πΏ2 for the ππ generalization error Els (π) = ∫π (π¦ − π(π₯))2 ππ leads to a variance-expectation bound with the form of Eπ2 ≤ 4Eπ, where π = (π¦ − π(π₯))2 − (π¦ − ππ (π₯))2 on (π, π). It plays an essential role in error analysis of kernel based regularized schemes. However, this identity relation and expectation-variance bound fail in the setting of the quantile regression. The reason is that the pinball loss is lack of strong convexity. If we add some noise condition on distribution π named π-quantile of π-average type π (see Definition 1), we can also get a similar identity relation which in turn enables us to have a varianceexpectation bound stated in the following which is proved by Steinwart and Christman [1]. Definition 1. Let π ∈ (0, ∞] and π ∈ [1, ∞). A distribution π on π × [−1, 1] is said to have a π-quantile of π-average type π if for ππ -almost every π₯ ∈ π, there exist a π-quantile π‘∗ ∈ R and constants πΌπ(⋅|π₯) ∈ (0, 2], ππ(⋅|π₯) > 0 such that for each π ∈ [0, πΌπ(⋅|π₯) ], π ((π‘∗ − π , π‘∗ ) | π₯) ≥ ππ(⋅|π₯) π π−1 , (5) π ((π‘∗ , π‘∗ + π ) | π₯) ≥ ππ(⋅|π₯) π π−1 , π−1 and that the function πΎ on π defined by πΎ(π₯) = ππ(⋅|π₯) πΌπ(⋅|π₯) satisfies πΎ−1 ∈ πΏπππ . We also need capacity of the hypothesis space H for our learning rates. Here in this paper, we measure the capacity by empirical covering numbers. Definition 2. Let (M, π) be a pseudometric space and π be a subset of M. For every π > 0, the covering number N(π, π, π) of π with respect to π and π is defined as the minimal number of balls of radius π whose union covers π, that is, N (π, π, π) β = min {β ∈ N : π ⊂ ∪βπ=1 π΅ (π π , π) for some {π π }π=1 ⊂ M} , (6) where π΅(π π , π) = {π ∈ M : π(π , π π ) ≤ π} is a ball in M. (π₯π )ππ=1 Definition 3. Let F be a set of functions on π, x = ⊂ π π π π and F|x = {(π(π₯π ))π=1 : π ∈ F} ⊂ R . Set N2,x (F, π) = N(F|x , π, π2 ). The β2 -empirical covering number of F is defined by N2 (F, π) = sup sup N2,x (F, π) , π∈N x∈ππ π > 0. (7) Here π2 is the normalized β2 -metric on the Euclidean space Rπ given by 1/2 1 πσ΅¨ σ΅¨2 π2 (a, b) = ( ∑σ΅¨σ΅¨σ΅¨ππ −ππ σ΅¨σ΅¨σ΅¨ ) π π=1 π π for a = (ππ )π=1 , b = (ππ )π=1 ∈ Rπ . (8) Assumption. Assume that the empirical covering number of the hypothesis space H is bounded for some π > 0 and π ∈ (0, 2), log N2 (H, π) ≤ ππ−π , ∀π > 0. (9) Theorem 4. Assume that π satisfies (5) with some π ∈ (0, ∞] and π ∈ [1, ∞). Denote π = ππ/(π + 1). One further assumes that ππ,π is uniquely defined. If ππ,π ∈ H and H satisfies (9) with π ∈ (0, 2), then for any 0 < πΏ < 1, with confidence 1 − πΏ, one has σ΅© σ΅©σ΅© Μ −π log 2 , σ΅©σ΅©πz − ππ,π σ΅©σ΅©σ΅© π ≤ πΆπ σ΅©πΏ ππ σ΅© πΏ (10) where π= 2 (π + 1) 4π (π + 1) − (2 − π) min {2 (π + 1) , ππ} (11) Μ is a constant independent of π and πΏ. and πΆ Remark 5. In the ERM scheme, we can choose ππ,π ∈ H which in turn makes the approximation error described by (23) equal to zero. However, it is impossible for the kernel based regularized scheme because of the appearance of the penalty term πβπβ2πΎ . If π ≤ 2, all conditional distributions around the quantile behave similar to the uniform distribution. In this case π = ππ/(π + 1) ≤ 2 and π = min{2/π, π/(π + 1)} = π/(π + 1) for all π > 0. Hence, π = 2(π + 1)/((2 + π)ππ + 4π). Furthermore, when π is large enough, the parameter π tends to π and the power index for the above learning rate arbitrarily approaches to 2/(2 + π)π which shows that the learning rate power index π for βπz −ππ,π βπΏπ is arbitrarily close to 2/(2+π) independent of ππ π. In particular, π can be arbitrarily small when H is smooth enough. In this case, the power index of the learning rates 2/(2 + π) can be arbitrarily close to 1 which is the optimal learning rate for the least square regression. Let us take some examples to demonstrate the above main result. Example 6. Let H be a unit ball of the sobolev space π»π with π > 0. Observe that the empirical covering number is bounded above by the uniform covering number defined in Definition 2. Hence we have (see [6, 7]) 1 π/π log N2 (H, π) ≤ πΆπ ( ) , π (12) where π is the dimension of the input space π and πΆπ > 0. Abstract and Applied Analysis 3 Under the same assumptions as Theorem 4, we get that by replacing π by π/π , for any 0 < πΏ < 1, with confidence 1 − πΏ, σ΅©σ΅© σ΅© Μ −π log 2 , σ΅©σ΅©πz − ππ,π σ΅©σ΅©σ΅© π ≤ πΆπ σ΅© σ΅©πΏ ππ πΏ (13) where 2 (π + 1) π= 4π (π + 1) − (2 − π/π ) min {2 (π + 1) , ππ} 2/(2+π/π ) independent of π. Furthermore, π can be arbitrarily large if the Sobolev space is smooth enough. In this special case, the learning rate power index arbitrarily approaches to 1. Example 7. Let H be a unit ball of the reproducing kernel Hilbert space Hπ generated by a Gaussian kernel (see [5]). Reference [7] tells us that (15) where πΆπ,π > 0 depends only on π and π > 0. Obviously, the right-hand side of (15) is bounded by πΆπ,π (1/π)π+1 . So from Theorem 4, we can get different learning rates with power index 2 (π + 1) . π= 4π (π + 1) − (1 − π) min {2 (π + 1) , ππ} (16) ππ which is very slow if π is large. However, in most data sets the data are concentrated on a much lower dimensional manifold embedded in the high dimensional space. In this setting an analysis that replaces π by the intrinsic dimension of the manifold would be of great interest (see [8] and references therein). 3. Error Analysis (19) > 0, one has 2 E {(πΏ π (π¦ − π (π₯)) − πΏ π (π¦ − ππ,π (π₯))) } π ≤ πΆπ (Eπ (π) − Eπ (ππ,π )) , (20) ∀π : π → π. The above result implies that we can get convergence rates of πz in the space πΏπππ by bounding the excess generalization error Eπ (πz ) − Eπ (ππ,π ). To bound Eπ (πz ) − Eπ (ππ,π ), we need a standard error decomposition procedure [6] and a concentration inequality. 3.1. Error Decomposition. Define the empirical error associated with the pinball loss πΏ π as Ez,π (π) = 1 π ∑πΏ (π¦ − π (π₯π )) π π=1 π π for π : π → [−1, 1] . (21) πH = arg minEπ (π) = arg min {Eπ (π) − Eπ (ππ,π )} . (22) π∈H π∈H Lemma 9. Let πΏ π be the pinball loss, πz be defined by (3) and πH ∈ H by (22). Then Eπ (πz ) − Eπ (ππ,π ) ≤ [Eπ (πz ) − Eπ (ππ,π )]−[Ez,π (πz ) − Ez,π (ππ,π )] +[Ez,π (πH )−Ez,π (ππ,π )]−[Eπ (πH ) − Eπ (ππ,π )] (23) +Eπ (πH ) − Eπ (ππ,π ) . Proof. The excess generalization error can be written as Define the noise-free error called generalization error associated with the pinball loss πΏ π as π πΆπ = σ΅© σ΅©π 22−π ππ σ΅©σ΅©σ΅©σ΅©πΎ−1 σ΅©σ΅©σ΅©σ΅©πΏπ ππ Define If π ≤ 2 and π is large enough, the power index of the π learning rates for βπz − ππ,π βπΏπ is arbitrarily close to 2/(3 + π) Eπ (π) = ∫ πΏ π (π¦ − π (π₯)) ππ 2 π π = min { , } ∈ (0, 1] , π π+1 (14) ππ ∀π > 0, 1/π σ΅© σ΅© σ΅©1/π σ΅©σ΅© σ΅©σ΅©π − ππ,π σ΅©σ΅©σ΅© π ≤ 21−1/π π1/π σ΅©σ΅©σ΅©πΎ−1 σ΅©σ΅©σ΅© π (Eπ (π) − Eπ (ππ,π )) . σ΅©πΏ ππ σ΅© σ΅©πΏ ππ σ΅© (18) Furthermore, with Μ is a constant independent of π and πΏ. and πΆ We carry out the same discussions on the case of π ≤ 2 and large enough π as Remark 5. Therefore the power index π of the learning rates for βπz − ππ,π βπΏπ is arbitrarily close to 1 π+1 log N (H, π) ≤ πΆπ,π (log ) , π Proposition 8. Let πΏ π be the pinball loss. Assume that π satisfies (5) with some π ∈ (0, ∞] and π ∈ [1, ∞). Then for all π : π → [−1, 1] one has for π : π → [−1, 1] . (17) Then the measurable function ππ,π is a minimizer of Eπ . Obviously, ππ,π (π₯) ∈ [−1, 1]. We need the following results from [1] for our error analysis. Eπ (πz ) − Eπ (ππ,π ) = [Eπ (πz ) − Ez,π (πz )] + [Ez,π (πz ) − Ez,π (πH )] + [Ez,π (πH ) − Eπ (πH )] (24) + [Eπ (πH ) − Eπ (ππ,π )] . The definition of πz implies that Ez,π (πz ) − Ez,π (πH ) ≤ 0. Furthermore, by subtracting and adding Eπ (ππ,π ) and Ez,π (ππ,π ) in the first term and third term, we see that Lemma 9 holds true. 4 Abstract and Applied Analysis We call the term (23) approximation error. It has been studied in [9]. 3.2. Concentration Inequality and Sample Error. Let us recall the one-sided Bernstein inequality as follows. Lemma 10. Let π be a random variable on a probability space π with variance π2 satisfying |π−E(π)| ≤ ππ for some constant ππ . Then for any 0 < πΏ < 1, with confidence 1 − πΏ, one has We apply Lemma 12 to a function set F, where F = {πΏ π (π¦ − π (π₯)) − πΏ π (π¦ − ππ,π (π₯)) : π ∈ H} . Proposition 13. Assume π on π × [−1, 1] satisfies the variance bound (20) with index π indicated in (19). If H satisfies (9) with π ∈ (0, 2), then for any 0 < πΏ < 1, with confidence 1 − πΏ/2, one has [Eπ (π) − Eπ (ππ,π )] − [Ez,π (π) − Ez,π (ππ,π )] 2ππ log (1/πΏ) 2π2 log (1/πΏ) 1 π +√ . ∑π (π§π ) − E (π) ≤ π π=1 3π π (25) Proposition 11. Let πH ∈ H. Assume that π on π × [−1, 1] satisfies the variance bound (20) with index π indicated in (19). For any 0 < πΏ < 1, with confidence 1 − πΏ/2, (3.6) can be bounded as [Ez,π (πH ) − Ez,π (ππ,π )] − [Eπ (πH ) − Eπ (ππ,π )] 4 log (2/πΏ) 2πΆ log (2/πΏ) 1/(2−π) ≤ +( π ) 3π π Proof. Let π(π§) = πΏ π (π¦ − πH (π₯)) − πΏ π (π¦ − ππ,π (π₯)) which satisfies |π| ≤ 2 and in turn |π − E(π)| ≤ 2. The variance bound (20) implies that π E(π − E (π))2 ≤ Eπ2 ≤ πΆπ (Eπ (πH ) − Eπ (ππ,π )) . (27) Using (25) on the random variable π(π§) = πΏ π (π¦ − πH (π₯)) − πΏ π (π¦ − ππ,π (π₯)), we can get the desired bound (28) with the help of Young’s inequality. Let us turn to estimate the sample error (3.5) involving the function πz which runs over a set of functions since z is a random sample itself. To estimate it, we use a concentration inequality below involving empirical covering numbers [10–12]. Lemma 12. Let F be a class of measurable functions on π. Assume that there are constants π΅, π > 0 and πΌ ∈ [0, 1] and βπβ∞ ≤ π΅ and Eπ2 ≤ π(Eπ)πΌ for every π ∈ F. If (7) holds, then there exists a constant ππσΈ depending only on π such that for any π‘ > 0, with probability at least 1 − π−π‘ , there holds Eπ − 1 π 1 πΌ ∑π (π§π ) ≤ π1−πΌ (Eπ) π π=1 2 + ππσΈ π + 2( ππ‘ 1/(2−πΌ) 18π΅π‘ + ) , π π (28) ∀π ∈ F, (2−π)/(2+π) π΅ π 2/(4−2πΌ+ππΌ) , ) π π 2/(2+π) ( ) }. π 1 [E (π) − Eπ (ππ,π )] 2 π 2 1 2/(4−2π+ππ) + πΆπ,π log ( ) , πΏ π (29) (31) ∀π ∈ H, where 1 πΆπ,π = ( + ππσΈ ) max {πΆπ(2−π)/(4−2π+ππ) π2/(4−2π+ππ) , 2 2(2−π)/(2+π) π2/(2+π) } + 2πΆπ1/(2−π) + 36. (32) Proof. Take π ∈ F with the form π(π§) = πΏ π (π¦ − π(π₯)) − πΏ π (π¦ − ππ,π (π₯)) where π ∈ H. Hence Eπ = Eπ (π) − Eπ (ππ,π ) and (1/π) ∑π π=1 π(π§π ) = Ez,π (π) − Ez,π (ππ,π ). The Lipschitz property of the pinball loss πΏ π implies that σ΅¨σ΅¨σ΅¨π (π§)σ΅¨σ΅¨σ΅¨ ≤ σ΅¨σ΅¨σ΅¨σ΅¨π (π₯) − ππ,π (π₯)σ΅¨σ΅¨σ΅¨σ΅¨ ≤ 2. (33) σ΅¨ σ΅¨ σ΅¨ σ΅¨ For π1 , π2 ∈ F, we have σ΅¨σ΅¨ σ΅¨σ΅¨ σ΅¨σ΅¨π1 (π§) − π2 (π§)σ΅¨σ΅¨ σ΅¨ σ΅¨ = σ΅¨σ΅¨σ΅¨πΏ π (π¦ − π1 (π₯)) − πΏ π (π¦ − π2 (π₯))σ΅¨σ΅¨σ΅¨ (34) σ΅¨σ΅¨ σ΅¨σ΅¨ ≤ σ΅¨σ΅¨π1 (π₯) − π2 (π₯)σ΅¨σ΅¨ , where π1 , π2 ∈ H. It follows that N2,z (F, π) ≤ N2,x (H, π) . (35) log N2 (F, π) ≤ ππ−π . (36) Hence Applying Lemma 12 with π΅ = 2, πΌ = π, and π = πΆπ , we know that for any 0 < πΏ < 1, with confidence 1 − πΏ/2, there holds Eπ − 1 π ∑π (π§π ) π π=1 πΆ log (2/πΏ) 1/(2−π) 1 π ≤ π1−π (Eπ) + ππσΈ π+2( π ) 2 π + where π := max {π(2−π)/(4−2πΌ+ππΌ) ( ≤ (26) + Eπ (πH ) − Eπ (ππ,π ) . (30) 36 log (2/πΏ) π πΆ log (2/πΏ) 1/(2−π) 1 1 ≤ Eπ + ( + ππσΈ ) π + 2( π ) 2 2 π + 36 log (2/πΏ) . π (37) Abstract and Applied Analysis 5 Here π ≤ max {πΆπ(2−π)/(4−2π+ππ) π2/(4−2π+ππ) , 2(2−π)/(2+π) π2/(2+π) } ×( 1 2/(4−2π+ππ) . ) π (38) Proof. Combining (18), (19), and (40), with confidence 1 − πΏ, we have σ΅©σ΅© σ΅©σ΅© σ΅©σ΅©σ΅©πz −ππ,π σ΅©σ΅©σ΅©πΏππ π Μ π (πH ) − Eπ (ππ,π )) ≤ πΆ(E 2/π(4−2π+ππ) Μ log 2 ( 1 ) +πΆ πΏ π 1/π Μ inf {Eπ (π) − Eπ (ππ,π )} ≤πΆ Note that π∈H πΆ log (2/πΏ) 1/(2−π) 36 log (2/πΏ) 1 + ( + ππσΈ ) π + 2( π ) 2 π π 2 1 2/(4−2π+ππ) ≤ πΆπ,π log ( ) , πΏ π (39) Μ= πΆ 2 (π + 1) , 4π (π + 1) − (2 − π) min {2 (π + 1) , ππ} σ΅© σ΅©1/π 4 2π1/π σ΅©σ΅©σ΅©σ΅©πΎ−1 σ΅©σ΅©σ΅©σ΅©πΏπ ( ππ 3 1/(2−π) + (2πΆπ ) + πΆπ,π ) 1/π (44) . Proof of Theorem 1. The assumption ππ,π ∈ H implies that inf {Eπ (π) − Eπ (ππ,π )} = 0. π∈H (45) Therefore, our desired result comes directly from Theorem 15. Eπ (πz ) − Eπ (ππ,π ) ≤ 2 (Eπ (πH ) − Eπ (ππ,π )) (40) 2 1 2/(4−2π+ππ) + 2πΆπ,π log ( ) . πΏ π The above bound follows directly from Propositions 11 and 13 with the fact that πz ∈ H. 3.3. Bounding the Total Error. Now we are in a position to present our general result on error analysis for algorithm (3). Theorem 15. Assume that π satisfies (5) with some π ∈ (0, ∞] and π ∈ [1, ∞). Denote πΎ = ππ/(π + 1). Further assume that H satisfies (9) with π ∈ (0, 2) and ππ,π is uniquely defined. Then for any 0 < πΏ < 1, with confidence 1 − πΏ, one has Μ −π log 2 , + πΆπ πΏ where π= Proposition 14. Assume π on π × [−1, 1] satisfies the variance bound (20) with index π indicated in (19). If H satisfies (9) with π ∈ (0, 2) Then for any 0 < πΏ < 1, with confidence 1 − πΏ, there holds σ΅©σ΅© σ΅© Μ inf {Eπ (π) − Eπ (ππ,π )}1/π σ΅©σ΅©πz − ππ,π σ΅©σ΅©σ΅© π ≤ πΆ σ΅© σ΅©πΏ ππ π∈H Μ −π log 2 , + πΆπ πΏ (43) where πΆπ,π is indicated in (32). Then our desired bound holds true. 8 log (2/πΏ) 2πΆ log (2/πΏ) 1/(2−π) + + 2( π ) 3π π 1/π (41) 4. Further Discussions In this paper, we studied ERM algorithm (3) for quantile regression and provide convergence and learning rates. We showed some essential differences between ERM scheme and kernel based regularized scheme for quantile regression. We also point out the difficulty to deal with quantile regression: the lack of strong convexity of the pinball loss. To overcome this difficulty, some noise condition on π is proposed to enable us to get a variance-expectation bound similar to the one for the least square regression. In our analysis we just consider π ∈ H and βπβπΆ(π) ≤ 1. The case βπβπΆ(π) ≤ π for π ≥ 1 would be interesting in the future work. The approximation error involving π can be estimated by the knowledge of interpolation space. In our setting, the sample is drawn independently from the distribution π. However, in many practical problems, the i.i.d condition is a little demanding, so it would be interesting to investigate the ERM scheme for quantile regression with nonidentical distributions [13, 14] or dependent sampling [15]. Acknowledgment This work described in this paper is supported by NSF of China under Grant 11001247 and 61170109. where π= 2 (π + 1) 4π (π + 1) − (2 − π) min {2 (π + 1) , ππ} Μ are constant independent of π and πΏ. and πΆ (42) References [1] I. Steinwart and A. Christman, How SVMs Can Estimate Quantile and the Median, vol. 20 of Advances in Neural Information Processing Systems, MIT Press, Cambridge, Mass, USA, 2008. 6 [2] D. H. Xiang, “Conditional quantiles with varying Gaussians,” Advances in Computational Mathematics, 2011. [3] D.-H. Xiang, T. Hu, and D.-X. Zhou, “Learning with varying insensitive loss,” Applied Mathematics Letters, vol. 24, no. 12, pp. 2107–2109, 2011. [4] D.-H. Xiang, T. Hu, and D.-X. Zhou, “Approximation analysis of learning algorithms for support vector regression and quantile regression,” Journal of Applied Mathematics, vol. 2012, Article ID 902139, 17 pages, 2012. [5] N. Aronszajn, “Theory of reproducing kernels,” Transactions of the American Mathematical Society, vol. 68, pp. 337–404, 1950. [6] F. Cucker and D.-X. Zhou, Learning Theory: An Approximation Theory Viewpoint, vol. 24, Cambridge University Press, Cambridge, UK, 2007. [7] D.-X. Zhou, “The covering number in learning theory,” Journal of Complexity, vol. 18, no. 3, pp. 739–767, 2002. [8] S. Mukherjee, Q. Wu, and D.-X. Zhou, “Learning gradients on manifolds,” Bernoulli, vol. 16, no. 1, pp. 181–207, 2010. [9] S. Smale and D.-X. Zhou, “Estimating the approximation error in learning theory,” Analysis and Applications, vol. 1, no. 1, pp. 17–41, 2003. [10] S. Mukherjee and Q. Wu, “Estimation of gradients and coordinate covariation in classification,” Journal of Machine Learning Research, vol. 7, pp. 2481–2514, 2006. [11] Y. Yao, “On complexity issues of online learning algorithms,” Institute of Electrical and Electronics Engineers, vol. 56, no. 12, pp. 6470–6481, 2010. [12] Y. Ying, “Convergence analysis of online algorithms,” Advances in Computational Mathematics, vol. 27, no. 3, pp. 273–291, 2007. [13] T. Hu and D.-X. Zhou, “Online learning with samples drawn from non-identical distributions,” Journal of Machine Learning Research, vol. 10, pp. 2873–2898, 2009. [14] S. Smale and D.-X. Zhou, “Online learning with Markov sampling,” Analysis and Applications, vol. 7, no. 1, pp. 87–113, 2009. [15] Z.-C. Guo and L. Shi, “Classification with non-i.i.d. sampling,” Mathematical and Computer Modelling, vol. 54, no. 5-6, pp. 1347–1364, 2011. Abstract and Applied Analysis Advances in Operations Research Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Advances in Decision Sciences Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Mathematical Problems in Engineering Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Journal of Algebra Hindawi Publishing Corporation http://www.hindawi.com Probability and Statistics Volume 2014 The Scientific World Journal Hindawi Publishing Corporation http://www.hindawi.com Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 International Journal of Differential Equations Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Volume 2014 Submit your manuscripts at http://www.hindawi.com International Journal of Advances in Combinatorics Hindawi Publishing Corporation http://www.hindawi.com Mathematical Physics Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Journal of Complex Analysis Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 International Journal of Mathematics and Mathematical Sciences Journal of Hindawi Publishing Corporation http://www.hindawi.com Stochastic Analysis Abstract and Applied Analysis Hindawi Publishing Corporation http://www.hindawi.com Hindawi Publishing Corporation http://www.hindawi.com International Journal of Mathematics Volume 2014 Volume 2014 Discrete Dynamics in Nature and Society Volume 2014 Volume 2014 Journal of Journal of Discrete Mathematics Journal of Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Applied Mathematics Journal of Function Spaces Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Optimization Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Volume 2014