ebook img

Sharp moment and exponential tail estimates for U-statistics PDF

0.12 MB·
Save to my drive
Quick download
Download
Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.

Preview Sharp moment and exponential tail estimates for U-statistics

SHARP MOMENT AND EXPONENTIAL TAIL ESTIMATES FOR U-STATISTICS E. Ostrovsky, L.Sirota. 6 Department of Mathematics, Bar-Ilan University, Ramat-Gan, 59200,Israel. 1 0 e-mail: [email protected] 2 n Department of Mathematics, Bar-Ilan University, Ramat-Gan, 59200,Israel. a e-mail: [email protected] J 1 Abstract. 3 ] T We obtain in this paper a non-asymptotic non-improvable up to multiplicative S constant moment and exponential tail estimates for distribution for U statistics . − h by means of martingale representation. t a We show also the exactness of obtained estimations in one way or another by m providing appropriate examples. [ 1 Key words: U statistics, kernel, rank, random variables, Osekowski, Rosen- v − thal, Iensen, Tchebychev, and triangle inequalities, martingales and martingale dif- 5 7 ferences, martingale representation, Lebesgue-Riesz and Grand Lebesgue norm and 1 spaces, symmetric function, lower and upper estimates, moments, examples, natural 0 0 functions and norming, tails of distribution. . 2 Mathematics Subject Classification (2002): primary 60G17; secondary 60E07; 0 6 60G70. 1 : v i X r a 1 Introduction. Notations. Statement of prob- lem. Let (Ω,F,P) be a probabilistic space, which will be presumed sufficiently rich when we construct examples (counterexamples). Let ξ(i) , i = 1,2,...,n, be in- { } dependent identically distributed (i., i.d.) random variables (r.v.) with values in the certain measurable space (X,S), Φ = Φ(x(1),x(2),...,x(d)) be a symmetric mea- surable non-trivial numerical function (kernel) of d variables: Φ : Xd R, U(n) = → U = U(n,Φ,d) = n −1 n U(n,Φ,d; ξ(i) ) = Φ(ξ(i ),ξ(i ),...,ξ(i(d))), n > d (1.0) 1 2 { } d! I∈XI(d,n) be a so-called U statistic. Denote degΦ = d, − 1 Φ = Φ(ξ(1),ξ(2),...,ξ(d)), r = rankΦ [1,2,...,d 1], (1.1) ∈ − σ(n) = σ = Var(U(n)), T(Φ,x) := supT(U(n) EU(n))/σ(n),x), (1.2) n − n>d q i.e. the uniform tail function for our U statistics under natural norming. − Let also I = I(n) = I(d;n) = i ;i ;...;i be the set of indices of the form 1 2 d { } I(n) = I(d;n) = ~i = i = i ,i ,...,i such that1 i < i < i < i < i 1 2 d 1 2 3 d−1 d { } { } { } ≤ ≤ n; J = J(n) = J(d;n) be the set of indices of the form (subset of I(d;n)) J(d;n) = J(n) = ~j = j = j ;j ;...;j such that 1 j < j ... < j n 1. 1 2 d−1 1 2 d−1 { } { } { } ≤ ≤ − Recall that σ2(n) = Var(U(n)) n−r, n . ≍ → ∞ The martingale representation for the U statistics as well as the exact value − for its variance σ2(n) = Var(U(n)) may be found, e.g. in [14], [19], chapter 1. Namely, d d U(n) EU(n) = U , n,m − m! m=r X where −1 n U = ... g (ξ(i(1)),ξ(i(2)),...,ξ(i(m))), n,m i m! XX 1≤i(1)<i(X2)...<i(m)≤n g (x(1),x(2),...,x(m)) = i m d ... Φ(y(1),y(2),...,y(m)) (δ (dy(s)) P(dy(s))) P(dy(s)). x(s) − × ZX ZX ZX s=1 s=m+1 Y Y The sequence n n S (n) = U m n,m → m! relative the natural filtration F = σ ξ(1),ξ(2),...,ξ(k) , F = ,X k 0 { } {∅ } forme a martingale. Herewith 2 d d 2 n −1 σ2(n) = VarΦ = m=r m! · m! · X 2 d r! n−r VarΦ+O(n−r−1), n . r! → ∞ Note in addition that it follows from Iensen inequality g (ξ(i(1)),ξ(i(2)),...,ξ(i(m))) Φ . i p p | | ≤ | | Here and in the future for any r.v. η the function T (x) will be denote its η tail function: def T (x) = max(P(η > x),P(η < x)), x > 0. (1.3) η − We denote as usually the L(p) norm of the r.v. η as follows: η = [E η p]1/p, p 1; p | | | | ≥ and correspondingly def M(p) = M(p,Φ) = M(p,Φ, ξ(i) ) = sup (U(n) EU(n))/σ(n) . (1.4) p { } n | − | We will derive the non-refined up to multiplicative constant moment and exponential tail estimations for distribution of normed U statistics, − indeed, to estimate the variables M(p,Φ, ξ(i) ) and T (x). U(n)/σ(n) { } Evidently, these estimates may be applied for building of a non-asymptotical confidence interval for unknown parameter by using the U statistics in the sta- − tistical estimation. There are many works about this problem; the next list is far from being com- plete: [3], [29], [12], [14], [19], [21], [27] etc.; see also reference therein. Notice that in the classical book [19] there are many examples of applying of the theory of U statistics. A new application, namely, in the modern adaptive − estimation in the non - parametrical statistics may be found in the article [2] and in the book [23], chapter 5, section 5.13. 2 Main result: moments estimation for U - statistics. It is reasonable to suppose that EΦ = 0, Var(Φ) (0, ); moreover, we can and ∈ ∞ will assume without loss of generality Var(Φ) = 1, as long as it is constant; and that 3 all the moments of r.v. Φ which are written below there exist; otherwise it nothing to prove. Theorem 2.1. Let EΦ = 0, VarΦ = 1, Φ < for some value p 2. p | | ∞ ≥ Then d U(n) p C(d,r) Φ , p 2. (2.1) p (cid:12)σ(n)(cid:12) ≤ ·"logp# ·| | ≥ (cid:12) (cid:12)p (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) Proof. 1. Previous result. The most recent (foregoing) results in this direction was obtained in [24]: U(n) pd C (d,r) Φ , p 2. (2.2) 0 p (cid:12)σ(n)(cid:12) ≤ ·"logp#·| | ≥ (cid:12) (cid:12)p (cid:12) (cid:12) Thus, we update(cid:12)a depe(cid:12)ndency on the degree p. See also [13]. (cid:12) (cid:12) 2. Outline of the proof. The inequality (2.2)wasobtainedin[24]bymeans of the so-called martingale representation for the U statistics, see [14], [19], chapters − 1,2; and using further the moment estimation for the centered homogeneous of the degree d polynomial martingales (ζ ,G ), see [24], of the form n n ζ pd n sup C (d) , (2.3) n (cid:12)(cid:12) Var(ζn)(cid:12)(cid:12) ≤ 2 logp (cid:12) (cid:12)p (cid:12)q (cid:12) (cid:12) (cid:12) if of course ζn p < . (cid:12) (cid:12) | | ∞ For the multiply series in rearrangement invariant spaces analogous result was obtained by S.V.Astashkin in [1]. More information about martingale inequalities may be found in the many works of D.L.Burkholder, see e.g. the articles [5]-[7]. A famous survey on the martingale inequalities belongs to G.Peshkir and A.N.Shirjaev [30]. But in the recent publication about martingales [27] relaying in turn on the famous result belonging to A.Osekowski [22] the estimate(2.4) was improved: d ζ p n sup C (d) . (2.4) n (cid:12)(cid:12) Var(ζn)(cid:12)(cid:12) ≤ 3 "logp# (cid:12) (cid:12)p (cid:12)q (cid:12) (cid:12) (cid:12) More exactly, the followi(cid:12)ng importa(cid:12)nt function was introduced by A.Osekowski (up to factor 2) in the article [22]: p 1/p p Os(p) d=ef 4 √2 +1 1+ , p 4; (2.5) ·(cid:18)4 (cid:19) · ln(p/2)! ≥ the case p [2,4) is simple and may be considered separately. ∈ Note that 4 Os(p) def K = K = sup 15.7858, (2.6) Os p≥4 "p/lnp# ≈ is the so-called Osekowski’s constant. Let us define the following numerical sequence γ(d), d = 1,2,... : γ(1) := K = K, (initial condition) and by the following recursion Os 1 d γ(d+1) = γ(d) K 1+ . (2.7) Os · · d (cid:18) (cid:19) Since 1 d 1+ e, d ≤ (cid:18) (cid:19) we conclude γ(d) Kd ed−1, d = 1,2,.... (2.8) ≤ Os · It is proved in [27] in particular that the ”constant” C (d) in (2.4) allows the 3 following simple estimate: C (d) γ(d). 3 ≤ The inequality (2.8) represents nothing more than d dimensional and mar- − tingale generalization of a classical Rosenthal’s inequality for sums of independent random variables, [31], see also [18], the exact values of constants in the Rosenthal’s inequality see in [26]. We apply further the more modern estimate (2.5) instead (2.4) into the consid- erations of the report [24], we obtain what is desired. 3. Some details. Let as before EΦ = 0, VarΦ = 1, Φ < for some value p | | ∞ p 2. Let also the sequence γ(d) be defined in (2.7) (and in (2.8)). Then ≥ d d n −1/2 p m U(n) γ(m) Φ , p 2. (2.9) | |p ≤ · m!· m! · lnp! ·| |p ≥ m=r X and in turn after evident simplification d p U(n) C(d,r) n−r/2 Φ , p 2. (2.10) p p | | ≤ "lnp# | | ≥ Proof. We can write using the martingale representation for U statistics − d d n −1 U(n) = µ , (2.11) j mX=r m! m! j∈JX(m,n) where (µ ,F ) is certain centered martingale, k k n cardJ(m,n) = , Varµ = VarΦ = 1. j m! 5 Denote for brevity n N = N(m,n) = , m! then d d U(n) = N−1/2(m,n) ζ(m,n). (2.12) m=r m! X We apply the triangle inequality for the L(p) norm: d d U(n) N−1/2(m,n) ζ(m,n) . (2.13) p p | | ≤ m=r m! | | X Each term ζ(m,n) is the centered polynomial martingale of degree m generated by the function Φ. One can apply the Osekowski’s inequality (2.4): m p ζ(m,n) γ(m) Φ . (2.14) p p | | ≤ "lnp# | | It remains to substitute into (2.13). 4. The assertion of theorem 2.1 follows immediately from the estimate (2.9), but this estimate gives us certain numerical estimate for the value U . n p | | Remark 2.1. As long as U U = σ , we deduce that every time when n p n 2 n | | ≥ | | Φ < , p | | ∞ U(n) n−r/2 σ(n), n . (2.15) p | | ≍ ≍ → ∞ We generalize further this relation on more general than L (Ω) spaces. p Let us discuss now the lower bounds for the theorem 2.1. To be more precise, we denote M(p,Φ, ξ(i) ) def K (p) = sup sup sup { } . (2.16) d {ξ(i)}06=Φ∈Lp n ( n−r/2 |Φ|p ) Theorem 2.2. We conclude taking into account formulated above our definition and restrictions d p K (p) , p 2. (2.17) d ≍ "lnp# ≥ Proof. It remains to prove only the lower bound for the value K (p). d The relation (2.12) has been conjectured (hypothesis) in the article [24], page 19 for the polynomial martingales. It was proved (before!) for the polynomial 6 martingales from appropriate independent random variables in [13], [18]. The case of arbitrary polynomial martingales was grounded in authors report [27]. We represent here a very simple example in order to obtain the bottom border for K (p) exactly for the U statistics still for arbitrary value d = 1,2,.... Let d − n = 1 and let a r.v. η has a standard Poisson distribution with unit parameter P(η = k) = e−1/k!, k = 0,1,2,... and define ξ = η 1; then ξ is centered, Var(ξ) = 1 and it is no hard to calculate − p ξ , p . p | | ∼ e lnp → ∞ · Let also ξ(i) be independent copies of ξ. Then d p d ξ(i) e−d , p . (2.18) p | | ∼ "lnp# → ∞ i=1 Y Therefore d p lim K (p) : e−d. (2.19) p→∞ d "lnp #  ≥     Thus, obtained in this section estimate (2.1) of theorem (2.1) is essentially non- improvable relative the parameter p for all the values of dimension d, of course, up to multiplicative constant. 3 Estimations of U-statistics in the Grand Lebesgue Spaces and in the exponential Orlicz space norms. Let ψ = ψ(p), p [2,b), b = const, 1 < b (or p [1,b]) be certain bounded ∈ ≤ ∞ ∈ from below: infψ(p) > 1 continuous inside the semi-open interval [2,b) numerical function. We can and will suppose b = sup p, ψ(p) < , { ∞} so that suppψ = [2,b) or suppψ = [2,b]. The set of all such a functions will be denoted by Ψ(b); Ψ := Ψ( ). ∞ For each such a function ψ Ψ(b) we define ∈ d p def ψ (p) = ψ(p). (3.1) d "lnp# · Evidently, ψ ( ) Ψ(b). d · ∈ 7 By definition, the (Banach) space Gψ = Gψ(b) consists on all the numerical valued random variables ζ defined on our probability space (Ω,F,P) and having { } a finite norm ζ def p ζ Gψ = sup | | < . (3.2) || || p∈(1,b)" ψ(p) # ∞ These spaces are suitable in particular for an investigation of the random variables and the random processes (fields) with exponential decreasing tails of distributions, the Central Limit Theorem inBanach spaces, study ofPartial Differential Equations etc., see e.g. [17], [4], [20], [23], chapter 1, [9]-[11], [15]-[16] etc. More detail, suppose 0 < ζ := ζ Gψ < . Define the function || || || || ∞ ν(p) = ν (p) = plnψ(p), 2 p < b ψ ≤ and put formally ν(p) := , p < 2 or p > b. Recall that the Young-Fenchel, ∞ or Legendre transform f∗(y) for arbitrary function f : R R is defined (in the → one-dimensional case) as follows f∗(y) d=ef sup(xy f(x)). x − It is known that T (y) exp ν∗(ln(y/ ζ )) , y > e ζ . (3.3) ζ ≤ − ψ || || ·|| || (cid:16) (cid:17) Conversely, if (3.3) there holds in the following version: T (y) exp ν∗(ln(y/K)) , y > e K, K = const > 0, (3.4) ζ ≤ − ψ · (cid:16) (cid:17) and the function ν (p), 2 p < is positive, continuous, convex and such that ζ ≤ ∞ lim ψ(p) = , p→∞ ∞ then ζ Gψ and besides ∈ ζ Gψ C(ψ) K. (3.5) || || ≤ · Moreover, let us introduce the exponential Orlicz space L(M) over the source probability space (Ω,F,P) with proper Young-Orlicz function M(u) := exp ν∗(ln u ) , u > e ψ | | | | (cid:16) (cid:17) or correspondingly M (u) := exp ν∗ (ln u ) , u > e d ψd | | | | (cid:16) (cid:17) and as ordinary M(u) = M (u) = exp(C u2) 1, u e. It is known [28] that the d − | | ≤ Gψ norm of arbitrary r.v. ζ is complete equivalent to the its norm in Orlicz space L(M) : 8 ζ Gψ C ζ L(M) C ζ Gψ, 1 C C < ; 1 2 1 2 || || ≤ || || ≤ || || ≤ ≤ ∞ ζ Gψ C ζ L(Md) C ζ Gψ , 1 C C < . d 3 4 d 3 4 || || ≤ || || ≤ || || ≤ ≤ ∞ Example 3.1. The estimate for the r.v. ξ of a form ξ C p1/m lnrp, p 2, p 1 | | ≤ ≥ where C = const > 0, m = const > 0, r = const, is quite equivalent to the 1 following tail estimate T (x) exp C (C ,m,r) xm log−mrx , x > e. ξ 2 1 ≤ − n o It is important to note that the inequality (3.4) may be applied still when the r.v. ξ does not have the exponential moment, i.e. does not satisfy the famous Kramer’s condition. Namely, let us consider next example. Example 3.2. Define the following Ψ function. − ψ (p) := exp C pβ , p [2, ), β = const > 0. [β] 3 ∈ ∞ (cid:16) (cid:17) The r.v. ξ belongs to the space Gψ if and only if [β] T (x) exp C (C ,β) [ln(1+x)]1+1/β , x 0. ξ 4 3 ≤ − ≥ (cid:16) (cid:17) See also [24]. Let us return to the source problem. Assume that there exists certain function ψ( ) Ψ(b), b = const (2, ) such that Φ Gψ(b). For instance, this function · ∈ ∈ ∞ ∈ may be picked by the following natural way: ψ (p) := Φ , (3.6) Φ p | | ifofcoursethereexistsandisfiniteatlastforsomevaluepgreatestthan2, obviously, with the correspondent value b. Theorem 3.1. We propose under formulated above conditions sup U(n)/σ(n) Gψ C(ψ) Φ Gψ. (3.7) d n || || ≤ || || Proof is very simple. Assume Φ Gψ (0, ). It follows immediately from the || || ∈ ∞ direct definition of the norm in the Grand Lebesgue Spaces Φ Φ Gψ ψ(p). p | | ≤ || || · We apply the inequality (2.1) of theorem 2.1 for the values p from the set p [2,b) : ∈ 9 d p sup U(n)/σ(n) C(d,r) Φ p p n | | ≤ ·"logp# ·| | ≤ d p C(d,r) Φ Gψ ψ(p) = ·"logp# ·|| || · C(d,r) Φ Gψ ψ (p), d ·|| || · or equally sup U(n)/σ(n) Gψ C(d,r) Φ Gψ. d n || || ≤ ·|| || Example 3.3. Suppose T (x) exp C xm log−mrx , x > e, (3.8) Φ 5 ≤ − n o where C = const > 0, m = const > 0, r = const. As we know, 5 Φ C (C ,m,r) p1/m lnrp, p 2. p 6 5 | | ≤ ≥ We use theorem 3.1 sup U(n)/σ(n) C pd+1/m (lnp)r−d,p 2, p 7 n | | ≤ ≥ and we conclude returning to the tail of distribution supT (x) exp C xm/(1+dm) log−m(r−d)/(1+dm)x , x e. (3.9) U(n)/σ(n) 8 n ≤ − ≥ n o Thus, we obtained in this way the exponential bounds for distribution for the normed U statistics, expressed only through the very simple source data (3.8). − But the authors are not convinced of the finality of these estimations in the considered case; cf., e.g. [3], [12], [19], chapter 2, [29]. Example 3.4. Assume now T (x) exp C [ln(1+x)]1+1/β , x 0. (3.10) Φ 9 ≤ − ≥ (cid:16) (cid:17) or more strictly exp C˜ [ln(1+x)]1+1/β T (x) 9 Φ − ≤ ≤ (cid:16) (cid:17) exp C [ln(1+x)]1+1/β , x 0,0 < C C˜ < . (3.10a) 9 9 9 − ≥ ≤ ∞ (cid:16) (cid:17) Then Φ exp C pβ , p 2, p 10 | | ≤ ≥ (cid:16) (cid:17) 10

See more

The list of books you might like

Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.