Distributed Trimmed Hill Estimator

Abstract

Proceeded from trimmed Hill estimators and distributed inference, a new distributed version of trimmed Hill estimator for heavy tail index is proposed. Considering the case where the number of observations involved in each machine can be either the same or different and either fixed or varying to the total sample size, its consistency and asymptotic normality are discussed. Simulation studies are particularized to show the new estimator performs almost in line with the trimmed Hill estimator.

Share and Cite:

Guo, T. (2023) Distributed Trimmed Hill Estimator. Journal of Applied Mathematics and Physics, 11, 4000-4015. doi: 10.4236/jamp.2023.1112256.

1. Introduction

Let { X 1 , X 2 , , X n } be independent and identically distributed (i.i.d.) random variables drawn from F, a distributed function which belongs to the max-domain of attraction of an extreme value distribution G γ with extreme value index γ . It is well-known that if F D ( G γ ) , mathematically, there exist constants a n > 0 and b n such that

l i m n F n ( a n x + b n ) = G γ ( x ) : = exp { ( 1 + γ x ) 1 / γ } , (1)

for all 1 + γ x > 0 . For a tail index γ > 0 , the limit relation (1) is equivalent to

lim t U ( t x ) U ( t ) = x γ , x > 0 (2)

where U = { 1 / ( 1 F ) } is the left-continuous inverse function of 1 / ( 1 F ) and a regular varying function with γ , see [1] .

The estimation of tail index for heavy-tailed distributions may be the one of the most studied problems in the extreme value theory. Since the numerous works of this aspect such as the Hill estimator, the Pickands estimator and the maximum likelihood estimator have already been explored referring to [1] [2] [3] [4] and [5] for detailed discussions and reviews.

This extreme value analysis often rely on high order statistics. However, in many applications, one may face the challenges when it quickly run out of data since the observations can be corrupted and this contamination can lead to severe bias in the estimation of the tail index. Considering the problem, plenty of researchers did a lot of work. Based on the classic Hill eatimator of γ :

γ ^ k ( n ) : = 1 k i = 1 k log ( X ( n i + 1 , n ) X ( n k , n ) ) , 1 k n 1 ,

where X 1, n X n , n are the associated order statistics of { X 1 , X 2 , , X n } i.i.d. random variables with unknown distribution F D ( G γ ) with γ > 0 , [6] trimmed a certain number of the largest order statistics in order to obtain a robust estimator of γ and (among other robust estimators) defined a trimmed version of the Hill estimator:

γ ^ k 0 , k t r i m ( n ) : = i = k 0 + 1 k c k 0 , k ( i ) log ( X ( n i + 1 , n ) X ( n k , n ) ) , 0 k 0 < k < n .

[7] then chose the weights c k 0 , k ( i ) so that the estimator is asymptotically optimal where,

c k 0 , k ( i ) = { k 0 + 1 k k 0 , i = k 0 + 1 1 k k 0 , i = k 0 + 2 , , k

and they also found the method for the trimming parameter which yields the trimmed Hill estimator that can adapt to the unknown level of contamination in the extremes. While removing the lower order statistics from the classical Hill estimator, [8] derived an alternative estimator of the tail index and it was shown to have lower variance than the classic Hill estimator. A number of reseachers also considered trimming but of the models rather than the data, see [9] and [10] . Moreover, the random censoring for heavy-tailed distribution was discussed in [11] [12] [13] and [14] . Contrary to the above, here we assume to have non-truncated heavy-tailed model and only the top order statistics are contaminated in the associated data.

The rapid emergence of massive datasets in various fields becomes more and more challenging to traditional statistical methods. Account of that, distributed inference theory which refers to analyzing data stored in distributed machines has been proposed. It is developed to deal with large-scale statistical optimization problems and requires a divide-and-conquer algorithm which estimates a desired quantity or parameter on each machine and transmits the results to a central machine often by simple averaging. With the conditions of [15] [16] and [17] [18] reported on a first attempt in distributed inference for extreme value index and proposed a distributed Hill estimator and establish its asymptotic theories.

In this paper, considering the massive datasets contaminated of the top order statistics we apply the method of distributed inference and then derive a new estimator of the extreme value index for heavy-tailed distributions. The new estimator can be used for the situation when large datasets are distributedly stored and cannot be combined into one oracle sample and the top order statistics are corrupted.

We assume that the i.i.d. observations { X 1 , X 2 , , X n } are stored in k machines with m observations each, i.e., n = m k and the operation mechanism of each machine is independent. Let M j ( 1 ) M j ( m ) denote the order statistics within the machine j. Suppose we have identified that the top d0 order statistics have been corrupted in each machine, we use the top d j exceedance radios M j ( i ) / M j ( d j ) and cut the same level of top d0 exceedance radios for i = 1, , d j and 0 < d 0 < d j < m to build the estimator in each machine. Then we take the average of the estimators from all machines and the distributed trimmed Hill estimator is defined as:

γ ^ D H t r i m = 1 k i = 1 k ( d 0 + 1 d j d 0 log M j ( d 0 + 1 ) M j ( d j + 1 ) + 1 d j d 0 i = d 0 + 2 d j log M j ( i ) M j ( d j + 1 ) ) . (3)

To derive the asymptotic normality of distributed trimmed Hill estimator γ ^ D H t r i m , we need impose the following condition on the sequences k and m,

m ( n ) , k ( n ) and m / log k as n . (4)

And we need the second order regular varying condition as follows: there exists an positive or negative function A with lim t A ( t ) = 0 and a real number ρ 0 , satisfying | A ( t ) | R V ρ , such that

l i m t U ( t x ) U ( t ) x γ A ( t ) = x γ x ρ 1 ρ (5)

for all x > 0 (see e.g. [1] , Corollary 2.3.4).

By (5), we have that there exists a function A 0 ( t ) such that A 0 ( t ) ~ A ( t ) as t , and for all ε > 0 , δ > 0 , there is a t 0 ( ε , δ ) > 0 such that for t x t 0 , t t 0 ,

| log U ( t x ) log U ( t ) γ log x A 0 ( t ) x ρ 1 ρ | ε x ρ x ± δ , where x ± δ = max { x δ , x δ } . (6)

By the adoption of (5) on the function L, we also have,

| log L ( t x ) log L ( t ) A 0 ( t ) x ρ 1 ρ | ε x ρ x ± δ , where x ± δ = max { x δ , x δ } , (7)

for more details on A 0 , see page 48 in [1] .

2. Main Results

In the homogeneous case where d 1 = = d k = d is a fixed integer, the following theorem shows the asymptotic normality of the distributed trimmed Hill estimator.

Theorem 2.1. Suppose F D ( G γ ) with γ > 0 and (4) and (5) hold. Let d 1 = = d k = d , where d > d 0 > 0 is a fixed integer. If [ k ( d d 0 ) ] 1 2 A ( m / d ) = O ( 1 ) , as n , then

[ k ( d d 0 ) ] 1 2 { γ ^ D H t r i m A ( m / d ) B ( m , d , d 0 , ρ ) } d N ( 0, γ 2 ) ,

where

B ( m , d , d 0 , ρ ) = 1 ρ ( d d 0 ) ( m d ) ρ Γ ( m + 1 ) Γ ( d ρ + 1 ) Γ ( d + 1 ) Γ ( m ρ + 1 ) { d 0 ( d d ρ ) d d 0 i = 1 d 0 ( d d ρ ) d i + 1 + d ρ 1 ρ }

with ρ < 0 and B ( m , d , d 0 , ρ ) = 1 with ρ = 0 .

In the heterogeneous case where { d j } j = 1 k are uniformly bounded positive integer series, i.e., sup j d j = d max < , the following theorem shows the asymptotic normality of the distributed trimmed Hill estimator.

Theorem 2.2. Suppose F D ( G γ ) with γ > 0 and (4) and (5) hold. Let { d j } j = 1 k be uniformly bounded positive integer series, i.e., sup j d j = d max < and 0 < d 0 < inf j d j = d min . If [ k ( d ¯ d 0 ) ] 1 2 A ( m / d ¯ ) = O ( 1 ) , as n , then

[ k ( d ¯ d 0 ) ] 1 2 { γ ^ D H t r i m γ A ( m / d ¯ ) k 1 j = 1 k ( d ¯ / d j ) ρ B ( m , d j , d 0 , ρ ) } d N ( 0 , γ 2 ) ,

where d ¯ = k 1 j = 1 k d j .

In the homogeneous case where d 1 = = d k = d and d = d ( m ) is an intermediate sequence, i.e., d = d ( m ) , d / m 0 as n , the following theorem shows the asymptotic normality of the distributed trimmed Hill estimator.

Theorem 2.3. Suppose F D ( G γ ) with γ > 0 , and (4) and (5) hold. Let d 1 = = d k = d , where d = d ( m ) and d / m 0 as n . If [ k ( d d 0 ) ] 1 / 2 A ( m / d ) = O ( 1 ) , as n , then

[ k ( d d 0 ) ] 1 2 { γ ^ D H t r i m γ A ( m / d ) H ( d , m , ρ ) } d N ( 0, γ 2 ) ,

where

H ( d , m , ρ ) = 1 1 ρ ( m d ) ρ Γ ( m + 1 ) Γ ( d ρ + 1 ) Γ ( d + 1 ) Γ ( m ρ + 1 ) d d d 0 .

3. Simulation Studies

In this section, we study the finite sample performance of the distributed trimmed Hill estimator γ ^ D H t r i m and compare it with [7] ’s estimator, i.e., the trimmed Hill estimator on the following three distributions which all belong to the max-domain of attraction of an extreme value distribution for varying parameters with three sub-cases for each distribution.

We obtain the mean value and mean squared error (MSE) for r = 2000 Monte Carlo simulations of all considered estimators of heavy-tailed models with sample size n = 10,000. We assume the contamination occurs in the top d0 order statictics in each machine and vary the level of d in the distributed trimmed Hill estimator to verify the theoretical results on the property we give in Section2 and to compare the finite sample performance of the distributed trimmed Hill estimator with that of the trimmed Hill estimator for different values of d. The sample { X 1 , X 2 , , X n } contains n = 10,000 observations stored in k machines with m observations each. We fix k = 20 and m = 500 and vary d from 30 to 100 with d0 = 8.The results are presented in Figures 1-3.

• The Fréchet distribution with distribution function

F ( x ) = exp ( x α ) , α > 0 ,

which implies γ = 1 / α and ρ = 1 . We consider the three parameters α = 1, 0.5 and 2.

Figure 1. Fréchet distribution, parameters α = 1, 0.5 and 2. Diagnostics of trimmed Hill estimator (coral) and distributed trimmed Hill estimator(skyblue) as a function of d.

Figure 2. Pareto (σ, ξ) distribution, sets of parameters σ = 1, ξ = 1; σ = 2, ξ = 0.5 and σ = 1, ξ = 2.5. Diagnostics of trimmed Hill estimator(coral) and distributed trimmed Hill estimator (skyblue) as a function of d.

• The Pareto(σ, ξ) distribution with distribution function

F ( x ) = 1 ( x σ ) 1 / ξ , x σ , ξ > 0 ,

which implies γ = ξ and ρ = ξ . We consider the three sets of parameters σ = 1, ξ = 1; σ = 2, ξ = 0.5 and σ = 1, ξ = 2.5.

• The Burr(τ, λ) distribution with distribution function

F ( x ) = 1 ( 1 + x τ ) λ , x > 0 , τ , λ > 0 ,

which implies γ = 1 τ λ and ρ = 1 λ .We consider the three sets of parameters τ = 2, λ = 0.5; τ = 3, λ = 0.5 and τ = 3, λ = 1.

For the Fréchet distribution, Figure 1 shows that as d increases, the MSE increases for the estimators with different α. For the Pareto distribution in Figure 2, the bias between the estimators and the true value is virtually zero for all levels of d. For the Burr distribution in Figure 3, we observe a trade off for the estimators with different sets of parameters: as d increases, the MSE increases when λ is low while the MSE decreases when λ takes a larger value.

Figure 3. Burr (τ, λ) distribution, sets of parameters τ = 2, λ = 0.5; τ = 3, λ = 0.5 and τ = 3, λ = 1. Diagnostics of trimmed Hill estimator (coral) and distributed trimmed Hill estimator (skyblue) as a function of d.

Figures 1-3 show that the difference in MSE between the distributed trimmed Hill estimator and the trimmed Hill estimator is not sizeable. Consequently, we can infer that when dealing with the estimation problem of extreme value index with massive and corrupted datasets the new estimator we derive performs well.

4. Proof

Recall that 1 / ( 1 F ) , { X 1 , X 2 , , X m } = d { U ( Z 1 ) , U ( Z 2 ) , , U ( Z m ) } , where { Z 1 , Z 2 , , Z n } is a random sample of Z with the distribution function 1 1 / z , z 1 . For each machine j, let Z j ( 1 ) Z j ( m ) denote the order statistics of the m Pareto (1) distributed variables corresponding to the m observations in this machine. Notting that { M j ( 1 ) , M j ( 2 ) , , M j ( m ) } = d { U ( Z j ( 1 ) ) , U ( Z j ( 2 ) ) , , U ( Z j ( m ) ) } , we have

γ ^ D H j t r i m = d d 0 + 1 d j d 0 log U ( Z j ( d 0 + 1 ) ) U ( Z j ( d j + 1 ) ) + 1 d j d 0 i = d 0 + 2 d j log U ( Z j ( i ) ) U ( Z j ( d j + 1 ) ) ,

and then

γ ^ D H t r i m = d 1 k j = 1 k ( d 0 + 1 d j d 0 log U ( Z j ( d 0 + 1 ) ) U ( Z j ( d j + 1 ) ) + 1 d j d 0 i = d 0 + 2 d j log U ( Z j ( i ) ) U ( Z j ( d j + 1 ) ) ) .

(2) implies that U R V γ , then U ( x ) = x γ L ( x ) , where L ( x ) is slowly varying function. Hence U ( Z j ( i ) ) = ( Z j ( i ) ) γ L ( Z j ( i ) ) and

γ ^ D H t r i m = d γ k j = 1 k ( d 0 + 1 d j d 0 log Z j ( d 0 + 1 ) Z j ( d j + 1 ) + 1 d j d 0 i = d 0 + 2 d j log Z j ( i ) Z j ( d j + 1 ) ) + 1 k j = 1 k ( d 0 + 1 d j d 0 log L ( Z j ( d 0 + 1 ) ) L ( Z j ( d j + 1 ) ) + 1 d j d 0 i = d 0 + 2 d j log L ( Z j ( i ) ) L ( Z j ( d j + 1 ) ) ) .

Lemma 4.1. Suppose F D ( G γ ) with γ > 0 , define

γ j * ( d 0 ) : = d 0 + 1 d j d 0 log Z j ( d 0 + 1 ) Z j ( d j + 1 ) + 1 d j d 0 i = d 0 + 2 d j log Z j ( i ) Z j ( d j + 1 ) ,

for j = 1, , k , and γ D H t r i m : = γ k j = 1 k γ j ( d 0 ) . Under the assumption of (4), γ D H t r i m P γ .

Proof. Note that { log Z i , i = 1 , 2 , , m } forms a random sample from the standard exponential distribution. In the machine j, for any i { d 0 + 1, , d j } by Rényi’s representation we have

log Z j ( i ) log Z j ( d j + 1 ) = d q = 1 d j i + 1 E j , q / ( d j q + 1 ) = d Y j ( i ) with E j ,1 , , E j , d j i.i.d. standard exponential, where Y j ( 1 ) Y j ( 2 ) Y i ( d ) are the order statistics of Exp (1) corresponding to the d j obeservation.

The joint distribution of γ j * ( d 0 ) , 0 d 0 d j 1 , can be expressed as follows:

{ γ j * ( d 0 ) } d 0 = 0 d j 1 = d { d 0 + 1 d j d 0 Y j ( d 0 + 1 ) + 1 d j d 0 i = d 0 + 2 d j Y j ( i ) } d 0 = 0 d j 1 = d { d 0 + 1 d j d 0 q = 1 d j d 0 E j , q d j q + 1 + 1 d j d 0 i = d 0 + 2 d j q = 1 d j i + 1 E j , q d j q + 1 } d 0 = 0 d j 1 = { d 0 + 1 d j d 0 q = 1 d j d 0 E j , q d j q + 1 + 1 d j d 0 q = 1 d j d 0 1 i = d 0 + 2 d j q + 1 E j , q d j q + 1 } d 0 = 0 d j 1 = { q = 1 d j d 0 1 ( d 0 + 1 d j d 0 + i = d 0 + 2 d j q + 1 1 d j d 0 ) E j , q d j q + 1 + E j , d j d 0 d j d 0 } d 0 = 0 d j 1 = { q = 1 d j d 0 E j , q d j d 0 } d 0 = 0 d j 1 (8)

By (8) it implies that E γ j * ( d 0 ) = 1 for d 0 = 0, , d j 1 , and by WLLN and (4) it follows that γ D H t r i m P γ as n .

Lemma 4.2. Under the condition of Theorem 3 and define

R d 0 , d : = 1 k j = 1 k ( d 0 + 1 d d 0 log L ( Z j ( d 0 + 1 ) ) L ( Z j ( d j + 1 ) ) + 1 d d 0 i = d 0 + 2 d log L ( Z j ( i ) ) L ( Z j ( d j + 1 ) ) ) ,

if h ( d ) = o ( d ) , as n , we have

[ k ( d d 0 ) ] 1 2 max 0 d 0 < h ( d ) | R d 0 , d A 0 ( m / d ) 1 ρ d d d 0 | P 0.

Proof. Note that

[ k ( d d 0 ) ] 1 2 max 0 d 0 < h ( d ) | R d 0 , d A 0 ( m / d ) 1 ρ d d d 0 | [ k ( d d 0 ) ] 1 2 max 0 d 0 < h ( d ) | R d 0 , d S d 0 , d | + [ k ( d d 0 ) ] 1 2 max 0 d 0 < h ( d ) | S d 0 , d A 0 ( m / d ) 1 ρ d d d 0 | , (9)

where

S d 0 , d = 1 k j = 1 k { A 0 ( Z j ( d + 1 ) ) d d 0 [ ( d 0 + 1 ) ( Z j ( d 0 + 1 ) / Z j ( d + 1 ) ) ρ 1 ρ + i = d 0 + 2 d ( Z j ( i ) / Z j ( d + 1 ) ) ρ 1 ρ ] } .

In the first term of (9), we have that

[ k ( d d 0 ) ] 1 2 max 0 d 0 < h ( d ) | R d 0 , d S d 0 , d | = [ k ( d d 0 ) ] 1 2 max 0 d 0 < h ( d ) d A 0 ( m / d ) d d 0 ( d d 0 d A 0 ( m / d ) | R d 0 , d S d 0 , d | ) [ k ( d d 0 ) ] 1 2 A 0 ( m / d ) 1 h ( d ) d max 0 d 0 < h ( d ) ( d d 0 d A 0 ( m / d ) | R d 0 , d S d 0 , d | ) . (10)

By the assumption of [ k ( d d 0 ) ] 1 / 2 A ( m / d ) = O ( 1 ) as n and A 0 ( t ) ~ A ( t ) as t , we can get that as n , [ k ( d d 0 ) ] 1 / 2 A 0 ( m / d ) = O ( 1 ) . By (7) choose ρ + δ < 0 and we can get that

max 0 d 0 < h ( d ) ( d d 0 d A 0 ( m / d ) | R d 0 , d S d 0 , d | ) max 0 d 0 < h ( d ) d 0 ε d k j = 1 k A 0 ( Z j ( d + 1 ) ) A 0 ( m / d ) ( Z j ( d 0 + 1 ) Z j ( d + 1 ) ) ρ + δ + max 0 d 0 < h ( d ) ε d k j = 1 k A 0 ( Z j ( d + 1 ) ) A 0 ( m / d ) i = d 0 + 1 d ( Z j ( i ) Z j ( d + 1 ) ) ρ + δ . (11)

Since Z j ( d 0 + 1 ) / Z j ( d + 1 ) is the ( d 0 + 1 ) t h order statistic from the standard Pareto distribution, [19] implies that d 0 Z j ( d 0 + 1 ) / ( d + 1 ) Z j ( d + 1 ) P 1 . Recall that A R V ρ and we have Z j ( d + 1 ) / ( m / d ) P 1 as n . Combining with A 0 ( t ) ~ A ( t ) as t , we can get that as n , A 0 ( Z j ( d + 1 ) ) / A 0 ( m / d ) P 1 and then

max 0 d 0 < h ( d ) d 0 ε d k j = 1 k A 0 ( Z j ( d + 1 ) ) A 0 ( m / d ) ( d + 1 d 0 ) ρ + δ ( d 0 Z j ( d 0 + 1 ) ( d + 1 ) Z j ( d + 1 ) ) ρ + δ P 0. (12)

Similarly, as n ,

max 0 d 0 < h ( d ) ε d k j = 1 k A 0 ( Z j ( d + 1 ) ) A 0 ( m / d ) i = d 0 + 1 d ( d + 1 i 1 ) ρ + δ ( ( i 1 ) Z j ( i ) ( d + 1 ) Z j ( d + 1 ) ) ρ + δ P 0. (13)

Combining with the (10) (11) (12) and (13), we can get that as n ,

[ k ( d d 0 ) ] 1 2 max 0 d 0 < h ( d ) | R d 0 , d S d 0 , d | P 0. (14)

In the second term of (9), we have that

[ k ( d d 0 ) ] 1 2 max 0 d 0 < h ( d ) | S d 0 , d A 0 ( m / d ) 1 ρ d d d 0 | [ k ( d d 0 ) ] 1 2 A 0 ( m / d ) 1 h ( d ) d max 0 d 0 < h ( d ) | d d 0 d S d 0 , d A 0 ( m / d ) 1 1 ρ | . (15)

It is known that

max 0 d 0 < h ( d ) | d d 0 d S d 0 , d A 0 ( m / d ) 1 1 ρ | max 0 d 0 < h ( d ) | d 0 d k j = 1 k A 0 ( Z j ( d + 1 ) ) A 0 ( m / d ) ( Z j ( d 0 + 1 ) / Z j ( d + 1 ) ) ρ 1 ρ | + max 0 d 0 < h ( d ) | 1 d k j = 1 k A 0 ( Z j ( d + 1 ) ) A 0 ( m / d ) i = d 0 + 1 d ( Z j ( i ) / Z j ( d + 1 ) ) ρ 1 ρ 1 1 ρ | , (16)

by WLLN for triangular array and Z j ( d + 1 ) is independent with Z j ( i ) / Z j ( d + 1 ) for i = d 0 + 1, , d and j = 1, , k , we have that n ,

max 0 d 0 < h ( d ) | d 0 d k j = 1 k A 0 ( Z j ( d + 1 ) ) A 0 ( m / d ) ( Z j ( d 0 + 1 ) / Z j ( d + 1 ) ) ρ 1 ρ | P 0 (17)

and

max 0 d 0 < h ( d ) | 1 d k j = 1 k A 0 ( Z j ( d + 1 ) ) A 0 ( m / d ) i = d 0 + 1 d ( Z j ( i ) / Z j ( d + 1 ) ) ρ 1 ρ 1 1 ρ | P 0. (18)

Combining with (17) and (18), it illustrates that n ,

max 0 d 0 < h ( d ) | d d 0 d S d 0 , d A 0 ( m / d ) 1 1 ρ | P 0

and finally we can get that n ,

[ k ( d d 0 ) ] 1 2 max 0 d 0 < h ( d ) | S d 0 , d A 0 ( m / d ) 1 ρ d d d 0 | P 0

which yield the Lemma.

Proof of Theorem 2.1. When ρ < 0 and by Lemma S.2 in [18] , we have that lim n { Z ( d + 1 ) > t 0 } = 1 , for any t 0 > 1 .Then by applying (6) twice with t = m / d and x = d Z j ( j ) / m , i = d 0 + 1 , , d + 1 and x = d Z j ( d + 1 ) / m we get that as n ,

log U ( Z j ( i ) ) log U ( Z j ( d + 1 ) ) γ ( log Z j ( i ) log Z j ( d + 1 ) ) A 0 ( m / d ) = ( d Z j ( j ) / m ) ρ 1 ρ ( d Z j ( d + 1 ) / m ) ρ 1 ρ + o p ( 1 ) { ( d Z j ( j ) / m ) ρ ± δ + ( d Z j ( d + 1 ) / m ) ρ ± δ } . (19)

Here, the o p ( 1 ) term is uniform for all 1 j k , d 0 + 1 i d + 1 and all k . We obtain that

[ k ( d d 0 ) ] 1 2 ( γ ^ D H t r i m γ ) : = I 1 + I 2 + I 3 ,

where

I 1 = γ { [ k ( d d 0 ) ] 1 2 [ 1 k j = 1 k ( d 0 + 1 d d 0 log Z j ( d 0 + 1 ) Z j ( d + 1 ) + i = d 0 + 2 d 1 d d 0 log Z j ( i ) Z j ( d + 1 ) ) 1 ] } ,

I 2 = [ k ( d d 0 ) ] 1 2 A 0 ( m / d ) ρ 1 k j = 1 k { ( d Z j ( d + 1 ) m ) ρ d 0 + 1 d d 0 ( ( Z j ( d 0 + 1 ) Z j ( d + 1 ) ) ρ 1 ) } + [ k ( d d 0 ) ] 1 2 A 0 ( m / d ) ρ 1 k j = 1 k { ( d Z j ( d + 1 ) m ) ρ i = d 0 + 2 d 1 d d 0 ( ( Z j ( i ) Z j ( d + 1 ) ) ρ 1 ) } ,

I 3 = o p ( 1 ) [ k ( d d 0 ) ] 1 2 A 0 ( m d ) 1 k j = 1 k { ( d Z j ( d + 1 ) m ) ρ ± δ d 0 + 1 d d 0 ( ( Z j ( d 0 + 1 ) Z j ( d + 1 ) ) ρ + δ + 1 ) } + o p ( 1 ) [ k ( d d 0 ) ] 1 2 A 0 ( m d ) 1 k j = 1 k { ( d Z j ( d + 1 ) m ) ρ ± δ i = d 0 + 2 d 1 d d 0 ( ( Z j ( i ) Z j ( d + 1 ) ) ρ + δ + 1 ) } .

By Lemma 4.1. and the central limit theorem, we have that I 1 d N ( 0, γ 2 ) , as n .

I 2 = [ k ( d d 0 ) ] 1 2 A 0 ( m / d ) ρ E { ( d Z 1 ( d + 1 ) m ) ρ } 1 k ( d d 0 ) j = 1 k ( d Z j ( d + 1 ) m ) ρ E { ( d Z 1 ( d + 1 ) m ) ρ } { d 0 ( ( Z j ( d 0 + 1 ) Z j ( d + 1 ) ) ρ 1 ) + i = d 0 + 1 d ( ( Z j ( i ) Z j ( d + 1 ) ) ρ 1 ) } .

By WLLN for triangular array and Z j ( d + 1 ) is independent with Z j ( i ) / Z j ( d + 1 ) for i = d 0 + 1, , d and j = 1, , k , we have that

1 k j = 1 k ( d Z j ( d + 1 ) m ) ρ E { ( d Z 1 ( d + 1 ) m ) ρ } { d 0 + 1 d d 0 ( ( Z j ( d 0 + 1 ) Z j ( d + 1 ) ) ρ 1 ) + i = d 0 + 2 d 1 d d 0 ( ( Z j ( i ) Z j ( d + 1 ) ) ρ 1 ) } P E { d 0 + 1 d d 0 ( ( Z j ( d 0 + 1 ) Z j ( d + 1 ) ) ρ 1 ) + i = d 0 + 2 d 1 d d 0 ( ( Z j ( i ) Z j ( d + 1 ) ) ρ 1 ) } = 1 d d 0 { d 0 ( d d ρ ) d d 0 i = 1 d 0 ( d d ρ ) d i + 1 + d ρ 1 ρ }

as n , where the second equality follows from a direct calculation. By the Stirling’s formula, it follows that,

E { ( d Z 1 ( d + 1 ) m ) ρ } = ( m d ) ρ Γ ( m + 1 ) Γ ( d ρ + 1 ) Γ ( d + 1 ) Γ ( m ρ + 1 ) d ρ Γ ( d ρ + 1 ) Γ ( d + 1 )

as m . Hence, combing with [ k ( d d 0 ) ] 1 / 2 A ( m / d ) = O ( 1 ) as n , we can replace A 0 by A and obtain that as n ,

I 2 = [ k ( d d 0 ) ] 1 2 A 0 ( m / d ) B ( m , d , d 0 , ρ ) { 1 + o p ( 1 ) } . (20)

Similarly, as for I 3 , we obtain that I 3 P 0 as n . Combining with I 1 d N ( 0, γ 2 ) and (20) as n the statement in Theorem 2.1 follows.

When ρ = 0 , (19) is equivalent to

log U ( Z j ( i ) ) log U ( Z j ( d + 1 ) ) γ ( log Z j ( i ) log Z j ( d + 1 ) ) A 0 ( m / d ) = log ( Z j ( i ) ) log ( Z j ( d + 1 ) ) + o p ( 1 ) { ( d Z j ( i ) / m ) ± δ + ( d Z j ( d + 1 ) / m ) ± δ } ,

as n , where o p ( 1 ) term is uniform for all 1 j k , all k and d 0 + 1 i d + 1 . Similarly, we obtain that

[ k ( d d 0 ) ] 1 2 ( γ ^ D H t r i m γ ) : = I 1 + I 2 + I 3 ,

where

I 1 = γ { [ k ( d d 0 ) ] 1 2 [ 1 k j = 1 k ( d 0 + 1 d d 0 log Z j ( d 0 + 1 ) Z j ( d + 1 ) + i = d 0 + 2 d 1 d d 0 log Z j ( i ) Z j ( d + 1 ) ) 1 ] } ,

I 2 = [ k ( d d 0 ) ] 1 2 A 0 ( m / d ) 1 k j = 1 k { d 0 + 1 d d 0 log Z j ( d 0 + 1 ) Z j ( d + 1 ) + 1 d d 0 i = d 0 + 2 d log Z j ( i ) Z j ( d + 1 ) } ,

I 3 = o p ( 1 ) [ k ( d d 0 ) ] 1 2 A 0 ( m / d ) 1 k j = 1 k { ( d Z j ( d + 1 ) m ) ± δ d 0 + 1 d d 0 ( ( Z j ( d 0 + 1 ) Z j ( d + 1 ) ) δ + 1 ) } + o p ( 1 ) [ k ( d d 0 ) ] 1 2 A 0 ( m / d ) 1 k j = 1 k { ( d Z j ( d + 1 ) m ) ± δ i = d 0 + 2 d 1 d d 0 ( ( Z j ( i ) Z j ( d + 1 ) ) δ + 1 ) } .

We can show that I 1 d N ( 0, γ 2 ) , I 2 = [ k ( d d 0 ) ] 1 2 A 0 ( m / d ) { 1 + o p ( 1 ) } and I 3 P 0 as n , similar to the proof above, the statement in Theorem 2.1 follows.

Proof of Theorem 2. 2. We only show the proof for ρ < 0 and the proof for ρ = 0 is similar.

By Lemma S.2 in [18] , we have lim n { Z ( d + 1 ) > t 0 } = 1 , for any t 0 > 1 . Then by applying (6) twice with t = m / d ¯ and x = d ¯ Z j ( i ) / m , i = d 0 + 1 , , d + 1 and x = d ¯ Z j ( d j + 1 ) / m and using the same method as shown in the proof of Theorem 2.1, we obtain that

[ k ( d ¯ d 0 ) ] 1 2 ( γ ^ D H t r i m γ ) : = I 1 + I 2 + I 3 ,

where

I 1 = γ { [ k ( d ¯ d 0 ) ] 1 2 [ 1 k j = 1 k ( d 0 + 1 d j d 0 log Z j ( d 0 + 1 ) Z j ( d j + 1 ) + i = d 0 + 2 d j 1 d j d 0 log Z j ( i ) Z j ( d j + 1 ) ) 1 ] } ,

I 2 = [ k ( d ¯ d 0 ) ] 1 2 A 0 ( m / d ¯ ) ρ 1 k j = 1 k { ( d ¯ d j ) ρ ( d j Z j ( d j + 1 ) m ) ρ d 0 + 1 d j d 0 ( ( Z j ( d 0 + 1 ) Z j ( d j + 1 ) ) ρ 1 ) } + [ k ( d ¯ d 0 ) ] 1 2 A 0 ( m / d ¯ ) ρ 1 k j = 1 k { ( d ¯ d j ) ρ ( d j Z j ( d j + 1 ) m ) ρ i = d 0 + 2 d j 1 d j d 0 ( ( Z j ( i ) Z j ( d j + 1 ) ) ρ 1 ) } ,

I 3 = o p ( 1 ) [ k ( d ¯ d 0 ) ] 1 2 A 0 ( m / d ¯ ) 1 k j = 1 k { ( d ¯ d j ) ρ ± δ ( d j Z j ( d j + 1 ) m ) ρ ± δ d 0 + 1 d j d 0 ( ( Z j ( d 0 + 1 ) Z j ( d j + 1 ) ) ρ + δ + 1 ) } + o p ( 1 ) [ k ( d ¯ d 0 ) ] 1 2 A 0 ( m / d ¯ ) 1 k j = 1 k { ( d ¯ Z j ( d j + 1 ) m ) ρ ± δ i = d 0 + 2 d j 1 d j d 0 ( ( Z j ( i ) Z j ( d j + 1 ) ) ρ + δ + 1 ) } .

as n .

By Lemma 4.1. and the central limit theorem, we have that I 1 d N ( 0, γ 2 ) as n .

As for I 2 , by WLLN for triangular array and for each j, Z j ( d + 1 ) is independent with Z j ( i ) / Z j ( d + 1 ) , where i = d 0 + 1, , d j , similar to the proof of Theorem 2.1, we have that

I 2 = [ k ( d ¯ d 0 ) ] 1 2 A 0 ( m / d ¯ ) 1 k j = 1 k ( d ¯ d j ) ρ B ( m , d j , d 0 , ρ ) { 1 + o p ( 1 ) } , (21)

as n .

Similarly, as for I 3 , we obtain that I 3 P 0 as n . Combining with I 1 d N ( 0, γ 2 ) and (21) as n the statement in Theorem 2.2 follows.

Proof of Theorem 2.3. We only show the proof for ρ < 0 and the proof for ρ = 0 is similar.

By Lemma S.2 in [18] , we have lim n { Z ( d + 1 ) > t 0 } = 1 , for any t 0 > 1 . Then by applying (6) twice with t = m / d and x = d Z j ( i ) / m , i = d 0 + 1 , , d + 1 and x = d Z j ( d + 1 ) / m and using the same method in the prood of Theorem 2.1, we obtain that

[ k ( d d 0 ) ] 1 2 ( γ ^ D H t r i m γ ) : = I 1 + I 2 + I 3 ,

where

I 1 = γ { [ k ( d d 0 ) ] 1 2 [ 1 k j = 1 k ( d 0 + 1 d d 0 log Z j ( d 0 + 1 ) Z j ( d + 1 ) + i = d 0 + 2 d 1 d d 0 log Z j ( i ) Z j ( d + 1 ) ) 1 ] } ,

I 2 = [ k ( d d 0 ) ] 1 2 A 0 ( m / d ) ρ 1 k j = 1 k { ( d Z j ( d + 1 ) m ) ρ d 0 + 1 d d 0 ( ( Z j ( d 0 + 1 ) Z j ( d + 1 ) ) ρ 1 ) } + [ k ( d d 0 ) ] 1 2 A 0 ( m / d ) ρ 1 k j = 1 k { ( d Z j ( d + 1 ) m ) ρ i = d 0 + 2 d 1 d d 0 ( ( Z j ( i ) Z j ( d + 1 ) ) ρ 1 ) } ,

I 3 = o p ( 1 ) [ k ( d d 0 ) ] 1 2 A 0 ( m d ) 1 k j = 1 k { ( d Z j ( d + 1 ) m ) ρ ± δ d 0 + 1 d d 0 ( ( Z j ( d 0 + 1 ) Z j ( d + 1 ) ) ρ + δ + 1 ) } + o p ( 1 ) [ k ( d d 0 ) ] 1 2 A 0 ( m d ) 1 k j = 1 k { ( d Z j ( d + 1 ) m ) ρ ± δ i = d 0 + 2 d 1 d d 0 ( ( Z j ( i ) Z j ( d + 1 ) ) ρ + δ + 1 ) } .

By Lemma 4.1. and the central limit theorem, we have that I 1 d N ( 0, γ 2 ) as n .

By WLLN for triangular array and Z 1 ( d + 1 ) is independent with Z j ( i ) / Z j ( d + 1 ) for i = d 0 + 1, , d and j = 1, , k , we have that

I 2 = [ k ( d d 0 ) ] 1 2 A 0 ( m / d ) E { ( d Z 1 ( d + 1 ) m ) ρ } d d d 0 1 1 ρ { 1 + o p ( 1 ) } . (22)

By the Stirling’s formula, it follows that

E { ( d Z 1 ( d + 1 ) m ) ρ } = ( m d ) ρ Γ ( m + 1 ) Γ ( d ρ + 1 ) Γ ( d + 1 ) Γ ( m ρ + 1 ) ~ 1

as m .

Similarly, as for I 3 , we obtain that I 3 P 0 as n . Combining with I 1 d N ( 0, γ 2 ) and (22) as n the statement in Theorem 2.3 follows.

Conflicts of Interest

The authors declare no conflicts of interest regarding the publication of this paper.

References

[1] de Haan, L. and Ferreira, A. (2006) Extreme Value: An Introduction. Springer, New York.
[2] Dekkers, A.L.M. and de Haan, L. (1989) On the Estimation of the Extreme-Value Index and Large Quantile Estimation. The Annals of Statistics, 17, 1795-1832.
https://doi.org/10.1214/aos/1176347396
[3] Dekkers, A.L.M., Einmahl, J.H.J. and de Haan, L. (1989) A Moment Estimator for the Index of an Extreme-Value Distribution. The Annals of Statistics, 17, 1833-1855.
https://doi.org/10.1214/aos/1176347397
[4] Beirlant, J., Goegebeur, Y., Teugels, J. and Segers, J. (2004) Statistics of Extremes: Theory and Applications. John Wiley & Sons, Ltd., New York.
https://doi.org/10.1002/0470012382
[5] Li, J., Peng, Z. and Nadarajah, S. (2010) Asymptotic Normality of Location Invariant Heavy Tail Index Estimator. Extremes, 13, 269-290.
https://doi.org/10.1007/s10687-009-0088-4
[6] Brazauskas, V. and Serfling, R. (2001) Robust Estimation of Tail Parameters for Twoparameter Pareto and Exponential Models via Generalized Quantile Statistics. Extremes, 3, 231-249.
[7] Bhattacharya, S., Kallitsis, M. and Stoev, S. (2019) Data-Adaptive Trimming of the Hill Estimator and Detection of Outliers in the Extremes of Heavy-Tailed Data. Electronic Journal of Statistics, 13, 1872-1925.
https://doi.org/10.1214/19-EJS1561
[8] Bladt, M., Albrecher, H. and Beirlant, J. (2020). Threshold Selection and Trimming in Extremes. Extremes, 23, 629-665.
https://doi.org/10.1007/s10687-020-00385-0
[9] Aban, I.B., Meerschaert, M.M. and Panorska, A.K. (2006) Parameter Estimation for the Truncated Pareto Distribution. Journal of the American Statistical Association, 101, 270-277.
https://doi.org/10.1198/016214505000000411
[10] Beirlant, J., Alves, I.F., Gomes, I. and Meerschaert, M.M. (2014) Extreme Value Statistics for Truncated Pareto-Type Distributions. John Wiley & Sons, Ltd., New York.
https://doi.org/10.1002/9781118445112.stat04418
[11] Beirlant, J., Guillou, A. and Dierckx, G. (2007) Estimation of the Extreme Value Index and Extreme Quantiles under Random Censoring. Extremes, 10, 151-174.
https://doi.org/10.1007/s10687-007-0039-x
[12] Einmahl, J.H., Fils-Villetard, A. and Guillou, A. (2008) Statistics of Extremes under Random Censoring. Bernoulli, 14, 207-227.
https://doi.org/10.3150/07-BEJ104
[13] Gomes, M.I. and Neves, M.M. (2011) Estimation of the Extreme Value Index for Randomly Censored Data. In Biometrical Letters, 48, 1-22.
[14] Bladt, M., Albrecher, H. and Beirlant, J. (2021) Trimmed Extreme Value Estimators for Censored Heavy-Tailed Data. Electronic Journal of Statistics, 15, 3112-3136.
https://doi.org/10.1214/21-EJS1857
[15] Battey, H., Fan, J., Liu, H., Lu, J. and Zhu, Z. (2015) Distributed Estimation and Inference with Statistical Guanrantees. arXiv: 1509.05457.
https://doi.org/10.48550/arXiv.1509.05457
[16] Li, R., Lin, D.K.J. and Li, B. (2013) Statistical Inference in Massive Data Sets. Applied Stochastic Models in Business and Industry, 29, 399-409.
https://doi.org/10.1002/asmb.1927
[17] Volgushev, S., Chao, S.-K. and Cheng, G. (2019). Distributed Inference for Quantile Regression Processes. The Annals of Statistics, 47, 1634-1662.
https://doi.org/10.1214/18-AOS1730
[18] Chen, L., Li, D. and Zhou, C. (2021). Distributed Inference for Extreme Value Index. Biometrika, 109, 257-264.
https://doi.org/10.1093/biomet/asab001
[19] Wellner, R.L. (1978) Limit Theorems for the Ratio of the Empirical Distribution Function to the True Distribution Fuction. Zeitschrift für Wahrscheinlichkeitstheorie und Verwandte Gebiete, 45, 73-88.
https://doi.org/10.1007/BF00635964

Copyright © 2024 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.