1. Introduction
Suppose are independent and identically distributed observations from a distribution, where is differentiable with a density which is positive in an interval and zero elsewhere. The order statistics of the sample is defined by the arrangement of from the smallest to largest denoted as. Then the p.d.f. of the order statistics, is given by
(1)
for details refer to [1].
Order statistics has been studied by statisticians for some time and has been applied to problems of statistical estimation [2], reliability analysis, image coding [3] etc. Some information theoretic aspects of order statistics have been discussed in the literature. Wong and Chen [4] showed that the difference between average entropy of order statistics and the entropy of a data distribution is a constant. Park [5] showed some recurrence relations for entropy of order statistics. Information properties of order statistics based on Shannon entropy [6] and Kullback-Leibler [7] measure using probability integral transformation have been studied by Ebrahimi et al. [8]. Arghami and Abbasnejad [9] studied Renyi entropy properties based on order statistics. The Renyi [10] entropy is a single parameter entropy. We consider a generalized two parameter, the Verma entropy [11], and study it in context with order statistics. Verma entropy plays a vital role as a measure of complexity and uncertainty in different areas such as physics, electronics and engineering to describe many chaotic systems. Considering the importance of this entropy measure, it will be worthwhile to study it in case of order statistics. The rest of the article is organized as follows:
In Section 2, we express generalized entropy of order statistics in terms of generalized entropy of order statistics of uniform distribution and study some of its properties. Section 3 provides bounds for entropy of order statistics. In Section 4, we derive an expression for residual generalized entropy of order statistics using residual generalized entropy for uniform distribution.
2. Generalized Entropy of Order Statistics
Let be a random variable having an absolutely continuous cdf and pdf, then Verma [11] entropy of the random variable with parameters is defined as:
(2)
where
is the Renyi entropy, and
is the Shannon entropy .
We use the probability integral transformation of the random variable where the distribution of U is the standard uniform distribution. If are the order statistics of a random sample from uniform distribution, then it is easy to see using (1) that has beta distribution with parameters and. Using probability integral transformation, entropy (2) of the random variable can be represented as
(3)
Next, we prove the following result:
Theorem 2.1 The generalized entropy of can be expressed as
(4)
where denotes the entropy of the beta distribution with parameters and, denotes expectation of over and is the beta density with parameters and
.
Proof: Since which implies. Thus, from (3) we have
(5)
It is easy to see that the entropy (2) for the beta distribution with parameters and (that is, the order statistics of uniform distribution) is given by
(6)
Using (6) in (5), the desired result (4) follows.
In particular, by taking, (4) reduces to
a result derived by Ebrahimi et al. [8].
Remark: In reliability engineering -outof- systems are very important kind of structures. A -out-of- system functions iff atleast
components out of components function. If denote the independent lifetimes of the components of such system, then the lifetime of the system is equal to the order statistic. The special case of and, that is for sample minima and maxima correspond to series and parallel systems respectively. In the following example, we calculate entropy (4) for sample maxima and minima for an exponential distribution.
Example 2.1 Let be a random variable having the exponential distribution with pdf
Here, and the expectation term is given by
For, from (6), we have
Hence, using (4)
which confirms that the sample minimum has an exponential distribution with parameter, since
where is an exponential variate with parameter. Also
Hence, the difference between the generalized entropy of first order statistics i.e. the sample minimum and the generalized entropy of parent distribution is independent of parameter, but it depends upon sample size. Similarly, for sample maximum, we have
It can be seen easily that the difference between and is
which is also independent of parameter.
3. Bounds for the Generalized Entropy of Order Statistics
In this section, we find the bounds for generalized entropy for order statistics (4) in terms of entropy (2). We prove the following result.
Theorem 3.1 For any random variable with, the entropy of the order statistics is bounded above as
(7)
where
and, bounded below as
(8)
where, , and is the mode of the distribution and is pdf of the random variable.
Proof: The mode of the beta distribution is. Thus,
For, from (4)
which gives (7).
From (4) we can write
Example 3.1 For the uniform distribution over the interval we have
and from (6),
and
Hence, using (7) we get
Further, for uniform distribution over the interval
,. Using (8) we get
Thus, for uniform distribution, we have
We can check that the bounds for are same as that of.
Example 3.2 For the exponential distribution with parameter, we have and
Thus, as calculated in Example 2.1
Using Theorem 3.1
Here we observe that the difference between upper bound and is, which is an increasing function of n. Thus, for the exponential distribution upper bound is not useful when sample size is large.
4. The Generalized Residual Entropy of Order Statistics
In reliability theory and survival analysis, usually denotes a duration such as the lifetime. The residual lifetime of the system when it is still operating at time, given by has the probability density
, where. Ebrahimi [12] proposed the entropy of the residual lifetime as
(9)
Obviously, when, it reduces to Shannon entropy.
The generalized residual entropy of the type is defined as
(10)
where. When, it reduces to (2).
We note that the density function and survival function of (refer to [13]), denoted by and, respectively are
(11)
where
(12)
and
(13)
where
(14)
and are known as the beta and incomplete beta functions respectively. In the next lemmawe derive an expression for for the dynamic version of as given by (6).
Lemma 4.1 Let be the order statistics based on a random sample of size from uniform distribution on. Then
(15)
Proof: For uniform distribution using (10), we have
(16)
Putting values from (11) and (13) in (16), we get the desired result (15).
If we put in (15), we get (6).
Using this, in the following theorem, we will show that the residual entropy of order statistics can be represented in terms of residual entropy of uniform distribution.
Theorem 4.1 Let be an absolutely continuous distribution function with density. Then, generalized residual entropy of the order statistics can be represented as
(17)
where
Proof: Using the probability integral transformation
and above lemma, the result follows.
Take in (17), it reduces to (4).
Example 4.1 Suppose that is exponentially distributed random variable with mean. Then,
and we have
For, Theorem 4.1 gives
Also
Hence
So, in the exponential case the difference between generalized residual entropy of the lifetime of a series system and residual generalized entropy of the lifetime of each component is independent of time.
5. Conclusion
The two parameters generalized entropy plays a vital role as a measure of complexity and uncertainty in different areas such as physics, electronics and engineering to describe many chaotic systems. Using probability integral transformation we have studied the generalized and generalized residual entropies based on order statistics. We have explored some properties of these entropies for exponential distribution.
6. Acknowledgements
The first author is thankful to the Center for Scientific and Industrial Research, India, to provide financial assistance for this work.