Professional Documents
Culture Documents
Entropies in Terms of Moments of Order Statistics
Entropies in Terms of Moments of Order Statistics
Entropies in Terms of Moments of Order Statistics
Narayanaswamy Balakrishnan ·
Francesco Buono · Maria Longobardi
arXiv:2009.02029v1 [math.ST] 4 Sep 2020
1 Introduction
where log is the natural logarithm and f is the probability density function
(pdf) of X. In the following we use F and F to indicate the cumulative dis-
tribution function (cdf) and the survival function (sf) of X, respectively.
N. Balakrishnan,
McMaster University, Hamilton, Ontario, Canada
E-mail: bala@mcmaster.ca
F. Buono
Universit degli Studi di Napoli Federico II, Naples, Italy
E-mail: francesco.buono3@unina.it
M. Longobardi
Universit degli Studi di Napoli Federico II, Naples, Italy
E-mail: maria.longobardi@unina.it
2 Narayanaswamy Balakrishnan et al.
In the literature, there are several different versions of entropy, each one
suitable for a specific situation. Rao et al., 2004, introduced the Cumulative
Residual Entropy (CRE) of X as
Z +∞
E(X) = − F (x) log F (x)dx.
0
Choosing k = 1 and k = n we get the smallest and the largest order statistic,
respectively. Their cdf and pdf are given by
Z +∞
E(X) = − (1 − F (x)) log(1 − F (x))dx
0
Z +∞
+∞
= −x(1 − F (x)) log(1 − F (x))0 − x log(1 − F (x))f (x)dx −
0
Z +∞
xf (x)dx
0
Z +∞
= x[− log(1 − F (x))]f (x)dx − E(X)
0
+∞
Z +∞ " X #
F (x)n
= x f (x)dx − E(X)
0 n=1
n
+∞
X 1
= µn+1:n+1 − E(X), (2)
n=1
n(n + 1)
+∞
X 1 1
E(X) = − µn+1:n+1 − E(X). (3)
n=1
n n+1
1 1
E(X) = 1 and E(Xn:n ) = 1 + + ···+ .
2 n
4 Narayanaswamy Balakrishnan et al.
where µ1:n+1 is the mean of the smallest order statistic from a sample of size
n + 1 from F , provided that limx→+∞ −xF (x) log F (x) = 0.
On cumulative entropies in terms of moments of order statistics 5
+∞
X 1 1
CE(X) = − − µ1:n+1 + E(X). (5)
n=1
n n+1
1
µ1:n = ,
n
+∞
X 1 1 1
CE(X) = − − +1
n=1
n n+1 n+1
+∞ +∞
X 1 X 1
=− + +1
n=1
n(n + 1) n=1 (n + 1)2
π2
= − 1,
6
Example 4 For the standard uniform distribution, using the fact that
1
µ1:n = ,
n+1
+∞
X 1 1
CE(X) = − +
n=1
n(n + 1)(n + 2) 2
+∞ +∞ +∞
1 1X1 X 1 1X 1
= − + −
2 2 n=1 n n=1 n + 1 2 n=1 n + 2
1
=
4
µn:n − µ = µ − µ1:n ,
x2 +∞ 1 +∞ 2
Z
E w (X) = − (1 − F (x)) log(1 − F (x))0 − x log(1 − F (x))f (x)dx
2 2 0
1 +∞ 2
Z
− x f (x)dx
2 0
Z +∞
1 1
= x2 [− log(1 − F (x))]f (x)dx − E(X 2 )
2 0 2
Z +∞ " X +∞
#
1 F (x)n 1
= x2 f (x)dx − E(X 2 )
2 0 n=1
n 2
+∞
1X 1 (2) 1
= µn+1:n+1 − E(X 2 ), (6)
2 n=1 n(n + 1) 2
(2)
where µn+1:n+1 is the second moment of the largest order statistic in a sample
2
of size n + 1, provided that limx→+∞ − x2 (1 − F (x)) log(1 − F (x)) = 0.
Example 5 For the standard uniform distribution, using the fact that
(2) n+1 1
µn+1:n+1 = and E(X 2 ) = ,
n+3 3
+∞
1X 1 n+1 1
E w (X) = −
2 n=1 n(n + 1) n + 3 6
+∞
1X 1 1 1
= − −
6 n=1 n n + 3 6
1 1 1 1 5
= 1+ + − = .
6 2 3 6 36
way
x2 +∞ 1 +∞ 2
Z
CE w (X) = − F (x) log F (x)0 + x log F (x)f (x)dx +
2 2 0
1 +∞ 2
Z
x f (x)dx
2 0
1 +∞ 2 1
Z
= x log[1 − (1 − F (x))]f (x)dx + E(X 2 )
2 0 2
Z +∞ X +∞ n
1 (1 − F (x)) 1
=− x2 f (x)dx + E(X 2 )
2 0 n=1
n 2
+∞
1X 1 (2) 1
=− µ + E(X 2 ), (7)
2 n=1 n(n + 1) 1:n+1 2
(2)
where µ1:n+1 is the second moment of the smallest order statistic in a sample
2
of size n + 1, provided that limx→+∞ − x2 F (x) log F (x) = 0.
3 Bounds
Let us consider a sample with parent distribution X such that E(X) = 0 and
E(X 2 ) = 1. Hartley and David, 1954, and Gumbel, 1954, have shown that
n−1
µn:n ≤ √ .
2n − 1
We relate µn:n with the mean of the largest statistic order from the stan-
dard distribution. In fact, by normalizing the random variable X with mean
µ and variance σ 2 we get
X −µ
Z= .
σ
Hence, the cdf FZ is given in terms of the cdf FX by
Then, cdf and pdf of the largest order statistic in a sample of size n are
n n−1
FZn:n (x) = FX (σx + µ), fZn:n (x) = nFX (σx + µ)fX (σx + µ)σ.
Theorem 2 Let X be DFR (decreasing failure rate). Then, we have the fol-
lowing lower bound for CE(X)
p
E(X 2 ) π2
CE(X) ≥ E(X) − √ 2− . (10)
2 6
Proof Let X be DFR. From Theorem 12 of Rychlik (2001) we know that for
a sample of size n, if
j
X 1
δj = ≤2 j ∈ {1, . . . , n}
n+1−k
k=1
then
δj p
E(Xj:n ) ≤ √ E(X 2 ).
2
1
For j = 1 we have δ1 = n ≤ 2 for all n ∈ N and we get
p
E(X 2 )
E(X1:n ) ≤ √ .
2n
Then, from (4) we get the following lower bound for CE(X)
+∞ p
X 1 E(X 2 )
CE(X) ≥ − √ + E(X)
n=1
n(n + 1)2 2
p
E(X 2 ) π2
= E(X) − √ 2− .
2 6
Remark 4 We note that we can not provide an analogous bound for E(X)
because δn ≤ 2 is not fulfilled for n ≥ 4.
From (2) and (4) we get the following expression for the sum of the cumu-
lative residual entropy and the cumulative entropy
+∞
X 1
E(X) + CE(X) = (µn+1:n+1 − µ1:n+1 ). (17)
n=1
n(n + 1)
Cal et al., 2017, showed a connection among (17) and the partition entropy
studied by Bowden, 2007.
Theorem 5 We have the following bound for the sum of the CRE and the
CE √
+∞
X 2σ
E(X) + CE(X) ≤ √ ≃ 3.09 σ. (18)
n=1
n n +1
Proof From Theorem 3.24 of Arnold and Balakrishnan, 1989, we know the
following bound for the difference between the expectation of the largest and
the smallest order statistics from a sample of size n + 1
p
µn+1:n+1 − µ1:n+1 ≤ σ 2(n + 1), (19)
and so using (19) in (17) we get the following bound for the sum of the CRE
and the CE
+∞ p +∞ √
X σ 2(n + 1) X 2σ
E(X) + CE(X) ≤ = √ ≃ 3.09 σ.
n=1
n(n + 1) n=1
n n +1
About a symmetric distribution, Arnold and Balakrishnan, 1989, showed
that if we have a sample X1 , . . . , Xn with parent distribution X symmetric
about the mean µ with variance 1, then
√
r
1
E(Xn:n ) − E(X1:n ) ≤ n 2 − B(n, n), (20)
2n − 1
where B(n, n) is the complete beta function.
Using the bound (20) for a non-negative parent distribution symmetric
about the mean µ and with variance σ 2 , we get
√
r
1
µn:n − µ1:n = σ (E(Zn:n ) − E(Z1:n )) ≤ σn 2 − B(n, n). (21)
2n − 1
12 Narayanaswamy Balakrishnan et al.
Conflict of interest
References