Professional Documents
Culture Documents
Approximation of Kolmogorov Smirnov Test Statistic
Approximation of Kolmogorov Smirnov Test Statistic
To cite this article: Long Bai & David Kalaj (2021) Approximation of Kolmogorov–Smirnov test
statistic, Stochastics, 93:7, 993-1027, DOI: 10.1080/17442508.2020.1844705
Article views: 63
China; b Department of Actuarial Science, University of Lausanne, Lausanne, Switzerland; c Faculty of Natural
Sciences and Mathematics, University of Montenegro, Podgorica, Montenegro
for large u, where W(x) is the multivariate Brownian sheet based AMS CLASSIFICATIONS
on a distribution function F. The asymptotic results are obtained for Primary 60G15; Secondary
general F and some important examples are also shown. 60G70
1. Introduction
Let X 1 , . . . , X m be n-dimensional independent random vectors with common continuous
distribution function (df) F. We shall assume without loss of generality that the marginal
dfs of F are the uniform df on [0, 1]. Define the empirical df Fm of F by
1
m
Fm (x) := I[0,x] (X i ), x = (x1 , . . . , xn ) ∈ [0, 1]n ,
m
i=1
where IA denote the indicator of the set A and [0, x] = ni=1 [0, xi ]. Let W denote the
(unpinned) Brownian sheet determined by F, i.e. W is a centred Gaussian random field
with covariance function
R(x, y) = E W(x)W(y) = F(x ∧ y), x, y ∈ [0, 1]n .
Further WF is the pinned Brownian sheet, which is a centred Gaussian random field with
covariance function
RF (x, y) = E WF (x)WF (y) = F(x ∧ y) − F(x)F(y), x, y ∈ [0, 1]n .
√
It is well known, see e.g.[1,10–12] that m(Fm − F) converges weakly to WF as m → ∞ in
the space of all bounded functions on [0, 1]n under the topology of uniform convergence.
Consequently, if
n √
Tm (F) := sup m(Fm (x) − F(x))
x∈[0,1]n
For WF , the pinned version of W on [0, 1]n , we have the following representation:
d
{WF (x)} = {W(x) − F(x)W(1)} , x ∈ [0, 1]n ,
d
where = means identical distribution.
Similarly to Brownian bridge, for WF we have another conditional representation (see
[1]), namely,
d
{WF (x)} = W(x)W(1) = 0 , x ∈ [0, 1]n .
In particular, we obtain the asymptotic results for some cases with important applica-
tion.
The rest of the paper is organized as follows. In Section 2, we state our main results
and some examples are given in Section 3. The proofs of the main results and some useful
lemmas are presented in Sections 4 and Appendix, respectively.
2. Main results
Before stating our main results, we need to introduce some notation. For x, y ∈ Rn ,
In the sequel, let (·) denote the survival function of a standard normal random variable,
and λk (A) is the Lebesgue measure on Rk of some measurable set A ⊂ Rk .
Theorem 2.1: Let F(x), x ∈ [0, 1]n , n ≥ 2, be a continuous df. Suppose that there exists a
x),
function h( x = (x1 , . . . , xn−1 ) ∈ L ⊆ [0, 1]n−1 with λn−1 (L) = 0 such that
1
D := {x = (
x, h( x ∈ L} = x ∈ [0, 1]n : F(x) =
x)) :
2
and λn−1 (∂L) = 0 (i.e. L is a Jordan measurable set). If h is continuously differentiable in the
interior of L and further
F(x) − F(y) −
n ai (z)(xi − yi )
lim sup sup
n i=1
= 0, (2)
δ→0 z∈D |x−z|, y−z ≤δ
| | i=1 xi − yi
x=y
where
K := 23(n−1) n−1
i=1 ai (
x, h(
x)) d
x ∈ (0, ∞). (4)
L
Corollary 2.2: If the df F(x), x ∈ [0, 1]n has a bounded positive density function f, then there
exists a continuously differentiable function h defined on the interior of a Jordan measurable
set L ⊂ [0, 1]n−1 with λn−1 (L) > 0 such that (3) holds with ai (x) = ∂F(x)/∂xi .
Consider the random field W ∗ (x) = B0 (F(x)), x ∈ [0, 1]n for a given df F(x), x ∈ [0, 1]n
with B0 a standard Brownian bridge. We have that W ∗ is a centred Gaussian random field,
and it has covariance function F(x) ∧ F(y) − F(x)F(y). In the special case that F(x) =
F(x1 ), x ∈ [0, 1]n with F a univariate df, then F(x) ∧ F(y) = F(x ∧ y), hence W ∗ has the
996 L. BAI AND D. KALAJ
same law as WF . These observations motivate the result of next theorem, where essentially
we use the fact that the tail asymptotics of the supremum of a Brownian bridge with trend
is known, see [3].
Theorem 2.3: Let F(x), x ∈ [0, 1]n be a n-dimensional continuous df. If there exist some
δ ∈ (0, 1/2) such that
F(x ∧ y) = F(x) ∧ F(y) (5)
holds for x, y ∈ Dδ with
1 1
Dδ := x ∈ [0, 1]n : − δ ≤ F(x) ≤ + δ ,
2 2
then for any w ∈ R
sup W(x) > uW(1) = w ∼ e−2u +2uw ,
2
P u → ∞. (6)
x∈[0,1]n
Remark 2.4: By the proof of Theorem 2.3 in Section 4, we know that if (5) holds for any
x, y ∈ [0, 1]n , then for any w ∈ R and u > 0 we have
P sup W(x) > uW(1) = w = e−2u +2uw .
2
x∈[0,1]n
Theorem 2.5: Let F(x), x ∈ [0, 1]n be a continuous df. If further there exists some δ ∈
(0, 1/4) such that
inf F(x ∧ y) > δ (7)
x,y∈Dδ
holds for Dδ := x ∈ E : 12 − δ ≤ F(x) ≤ 12 + δ where Jordan measurable set E ⊆ [0, 1]n
satisfies that there exists x0 ∈ E such that F(x0 ) = 12 , then we have for w ∈ R as u → ∞
P sup |W(x)| > uW(1) = w
x∈E
∼ P sup W(x) > uW(1) = w + P sup W(x) > uW(1) = −w . (8)
x∈E x∈E
Remarks 2.6: (i) When E = [0, 1]n , for F(x) satisfying (5) in Theorem 2.3, (7) always
holds. In fact, for x, y ∈ Dδ with δ ∈ (0, 14 )
1 1
− δ > > δ.
F(x ∧ y) = F(x) ∧ F(y) ≥
2 4
Further by (6) and (8), under the conditions of Theorem 2.3, we have
P sup |W(x)| > uW(1) = w ∼ ce−2u +2u|w| , u → ∞,
2
(9)
x∈[0,1]n
where c = 1 if w = 0 and c = 2 if w = 0.
STOCHASTICS 997
(ii) Under the conditions of Theorem 2.1, if further (7) holds for E = [0, 1]n , we have
by (8) in Theorem 2.5
sup |W(x)| > uW(1) = w ∼ cKu2(n−1) e−2u +2uw ,
2
P u → ∞, (10)
x∈[0,1]n
3. Applications
In this part, we give the asymptotic results of (1) for some particular df F. First, we give
several two-dimensional cases.
Proposition 3.1: For (i) − −(ii) below and w ∈ R, we have that both (3) and (10) with
n = 2 hold.
Remarks 3.2: (i) In [1][Theorem 3.1], the following upper bound for case in Proposi-
tion 3.1 is given as
∞
(8i2 u2 − 2)e−2i u ,
2 2
P sup W(x)W(1) = 0 > u ≤ u > 0.
x∈[0,1]2 i=1
Comparing with our exact result, we see that the prefactor for this upper bound is two
times our constant K = 4.
(ii) In the light of [1][Theorem 3.1] for any two-dimensional df F on [0, 1]2 and any
u>0
P sup WF (x) > u ≤ P sup WG (x) > u ,
x∈[0,1]2 x∈[0,1]2
with G(x) = (x1 + x2 − 1)+ , x ∈ [0, 1]2 . Consequently, our result in Proposition 3.1 gives
an asymptotic upper bound for any two-dimensional df F on [0, 1]2 .
Proposition 3.3: For (i) − −(ii) below both (3) and (10) hold for any w ∈ R. Moreover we
have: (i) If F(x) = ni=1 xi , x ∈ [0, 1]n , then
−1 n−1 1
K=2 2(n−1)
(n−1
i=1 xi ) dx, L=
x ∈ [0, 1] n−1
: ≤ i=1 xi ≤ 1 .
L 2
(ii) If
F(x) = d min xi + (1 − d)ni=1 xi , d ∈ (0, 1),
1≤i≤n
then
n−2
(n−1) n−1
i=1 xi
K = n(4d) x,
n−1 d
L d + (1 − d)n−1
i=1 xi
1
x ∈ [0, 1]n−1 :
L= ≤ min xi .
2d + 2(1 − d)n−1
i=1 xi
1≤i≤n−1
Remark 3.4: (i) The result of Proposition 3.3, i) for n = 2 and w = 0 agrees with the claim
of [13][Theorem 1].
(ii) In [1][Theorem 2.1], a lower bound for the n-dimensional case is given by
n−1
(2u2 − 2uw)i
P sup W(x)W(1) = w > u ≥ e−2u +2uw
2
, u > w.
x∈[0,1]n i=0
i!
Since
n−1
(2u2 − 2uw)i (2u2 )n−1
∼ , u → ∞,
i=0
i! (n − 1)!
in the particular case n = 3 we have 16(ln 2)2 ≥ 2.
Next statement shows a particular case that satisfies the hypotheses in Theorem 2.3.
and
|W(x)| W(1) = w > u ∼ ce−2u +2u|w| ,
2
P sup u → ∞,
x∈[0,1]n
with c = 1 if w = 0 and c = 2 if w = 0.
Remark 3.6: In Proposition 3.5, if d = 12 , F(x) = min1≤i≤n xi , i.e. the upper copula.
STOCHASTICS 999
4. Proofs
Proof of Theorem 2.1: Hereafter, we denote by Qi , i ∈ N some positive constants that may
differ from line to line.
By the monotonicity and continuity of df F(x), x ∈ [0, 1]n , h(
x) is a continuous function
over L ⊂ [0, 1]n−1 which is strictly decreasing along every line parallel to the axes (and so
on all increasing paths).
Then for x ∈ L and xn = h(x) we set
ãi (
x) := ai (
x, h(
x)), i = 1, . . . , n,
and
ai := inf ai (x) > 0, ai := sup ai (x) < ∞, i = 1, . . . , n,
x∈D x∈D
where we use the fact that ai (x)’s are continuous and positive function.
We have for u > w
P sup W(x) > uW(1) = w = P sup (WF (x) + F(x)w) > u ,
x∈[0,1]n x∈[0,1]n
which attains its maximum equal to 14 over [0, 1]n at z with F(z) = 12 . Hence σF (x) attains
its maximum equal to 12 at x ∈ D and as F(x) → 12
1 1 2
− σF (x) ∼ F(x) − . (11)
2 2
Since by (2), there exist ε1 ∈ 0, min1≤i≤n ai , for any z ∈ D, if |x − z| , y − z < δ
n
n
F(x) − F(y) ≤ ai (z)(xi − yi ) + ε1 xi − yi ≤ Q1 x − y , (12)
i=1 i=1
n
n
F(x) − F(y) ≥ ai (z)(xi − yi ) − ε1 xi − yi ≥ Q2 x − y . (13)
i=1 i=1
By (13), we have F(x) is strictly increasing along every line parallel to the axes in
and for any δ0 ∈ (0, 1/2) we can take δ ∈ (0, 12 ) small enough such that
with δ0 → 0 as δ → 0 and
2
1 1
− δ0 ≤ σF2 (x) ≤ , x ∈ E(δ).
2 4
1000 L. BAI AND D. KALAJ
WF (x) WF (y)
For the correlation function rF (x, y) := Cov σF (x) , σF (y) , we have for any z ∈ D, if
|x − z| , y − z ≤ δ
E WF (x)WF (y)
1 − rF (x, y) = 1 −
σF (x)σF (y)
2 2
E WF (x) − WF (y) − σF (x) − σF (y)
= . (16)
2σF (x)σF (y)
By (15), we have
2 2
2 σF (x) − σF2 (y)
σF (x) − σF (y) = 2
σF (x) + σF (y)
1 2
≤ (F(x) − F(y)) − (F 2 (x) − F 2 (y))
(1 − 2δ0 )2
1
= (F(x) − F(y))2 (1 − (F(x) + F(y)))2
(1 − 2δ0 )2
4δ02 Q22 2
≤ x − y
(1 − 2δ0 )2
and
2
2
E WF (x) − WF (y) =E W(x) − W(y) − (F(x) − F(y))W(1)
n
F(x) + F(y) − 2F(x ∧ y) ≤ z) + ε1 ) xi − yi ,
(ãi (
i=1
n
F(x) + F(y) − 2F(x ∧ y) ≥ z) − ε1 ) xi − yi .
(ãi (
i=1
2
1 − rF (x, y) ≤ F(x) + F(y) − 2F(x ∧ y)
(1 − 2δ0 ) 2
n
2
≤ (ãi (
z) + ε1 ) xi − yi
(1 − 2δ0 )2
i=1
n
≤ 2(1 + ε2 ) z) + ε1 ) xi − yi
(ãi (
i=1
STOCHASTICS 1001
and
2
2 4δ02 Q22 2
1 − rF (x, y) ≥ E W F (x) − W F (y) − x − y
(1 + 2δ0 ) 2 (1 − 2δ0 ) 2
2 2
2
≥ F(x) + F(y) − 2F(x ∧ y) − Q 2 x − y
(1 + 2δ0 )2
4δ02 Q22 2
− x − y
(1 − 2δ0 ) 2
n
2 4δ02 Q22
≥ ã i (
z) x i − y i
− Q 2
+ x − y2
2
(1 + 2δ0 )2 (1 − 2δ0 )2
i=1
n
≥ 2(1 + ε4 ) ãi (z) xi − yi ,
i=1
2 1 2
≤ ≤ .
(1 + 2δ)2 2σF (x)σF (y) (1 − 2δ)2
Hence
1 − rF (x, y)
− 1 = 0.
lim sup sup
n
(17)
δ→0 z∈D |x−z|, y−z <δ 2 i=1 ai (z) xi − yi
| |
x=y
and for x = y, x, y ∈ E(δ), if F(x) = F(y), F(x ∧ y) < F(x) ∧ F(y) since F(x) is strictly
increasing along every line parallel to the axes in E(δ). Then in (18), at least one of the two
inequality strictly holds implying
and
x, L) ≤ δ, xn − q(
E0 (δ) = {x : ρ( x) ≤ δ},
where
x), if
h( x ∈ L,
x) =
q(
y) + n−1
h( y)(xi − yi ), if
i=1 ã( x∈/ L and z ∈ L : |
y ∈ { z −
x| = ρ(
x, L)}.
1
σm2 = sup σF2 (x) < .
x∈[0,1]n \E0 (δ) 4
where
1 (u) = P sup (WF (x) + F(x)w) > u .
x∈E0 (δ)
Notice that the variance function σX2 (x) of X(x) attains its maximum equal to 1 at D.
In (2) for x ∈ E0 (δ) taking y = z = ( x, h(
x)) leading to
F(x) − 1/2 − ãn ( x))
x) (xn − h(
lim sup sup = 0, (23)
δ→0 x∈L |xn −h(
x)|≤δ |xn − h(
x)|
xn =h(
x)
and
g(x)
lim sup sup − 1 = 0. (25)
δ→0 x)|<δ 2wãn (
x∈L |xn −h( x)(xn − h(
x))
xn =h(
x)
which implies that for any small δ ∈ (0, 1) there exist a constant Q3 such that
h(
x) − h(y)
sup sup
n−1 ≤ Q3 .
y∈L |
x, y|≤δ
x−
xi − yi
i=1
x=
y
n
n
2(1 − ε)
ai (z) xi − yi ≤ 1 − rX (x, y) ≤ 2(1 + ε) ai (z) xi − yi . (29)
i=1 i=1
1004 L. BAI AND D. KALAJ
where
1 (u) = P sup (X(x) + g(x)) > μ, sup (X(x) + g(x)) > μ ,
x∈Dk x∈Dl
(
k,
l)∈K1
2 (u) = P sup (X(x) + g(x)) > μ, sup (X(x) + g(x)) > μ .
x∈Dk x∈Dl
(
k,
l)∈K2
as μ → ∞, δ → 0. Since ai (x, h(
x)),
x ∈ Lis positive continuous and L is Jordan measur-
able with positive Lebesgue measure, then L n−1 i=1 ia (
x, x))
h( x ∈ (0, ∞).
d
STOCHASTICS 1005
Similarly,
xu2(n−1) e−2u +2uw
2
P sup (X(x) + g(x)) > μ ∼ 2 3(n−1)
n−1
i=1 ai (
x, h(
x)) d
x∈Dk L
k∈L2
(33)
as μ → ∞, δ → 0. Next we will show that
i (u), i = 1, 2 as u → ∞ are both negligible
compared with
P sup (X(x) + g(x)) > μ .
x∈Dk
k∈L2
For any (
k,
l) ∈ K1 (u), without loss of generality, we assume that k1 + 1 = l1 . Let
!
D1k = k1 δ, (k1 + 1)δ − δ 2 × n−1
j=2 [kj δ, (kj + 1)δ],
!
D2k = (k1 + 1)δ − δ 2 , (k1 + 1)δ × n−1 j=2 [kj δ, (kj + 1)δ].
Then
P sup (X(x) + g(x)) > μ, sup (X(x) + g(x)) > μ
x∈Dk x∈Dl
⎧ ⎫
⎨ ⎬
≤P sup (X(x) + g(x)) > μ, sup (X(x) + g(x)) > μ
⎩x∈D1 (u) x∈D(u) ⎭
k l
⎧ ⎫
⎨ ⎬
+P sup (X(x) + g(x)) > μ .
⎩x∈D2 ⎭
k
⎧ ⎫
⎨ ⎬
11 (u) := P sup (X(x) + g(x)) > μ
⎩x∈D2 ⎭
k∈L2
k
π (2w)2 2n−1
∼ 2n δ n n−1
i=1 ãi (c
k) e 8 μ (μ)
2
k∈L2
2(n−1) −2u2 +2wu
2(n−1) −2u2 +2uw
∼ 23(n−1) δ n n−1 ã
i=1 i (c k ) u e = o u e
k∈L2
as u → ∞, δ → 0.
1006 L. BAI AND D. KALAJ
(2(μ − Q) − Q)2
≤ exp −
2(4 − Qδ)
(
k,
l)∈K1
= o u2(n−1) e−2u +2uw , u → ∞.
2
(34)
Inserting (32), (33), (35) and (36) into (30) and (31) implies
P sup (X(x) + g(x)) > μ
x∈E0 (δ)
xu2(n−1) e−2u +2uw ,
2
∼ 23(n−1) n−1
i=1 ai (
x, h(
x)) d u → ∞,
L
which combined with (21) and (22) establishes the claim (3).
STOCHASTICS 1007
Proof of Corollary 2.2: Since we assume that f is positive in [0, 1]n , we have for any x ∈
(0, 1)n−1 the ith partial derivative of F denoted by ai (x) is positive and continuous. Let
Q = supx∈[0,1]n f (x) which is finite and positive by the assumption and D = {x ∈ [0, 1]n :
F(x) = 12 }. Using Taylor expansion, we have
n
n
F(x) − F(z) − ai (z)(xi − zi ) ≤ Q |xi − zi | 2
i=1 i=1
In view of Theorem A.3 the set D is not empty and moreover its projection on [0, 1]n−1
denoted by L is Jordan measurable with positive Lebesgue measure (with respect to λn−1 ).
Set Lo is the interior of L. By the positivity of the partial derivatives on interior of [0, 1]n
and the fact that F is strictly increasing on [0, 1]n−1 , we have that for any x ∈ Lo there is
only one xn such that F( x, xn ) = 1/2. Consequently, xn = g( x) for some bijective function
g for anyx ∈ L . Since F is continuously differentiable on L , for any x ∈ D with
o o x ∈ Lo by
the implicit function theorem there exists ε > 0 such that for any y ∈ Oε (x) = {
z ∈ I n−1 :
|
z − x| < ε} ⊆ L , we have F(
o y)) = 1/2. By the above, λn−1 (L ) ≥ λn−1 (Oε (
y, hx ( o x)) > 0
and hx does not depend on x and hx ( x) = g(
x) for any x ∈ Lo . It follows thus that g is
continuously differentiable on Lo . Moreover for any x ∈ Lo
∂h(
x) ai (
x, h(
x))
=− < 0, i = 1, . . . , n − 1,
∂xi an (
x, h(
x))
and thus h is continuously differentiable in Lo . Hence the proof follows by Theorem 2.1
since (2), the continuous differentiability of h in Lo and the Jordan measurability of L are
satisfied.
1
which attains its maximum equal to 4 over [0, 1]n at (z) which satisfies F(z) = 12 .
1008 L. BAI AND D. KALAJ
1
sup σF2 (x) = − δ2.
x∈[0,1]n \Dδ 4
P sup W(x) > uW(1) = w = P sup (WF (x) + F(x)w) > u
x∈[0,1]n \Dδ x∈[0,1]n \Dδ
≤P sup WF (x) > u − Q1
x∈[0,1]n \Dδ
(u − Q1 − Q2 )2
≤ exp −
2σm2
−2u+2uw
=o e , (37)
where Q1 = supx∈[0,1]n F(x) and Q2 = E supx∈[0,1]n WF (x) < ∞. By (5), we know that
WF (x), x ∈ Dδ is a Gaussian fields with covariance function
E WF (x)WF (y) = F(x ∧ y) − F(x)F(y)
= F(x) ∧ F(y) − F(x)F(y)
= E B0 (F(x))B0 (F(y)) , (38)
where B0 (t) = B(t) − tB(1) is the standard Brownian bridge. Then by [3] [Example 3.12]
P sup W(x) > uW(1) = w = P sup (WF (x) + F(x)w) > u
x∈Dδ x∈Dδ
=P sup (B0 (F(x)) + F(x)w) > u
x∈Dδ
⎧ ⎫
⎨ ⎬
=P sup (B0 (F(x)) + F(x)w) > u
⎩F(x)∈[ 1 −δ, 1 +δ] ⎭
2 2
⎧ ⎫
⎨ ⎬
=P sup (B0 (x) + wx) > u
⎩x∈[ 1 −δ, 1 +δ] ⎭
2 2
−2u2 +2uw
∼e , u → ∞,
STOCHASTICS 1009
implies
sup W(x) > uW(1) = w ∼ e−2u +2uw ,
2
P u → ∞. (39)
x∈[0,1]n
If (5) holds for x, y ∈ [0, 1]n , then by (38) for any u > 0
P sup W(x) > uW(1) = w = P sup (WF (x) + F(x)w) > u
x∈[0,1]n x∈[0,1]n
=P sup (B0 (F(x)) + F(x)w) > u
F(x)∈[0,1]
=P sup (B0 (x) + wx) > u
x∈[0,1]
= e−2u
2 +2uw
,
where the last equation is well known, see, e.g. [4] [Lemma 2.7].
1
Var(W(x0 ) − F(x0 )W(1)) = F(x0 ) − F 2 (x0 ) = ,
4
1010 L. BAI AND D. KALAJ
we have
J1 (u) = P sup (W(x) − F(x)W(1) + F(x)w) > u
x∈E
1
≥ P W(x0 ) − F(x0 )W(1) + w > u = (2u − w)
2
and
J2 (u) = P inf (W(x) − F(x)W(1) + F(x)w) < −u
x∈E
1
≥ P W(x0 ) − F(x0 )W(1) + w < −u = (2u + w).
2
J3 (u) = o ((2u)) , u → ∞.
1
σm2 := sup Var (W(x) − F(x)W(1)) = sup F(x)(1 − F(x)) < − δ2.
x∈E \Dδ x∈E \Dδ 4
and
J32 (u) = P inf (W(x) − F(x)W(1) + F(x)w) < −u
y∈E \Dδ
=P sup (−W(x) + F(x)W(1) − F(x)w) > u
x∈E \Dδ
≤P sup (W(x) − F(x)W(1)) > u − |w|
x∈E \Dδ
(u − |w| − Q1 )2
≤ exp − = o ((2u)) ,
2σm2
where we use the symmetry of (W(x) − F(x)W(1)) and Q1 := supx∈[0,1]n \Dδ E {W(x)−
F(x)W(1)} ∈ (0, ∞). Further, by (7)
:= sup Var (W(x) − W(y)) − (F(x) − F(y))W(1)
(x,y)∈Dδ ×Dδ
= sup F(x) + F(y) − 2F(x ∧ y) − (F(x) − F(y))2
(x,y)∈Dδ ×Dδ
≤ sup F(x) + F(y) − F 2 (x) − F 2 (y) + 2F(x)F(y) − inf 2F(x ∧ y)
(x,y)∈Dδ ×Dδ (x,y)∈Dδ ×Dδ
< 1 + 2δ − 2δ = 1,
sup! x + y − x2 − y2 + 2xy = 1 + 2δ.
1 1
(x,y)∈ 2 −δ, 2 +δ ×[ 12 −δ, 12 +δ]
J33 (u) = P sup (W(x) − F(x)W(1) + F(x)w) > u,
x∈Dδ
inf W(y) − F(y)W(1) + F(y)w < −u
y∈Dδ
=P sup (W(x) − F(x)W(1) + F(x)w) > u,
x∈Dδ
sup −W(y) + F(y)W(1) − F(y)w > u
y∈Dδ
1012 L. BAI AND D. KALAJ
≤P sup (W(x) − W(y)) − (F(x) − F(y))W(1)
(x,y)∈Dδ ×Dδ
+(F(x) − F(y))w > 2u
≤P sup (W(x) − W(y)) − (F(x) − F(y))W(1) > 2u − |w|
(x,y)∈Dδ ×Dδ
(2u − |w| − Q2 )2
≤ exp − = o ((2u)) , (40)
2
where Q2 = sup(x,y)∈Dδ ×Dδ (W(x) − W(y)) − (F(x) − F(y))W(1) < ∞. Hence, as
u→∞
2
F(x) − F(y) = (xi − yi ),
i=1
where
J1 (u) = P sup |W(x)| > uW(1) = w ,
x∈E
⎧ ⎫
⎨ ⎬
J2 (u) = P sup! |W(x)| > uW(1) = w
⎩x∈ 0, 1 +δ ×[0,1] ⎭
2
⎧ ⎫
⎨ ⎬
J3 (u) = P sup |W(x)| > uW(1) = w .
⎩x∈[0,1]× 0, 1 +δ ! ⎭
2
!
For J1 (u), by Theorem 2.1 with a1 (x, h(x)) = 1, x ∈ L1 = 12 + δ, 1 and Theorem 2.5
J1 (u) ∼ P sup W(x) > uW(1) = w + P sup W(x) > uW(1) = −w
x∈E x∈E
−2u2 +2u|w|
∼ c(4 − 8δ)u2 e
∼ 4cu2 e−2u
2 +2u|w|
, u → ∞, δ → 0
holds with c = 1 for w! = 0 and c = 2 for w = 0. Since by Theorem 2.1 with a1 (x, h(x)) =
1, x ∈ L2 = 12 , 12 + δ
⎧ ⎫
⎨ ⎬
J2 (u) ≤ P sup! W(x) > uW(1) = w
⎩x∈ 0, 1 +δ ×[0,1] ⎭
2
+P inf! W(x) < −uW(1) = w
x∈ 0, 12 +δ ×[0,1]
⎧ ⎫
⎨ ⎬
≤P sup! W(x) > uW(1) = w
⎩x∈ 0, 1 +δ ×[0,1] ⎭
2
⎧ ⎫
⎨ ⎬
+P sup! W(x) > uW(1) = −w
⎩x∈ 0, 1 +δ ×[0,1] ⎭
2
−2u2 +2uw
+ 8δu2 e−2u −2uw
2
∼ 8δu2 e
∼ o u2 e−2u +2u|w| , u → ∞, δ → 0.
2
Similarly,
J3 (u) = o u2 e−2u +2u|w| ,
2
u → ∞, δ → 0.
Thus by (42) and (43)
sup |W(x)| > uW(1) = w ∼ J1 (u) ∼ 4cu2 e−2u +2u|w| ,
2
P u → ∞.
x∈[0,1]2
1014 L. BAI AND D. KALAJ
which show that (7) holds. Thus by Theorems 2.1 and 2.5
P sup |W(x)| > uW(1) = w ∼ cKu2 e−2u +2u|w|
2
x∈[0,1]2
which show that (7) holds. Thus by Corollary 2.2 and Theorem 2.5
P sup |W(x)| > uW(1) = w ∼ P sup W(x) > uW(1) = w
x∈[0,1]n x∈[0,1]n
+P sup W(x) > uW(1) = −w
x∈[0,1]n
∼ cKu2(n−1) e−2u
2 +2u|w|
We have
n+1
P sup W(x)W(1) = w > u ≤ P sup W(x)W(1) = w > u ,
x∈[0,1]n i=1 x∈Ei (δ)
n
P sup W(x)W(1) = w > u ≥ P sup W(x)W(1) = w > u
x∈[0,1]n i=1 x∈Ei (δ)
− P sup W(x)W(1) = w > u,
1≤i<j≤n x∈Ei (δ)
sup W(x)W(1) = w > u .
x∈Ej (δ)
For x ∈ En (δ), we have F(x) = xn d + (1 − d)n−1
i=1 xi ,
1
Dn = {x ∈ En (δ) : F(x) = } = {(
x, h(
x)) :
x ∈ L} ,
2
1
L(δ) = x ∈ [0, 1]n−1 : h(
x) ≤ min xi − δ , h( x) = ,
1≤i≤n−1 2d + 2(1 − d)n−1
i=1 xi
Further, we have
P supx∈En (δ) W(x)W(1) = w > u
lim lim
u2(n−1) e−2u +2uw
2
δ→0 u→∞
n−1 n−2
i=1 xi
= (4d)(n−1) x,
n−1 d
L(δ) d + (1 − d)n−1
i=1 x i
where
1
x ∈ [0, 1]n−1 :
L= ≤ min xi .
2d + 2(1 − d)n−1
i=1 xi
1≤i≤n−1
Similarly, for i = 1, . . . , n − 1
P sup Wi (x)Wi (1) = w > u
x∈En+1 (δ)
n−2
(n−1) n−1
i=1 xi
xu2(n−1) e−2u
2 +2uw
∼ (4d) n−1 d , u → ∞.
Ln+1 (δ) d + (1 − d)n−1
i=1 xi
STOCHASTICS 1017
Since limδ→0 ∪ni=1 Ei (δ) = [0, 1]n and limδ→0 En+1 (δ) = ∅, then limδ→0 λn−1 (Ln+1 (δ))
= 0. Thus we have
P sup W(x)W(1) = w > u
x∈En+1 (δ)
n−2
(n−1) n−1
i=1 xi
xu2(n−1) e−2u
2 +2uw
≤ n(4d) n−1 d
Ln+1 (δ) d + (1 − d)n−1
i=1 xi
= o u2(n−1) e−2u
2 +2uw
, u → ∞, δ → 0.
Since
σm2 := sup Var W(x) + W(y) − (F(x) + F(y))W(1)
(x,y)∈Ei (δ)×Ej (δ)
2
= sup E W(x) + W(y) − (F(x) + F(y))W(1)
(x,y)∈Ei (δ)×Ej (δ)
!
= sup F(x) + F(y) 1 − F(x) + F(y) + 2F(x ∧ y)
(x,y)∈Ei (δ)×Ej (δ)
!
< sup F(x) + F(y) 2 − F(x) + F(y)
(x,y)∈Ei (δ)×Ej (δ)
≤ 1,
Thus we have
P sup W(x)W(1) = w > u, sup W(x)W(1) = w > u
1≤i<j≤n x∈Ei (δ) x∈Ej (δ)
2
(2u − 2w)
≤ n2 exp −
2σm2
= o u2(n−1) e−2u +2uw , u → ∞.
2
% &
Specially, when n = 2, we have L = √ 1 ,1 and
(1−d)2 +1+d
P sup W(x)W(1) = w > u
x∈[0,1]n
8d 1 d
u2 e−2u +2uw , u → ∞.
2
∼ ln 1+ −
1−d (1 − d)2 1−d
Further, if we take E = [0, 1]n and Dδ := x ∈ [0, 1]n : 12 − δ ≤ F(x) ≤ 1
+ δ with δ ∈
n 2
(0, 14 ) such that d 12 − δ + (1 − d) 12 − δ > δ, then we have
1
min xi ≥ − δ, x ∈ Dδ ,
1≤i≤n 2
and
inf F(x ∧ y) = inf d min xi ∧ yi + (1 − d)ni=1 (xi ∧ yi )
x,y∈Dδ x,y∈Dδ 1≤i≤n
n
1 1
≥d − δ + (1 − d) − δ > δ,
2 2
STOCHASTICS 1019
which show that (7) holds. Thus by Corollary 2.2 and Theorem 2.5
P sup |W(x)| > uW(1) = w
x∈[0,1]n
∼P sup W(x) > uW(1) = w + P sup W(x) > uW(1) = −w
x∈[0,1]n x∈[0,1]n
∼ cP sup W(x)W(1) = |w| > u
x∈[0,1]n
Proof of Proposition 3.5: Clearly, for F(x), (5) holds for x, y ∈ [0, 1]2 , hence the claim
follows by Theorems 2.3, 2.5 and Remark 2.6.
Acknowledgements
Thanks to Professor Enkelejd Hashorva who give many useful suggestions which greatly improved
our manuscript. Thanks to Swiss National Science Foundation Grant no. 200021-175752, National
Natural Science Foundation of China Grant no. 11901469 and The Natural Science Foundation of
the Jiangsu Higher Education Institutions of China grant no. 19KJB110022.
Disclosure statement
No potential conflict of interest was reported by the author(s).
Funding
Thanks to Swiss National Science Foundation grant no. 200021-175752, National Natural Science
Foundation of China grant no. 11901469 and The Natural Science Foundation of the Jiangsu Higher
Education Institutions of China grant no. 19KJB110022.
ORCID
Long Bai http://orcid.org/0000-0001-6947-6582
References
[1] R. Adler and L. Brown, Tail behavior for suprema of empirical processes, Ann. Probab. 14(1)
(1986), pp. 1–30.
[2] R. Adler and J. Taylor, Random Fields and Geometry. Springer Monographs in Mathematics,
Springer, New York, 2007.
[3] L. Bai, K. Dȩbicki, E. Hashorva, and L. Ji, Extremes of threshold-dependent Gaussian processes,
Science China Math. 61(11) (2018), pp. 1971–2002.
[4] W. Bischoff, F. Miller, E. Hashorva, and J. Hüsler, Asymptotics of a boundary crossing probabil-
ity of a Brownian bridge with general trend, Methodology Comput. Appl. Prob. 5 (2003), pp.
271–287.
[5] H.P. Chan and T.L. Lai, Maxima of asymptotically Gaussian random fields and moderate
deviation approximations to boundary crossing probabilities of sums of random variables with
multidimensional indices, Ann. Probab. 34(1) (2006), pp. 80–121.
1020 L. BAI AND D. KALAJ
[6] K. Dȩbicki and E. Hashorva, Approximation of supremum of max-stable stationary processes and
Pickands constants, J. Theor. Prob. 34 (2019), pp. 1–21.
[7] K. Dȩbicki, S. Engelke, and E. Hashorva, Generalized Pickands constants and stationary max-
stable processes, Extremes 20(3) (2017), pp. 493–517.
[8] K. Dȩbicki, E. Hashorva, and P. Liu, Extremes of Gaussian random fields with regularly varying
dependence structure, Extremes 20 (2017), pp. 333–392.
[9] K. Dȩbicki, E. Hashorva, and P. Liu, Uniform tail approximation of homogenous functionals of
Gaussian fields, Adv. Appl. Prob. 49 (2017), pp. 1037–1066.
[10] M. Donsker, Justification and extension of Doob’s heuristic approach to the kolmogorov-smirnov
theorems, Ann. Math. Statist 23 (1952), pp. 277–281.
[11] R.M. DuDley, Weak convergence of probabilities on nonseparable metric spaces and empirical
measures on Euclidean spaces, Illinois. J. Math 10 (1966), pp. 109–126.
[12] R.M. DuDley, Measures on non-separable metric space, Illinois. J. Math 11 (1967), pp. 449–453.
[13] M. Hogan and D. Siegmund, Large deviations for the maxima of some random fields, Adv. App.
Math. 7 (1986), pp. 2–22.
[14] V. Piterbarg, Asymptotic Methods in the Theory of Gaussian Processes and Fields, Translations
of Mathematical Monographs Vol. 148, American Mathematical Society, Providence, RI, 1996.
Appendix
Before stating the proofs of next lemmas, we introduce some notation. Define the random fields
Y(x), x ∈ Rn by
n
Y(x) = B(i) (xi ) (A1)
i=1
where B(i) s are independent standard Brownian motions. We define
√
Hc [0, λ] = E sup e 2Y(x)−Var(Y(x))
x∈[0,c∗λ]
See the recent contributions [6,7] for various results on Pickands constants.
Lemma A.1: Let h( x),x = (x1 , . . . , xn−1 ) ∈ Rn−1 is a continuous differentiable function. Assume
that X(x), x ∈ E, with E = {x : xi ∈ [Si , Ti ], i = 1, . . . , n − 1, |xn − h( x)| ≤ T}, T > 0, 0 ≤ Si <
Ti , i = 1, . . . , n − 1 is a Gaussian field which has continuous sample paths with variance function
σ 2 (x) and correlation function r(x, y) and g(x), x ∈ E is a continuous function. Further, σ 2 (x) attains
it maximum at E = {x : xi ∈ [Si , Ti ], i = 1, . . . , n − 1, xn = h(x)} which satisfies for x ∈ E
1 − σ (x)
lim sup 2
− 1 =0
(A3)
δ→0 |x−h( x)|≤δ b (xn − h( x))
x∈E\
E
and
g(x)
lim sup − 1 = 0, (A4)
x)|≤δ c (xn − h(
δ→0 |x−h( x))
x∈E\
E
for constants b > 0 and c ∈ R.
STOCHASTICS 1021
For any z ∈
E, x, y ∈ E
n
1 − r(x, y) ∼ ci xi − yi , |x − z| , y − z → 0, (A5)
i=1
where ci > 0 are constants and there exist positive constants C1 and C2 such that
n
n
C1 xi − yi < 1 − r(x, y) < C2 xi − yi , x, y ∈ E. (A6)
i=1 i=1
then as u → ∞
∞
n
e−bx
2 +cx
P sup(X(x) + g(x)) > u ∼ n−1
i=1 (Ti − S i ) i=1 ci dxu2n−1 (u)
x∈E −∞
π c2 2n−1
= n−1
i=1 (Ti − Si ) ni=1 ci e 4b u (u).
b
Proof of Lemma A.1: In the following proof, without loss of generality, we assume that c > 0.
First for ε1 , ε ∈ (0, 1) and λ > 0 we introduce the following notation:
E0 = n−1 x = (x1 , . . . , xn−1 ), E(ε1 ) = {x ∈ E : |xn − h(
i=1 [Si , Ti ], x ∈ E0 } ,
x)| ≤ ε1 ,
% & ' (
ln u kλ (k + 1)λ u ln u
E(u) = x : |xn − h( x)| ≤ ,x ∈ E0 , Jk (u) = , , M n (u) = ,
u u2 u2 λ
k = (k1 , . . . , kn−1 ), Dk (u) = ni=1 Jki (u), k ∈ Nn , Dk (u) = n−1
i=1 Jki (u), k ∈ N
n−1
,
)
n−1 )
n−1
M1 (u) = k : D(u) ⊂
k [Si , Ti ] , M2 (u) = k : D(u) ∩ k [Si , Ti ] = ∅ ,
i=1 i=1
where
0 (u) := P sup (X(x) + g(x)) > u , 1 (u) := P sup (X(x) + g(x)) > u ,
x∈E(u) x∈E\E(ε1 )
2 (u) := P sup (X(x) + g(x)) > u .
x∈E(ε1 )\E(u)
1022 L. BAI AND D. KALAJ
By (A3) and (A4), we have for any small ε ∈ (0, 1) there exists ε1 small enough such that
and
x)) − ε |xn − h(
c(xn − h( x)| ≤ g( x)) + ε |xn − h(
x) ≤ c(xn − h( x)| (A10)
hold for x ∈ E(ε1 ).
By continuity of σ (x) and E(ε1 ) ⊃
E, we have
with δ1 ∈ (0, 1), which combined with Borell-TIS inequality as in [2] leads
u−Q1 −Q2 )2
−(
1 (u) ≤ e = o ((u)) , u → ∞,
2(1−δ1 )2 (A11)
where Q1 = supx∈E\E(ε1 ) g(x) < ∞ and Q2 = E supx∈E\E(ε1 ) X(x) < ∞.
In light of (A9) and (A10), we have for u large enough
1 ln u 2
inf ≥ 1 + Q3
x∈E(ε1 )\E(u) σ (x) u
and
ln u
sup g(x) ≤ Q4 .
x∈E(ε1 )\E(u) u
By (A6), we have
2
n
E X(x) − X(y) = 2(1 − r(x, y)) ≤ 2C2 xi − yi ,
i=1
which combined with [14] [ Theorem 8.1] derive for u large enough
ln u
2 (u) ≤ P sup X(x) > u − Q4
x∈E(ε1 )\E(u) u
ln u ln u 2
≤P sup X(x) > u − Q4 1 + Q3
x∈E(ε1 )\E(u) u u
ln u ln u 2
≤ Q5 u2n u − Q4 1 + Q3
u u
= o ((u)) , u → ∞, (A12)
By (A8), (A11), (A12) and the fact 0 (u) ≥ P {X(z) > u} = (u) with z ∈
E we have
P sup(X(x) + g(x)) > u ∼ 0 (u), u → ∞. (A13)
x∈E
and
sup sup (xn − h( y))2
x))2 − (yn − h(
k∈L2 (u) x,y∈Dk (u)
≤ sup x)| + yn − h(
sup (|xn − h( y)) (xn − h( y))
x)) − (yn − h(
k∈L2 (u) x,y∈Dk (u)
λ ln u
≤ Q7 .
u3
In the view of (A9) and (A10), we notice that for any k ∈ L2 (u)
1
u+
k := sup (u − g(x))
x∈Dk (u) σ (x)
x)) + ε |xn − h(
≤ sup (u − c(xn − h( x))2 )
x)|)(1 + (1 + ε)b(xn − h(
x∈Dk (u)
2
x))) + ε sup |xn − h(
≤ u − inf (c(xn − h( x)| 1 + (1 + ε)b sup (xn − h(
x))
x∈Dk (u) x∈Dk (u) x∈Dk (u)
λ λ ln u
≤ u − inf (cy) + ε sup y + Q8 2 1 + (1 + ε)b sup (y)2 + Q9 3
y∈Jl (u) y∈Jl (u) u y∈Jl (u) u
=: u+
l ,
where l satisfies that
* +
Jl (u) ∩ inf (xn − h(
x)), sup (xn − h(
x)) = ∅. (A14)
x∈Dk (u) x∈Dk (u)
Similarly, we define
1
u−
k := inf (u − g(x))
x∈Dk (u) σ (x)
and
λ λ ln u
u−
l = u − sup (cy) − ε sup y − Q8 2 1 + (1 − ε)b inf (y)2 − Q9 ,
y∈Jl (u) y∈Jl (u) u y∈Jl (u) u3
where u−
k ≥ u−
with l satisfying (A14).
l
Considering k ∈ L2 (u), if we fix
k first, for all kn such that k ∈ L2 (u) we can chose l satisfy-
ing (A14) from −Mn (u) − 1 to Mn (u) + 1.
Bonferroni inequality leads to
0 (u) ≤ P sup (X(x) + g(x)) > u , (A15)
k∈L2 (u) x∈Dk (u)
3
0 (u) ≥ P sup (X(x) + g(x)) > u − Ai (u), (A16)
k∈L1 (u) x∈Dk (u) i=1
where
Ai (u) = P sup (X(x) + g(x)) > u, sup (X(x) + g(x)) > u
(k,l)∈Ki (u) x∈Dk (u) x∈Dl (u)
≤ P sup X(x) > u−
k, sup X(x) > u−
l , i = 1, 2, 3.
(k,l)∈Ki (u) x∈Dk (u) x∈Dl (u)
1024 L. BAI AND D. KALAJ
We set
Xu,k (x) = X(k1 u−2 λ + x1 , . . . , kn u−2 λ + xn ), x ∈ D0 (u), k ∈ L2 (u).
Then by (A5) and Lemma A.2 that
−
P sup
x∈D0 (u) Xu,k (x) > uk
lim sup − − Hc [0, λ] = 0, (A17)
u→∞ k∈L (u) (u )
2 k
as u → ∞, λ → ∞, ε → 0. Similarly,
+
P sup (X(x) + g(x)) > u ≥ P sup Xu,k (x) > uk
k∈L1 (u) x∈Dk,l (u) k∈L1 (u) x∈D0 (u)
∞
n
e−bx
2 +cx
≥ n−1
i=1 (Ti − Si ) i=1 ci dxu2n−1 (u), u → ∞, λ → ∞, ε → 0. (A19)
−∞
Next we will show that Ai (u), i = 1, 2, 3 are all negligible compared with
P sup (X(x) + g(x)) > u .
k∈L1 (u) x∈Dk (u)
For any (k, l) ∈ K1 (u), without loss of generality, we assume that k1 + 1 = l1 . Let
% &
λ √ 1
Dk (u) = k 2 , ((k + 1)λ − λ) 2 × nj=2 Jkj (u),
1
u u
% &
√ 1 λ
D2k (u) = ((k + 1)λ − λ) 2 , (k + 1) 2 , × nj=2 Jkj (u).
u u
STOCHASTICS 1025
Then
P sup X(x) > u−ε
k , sup X(x) > u−ε
l
x∈Dk (u) x∈Dl (u)
⎧ ⎫ ⎧ ⎫
⎨ ⎬ ⎨ ⎬
≤P sup X(x) > u−ε −ε −ε
k , sup X(x) > ul ⎭ + P ⎩ sup X(x) > uk ⎭ .
⎩x∈D1 (u) x∈Dl (u) x∈D2 (u)
k k
√
where λ1 = ( λ, λ, . . . , λ). Moreover, in the light of (A5) and [8] [Lemma 5.4] we have for u large
enough
⎧ ⎫
⎨ ⎬
P sup X(x) > u−ε k , sup X(x) > u −ε 4 −Q8 λ1/4
l ⎭ ≤ Q7 λ e (min(u−ε −ε
k , ul )),
⎩x∈D1 (u) x∈Dl (u)
k
holds with δ ∈ (0, 1) for (k, l) ∈ K3 (u), x ∈ Dk (u), y ∈ Dl (u). Further, Borell-TIS inequality leads to
A3 (u) ≤ P sup X(x) + X(x) > 2(u − Q11 )
(k,l)∈K3 (u) (x,y)∈Dk (u)×Dl (u)
(2u−2Q11 −Q12 )2
≤ e− 2(4−δ)
(k,l)∈K3 (u)
(2u−2Q11 −Q12 )2
≤ Qu2n e− 2(4−δ)
2n−1
=o u (u) , u → ∞, (A22)
where Q11 = supx∈E g(x) < ∞ and Q12 = 2E supx∈E X(x) < ∞.
Inserting (A18)–(A22) into (A15) and (A16) yields that
P sup (X(x) + g(x)) > u
x∈E(u)
∞
n
e−bx
2 +cx
∼ n−1
i=1 (Ti − Si ) i=1 ci dxu2n−1 (u), u → ∞,
−∞
Lemma A.2: Let Xu,k (x), k ∈ Ku , u ∈ D(u) be a family of centred Gaussian fields with continuous
sample paths where D(u) = ni=1 [0, λi u−2 ] for some λ > 0. Let further uk , k ∈ Ku be given positive
constants satisfying
u
k
lim sup − 1 = 0. (A23)
u→∞ k∈K
u
u
If Xu,k has unit variance, and correlation function rk (not depending on u) satisfying (A5) uniformly
with respect to k ∈ Ku , then
P sup
x∈D(u) Xu,k (x) > uk
lim sup − Hc [0, λ] = 0.
u→∞ k∈K (u k )
u
Proof of Lemma A.2: It follows along the same lines of [9][Theorem 2.1].
Theorem A.3: Assume that f : [0, 1]n → R is a measurable bounded positive function. Let
x1 xn
F(x1 , . . . , xn ) = ... f (t1 , . . . , tn )dt1 . . . dtn
0 0
and assume that F(1, . . . , 1) = 1. Then the projection of the level set D = {(x1 , . . . , xn ) :
F(x1 , . . . , xn ) = 1/2} on Rn−1 is a Jordan measurable set of nonzero measure.
Proof of Theorem A.3: The assumption on f tells us that there is a positive constant M so that
0 < f (x) ≤ M (A24)
for every point x ∈ [0, 1]n . Since f is defined on [0, 1]n and positive, we can extend it on Rn subject to
the condition (A24). This can be done in an explicit way by using symmetries w.r.t. to all sides of the
cube [0, 1]n . Namely, we say that two points x = (x1 , . . . , xn ), y = (y1 , . . . , yn ) ∈ Rn are equivalent if
there are integers i, k, m so that xi − yi = 2k or xi + yi = 2m + 1. Now every y ∈ Rn has its unique
representatives x in [0, 1]n . Then we set f (y) = f (x) for y ∈ [0, 1]n .
STOCHASTICS 1027
Thus B has Jordan n−1 measure equal to zero. This implies that ∂L has Lebesgue measure 0, with
respect to Lebesgues measure on Rn−1 and in particular L is measurable. Further we show that
its measure is not zero. Since F(1, . . . , 1, 1) = 1 and F(0, . . . , 0, 0) = 0, it follows that there is t ∈
(0, 1) so that F(t, . . . , t, t) = 1/2. Then h(t, . . . , t) = t, and thus by the continuity it follows that
T = (t, . . . , t) is an interior point of L, namely for
> 0 small enough so that t +
< 1, there is a
ball B centred at T and with a positive radius δ such that h(B) ⊂ (t −
, t +
). Consequently, L has
a positive Lebesgue measure establishing the proof.