Professional Documents
Culture Documents
Tema Exposicion
Tema Exposicion
Remark 6.1 The problems in Examples 6.1– 6.6 are parametric in the sense
that the distribution of the sampled population is specified. The Monte Carlo
approach here is sometimes called parametric bootstrap. The ordinary boot-
strap discussed in Chapter 7 is a different procedure. In “parametric” boot-
strap, the pseudo random samples are generated from a given probability distri-
bution. In the “ordinary” bootstrap, the samples are generated by resampling
from an observed sample. Bootstrap methods in this book refer to resampling
methods.
Monte Carlo methods for estimation, including several types of bootstrap
confidence interval estimates, are covered in Chapter 7. Bootstrap and jack-
knife methods for estimating the bias and standard error of an estimate are
also covered in Chapter 7. The remainder of this chapter focuses on hypothesis
tests, which are also covered in Chapter 8.
α = sup π(θ).
θ∈Θ0
The probability of Type I error is the conditional probability that the null
hypothesis is rejected given that H0 is true. Thus, if the test procedure is
replicated a large number of times under the conditions of the null hypothesis,
the observed Type I error rate should be at most (approximately) α.
If T is the test statistic and T ∗ is the observed value of the test statistic,
then T ∗ is significant if the test decision based on T ∗ is to reject H0 . The
significance probability or p-value is the smallest possible value of α such that
the observed test statistic would be significant.
For the Monte Carlo experiment above, the parameter estimated is a proba-
bility and the estimate, the observed Type I error rate, is a sample proportion.
If we denote the observed Type I error rate by p̂, then an estimate of se(p̂) is
p̂(1 − p̂) 0.5
se(p̂) = ≤√ .
m m
X − 500
T∗ = √ ∼ t(19),
S/ 20
n <- 20
alpha <- .05
mu0 <- 500
sigma <- 100
where the sampling distribution of the skewness statistic is derived under the
assumption of normality. √
However, the convergence of b1 to its limit distribution is rather slow and
the asymptotic distribution is not a good approximation for small to moderate
sample sizes.
Assess the Type I error rate for a skewness
√ test of normality at α = 0.05
based on the asymptotic distribution of b1 for sample sizes n = 10, 20, 30,
50, 100, and 500.
The vector of critical values cv for each of the sample sizes n = 10, 20, 30,
50, 100, and 500 are computed under the normal limit distribution and stored
in cv.
sk <- function(x) {
#computes the sample skewness coeff.
xbar <- mean(x)
m3 <- mean((x - xbar)^3)
m2 <- mean((x - xbar)^2)
return( m3 / m2^1.5 )
}
In the code below, the outer loop varies the sample size n and the inner
loop is the simulation for the current n. In the simulation, the test decisions
are saved as 1 (reject H0 ) or 0 (do not reject H0 ) in the vector sktests.
When the simulation for n = 10 ends, the mean of sktests gives the sample
proportion of significant tests for n = 10. This result is saved in p.reject[1].
Then the simulation is repeated for n = 20, 30, 50, 100, 500, and saved in
p.reject[2:6].
#n is a vector of sample sizes
#we are doing length(n) different simulations
for (i in 1:length(n)) {
sktests <- numeric(m) #test decisions
for (j in 1:m) {
x <- rnorm(n[i])
#test decision is 1 (reject) or 0
sktests[j] <- as.integer(abs(sk(x)) >= cv[i] )
}
p.reject[i] <- mean(sktests) #proportion rejected
}
> p.reject
[1] 0.0129 0.0272 0.0339 0.0415 0.0464 0.0539
The results of the simulation are the empirical estimates of Type I error rate
summarized below.
n 10 20 30 50 100 500
estimate 0.0129 0.0272 0.0339 0.0415 0.0464 0.0539
Monte Carlo Methods in Inference 167
n 10 20 30 50 100 500
estimate 0.0548 0.0515 0.0543 0.0514 0.0511 0.0479
These estimates are closer to the nominal level α = 0.05. On skewness tests
and other classical tests of normality see [58] or [270].