Professional Documents
Culture Documents
Measurement of Technical Efficiency. A Brief Survey On Parametric and Non-Parametric Techniques
Measurement of Technical Efficiency. A Brief Survey On Parametric and Non-Parametric Techniques
Francesco Porcelli
January 2009
1
Francesco Porcelli - Measurement of Technical Efficiency 2
Contents
1 Measurement of efficiency 3
1.1 Koopmans and Debreu-Farrell measure of technical efficiency . . . . . . . . . . . . . . 3
1.1.1 Output approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.1.2 Input approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.1.3 Features of Debreu-Farrell measure of technical efficiency . . . . . . . . . . . . 6
1.2 Measure of allocative efficiency . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.3 Techniques for measurement of technical efficiency . . . . . . . . . . . . . . . . . . . . 9
References 26
Francesco Porcelli - Measurement of Technical Efficiency 3
1 Measurement of efficiency
Following Lovell [1993], the productivity of a production unit can be measured by the ratio of its
output to its input. However productivity varies according to differences in production technology,
production process and differences in the environment in which production occurs. The main interest
here is in isolating the efficiency component in order to measure its contribution to productivity.
Producers are efficient if they have produced as much as possible with the inputs they have actually
employed and if they have produced that output at minimum cost [Greene, 1997]. It is important,
however, to be aware that efficiency is only one part of the overall performance; as reported in Fig-
ure 1 a complete analysis also involves the measurement of effectiveness, and the degree to which
a system achieves programmes and policy objectives in terms of outcomes, accessibility, quality and
appropriateness [Worthington and Dollery, 2000].
PERFORMANCE
Efficiency Effectiveness
Resource management
Outcomes Accessibility
Quality Appropriateness
1. allocative (or price) efficiency refers to the ability to combine inputs and outputs in optimal
proportions in the light of prevailing prices, and is measured in terms of behavioural goal of
the production unit like, for example, observed vs optimum cost or observed profit vs optimum
profit.
2. technical efficiency is measured as the ratio between the observed output and the maximum
output, under the assumption of fixed input, or, alternatively, as the ratio between the observed
input and the minimum input under the assumption of fixed output.
In the literature there are two main definitions of technical efficiency.
(a) According to Koopmans [1951] "a producer is technically efficient if an increase in an output
requires a reduction in at least one other output or an increase in at least one input, and if
Francesco Porcelli - Measurement of Technical Efficiency 4
a reduction in any input requires an increase in at least one other input or a reduction in at
least one output".
(b) Differently, Debreu [1951] and Farrell [1957] defined the following measure of technical effi-
ciency known as the Debreu-Farrell measure: "one minus the maximum equiproportion-
ate reduction in all inputs that still allows the production of given outputs, a value of one
indicates technical efficiency and a score less than unity indicates the severity of technical
inefficiency".
Finally, both technical and allocative efficiency can be measured by two main approaches:
1. the input approach if one is considering the ability to avoid waste by producing as much output
as input usage allows, i.e. we evaluate the ability to minimise inputs keeping outputs fixed;
2. the output approach if one is considering the ability to avoid waste by using as little input as
output production allows, i.e. we evaluate the ability to maximise outputs keeping inputs fixed.
It is useful, now, to derive both the output and input oriented Debreu-Farrell efficiency measures that
differ form Koopmans’ measure only in the presence of input and output slacks as we will see in Section
1.1.3.
• the production frontier (or isoquant from the output perspective) IsoqP (x) = {y : y ∈ P (x), ey ∈
/
P (x), e ∈ (1, +∞)}
• the Shephard [1953, 1970] output distance function DO (x, y) = min{e : ( ye ) ∈ P (x)}.
Figure 2 displays a piece-wise production frontier IsoqP (x) and the corresponding output set P (x)
considering a two-output (Y1 and Y2 ) one input (x) technology.
Output combinations that lie on the isoquant, for example, yc and yd , will identify fully efficient
producers. Conversely, output combinations that are inside the production frontier, for example, ya
and yb , will identify inefficient producers.
In this example the expansion of ya and yb toward the production frontier along the rays that
start from the origin corresponds to the Debreu-Farrell output-oriented technical efficiency measures,
respectively ea and eb . Then ea ya and eb yb are the projected or ideal output combinations, i.e. the
output combination that producers a and b should have attained if fully efficient.
Y2
IsoqP (x) yc
b
ea ya
P (x) b
b
ya
b
yd
b eb yb
b
yb
O Y1
• the Shephard [1953, 1970] input distance function DI (y, x) = max{ρ : ( xρ ) ∈ L(y)}
Ef f L(y) ⊆ IsoqL(y)
IsoqL(y) = {x : DI (y, x) = 1}
Francesco Porcelli - Measurement of Technical Efficiency 6
b Xb
ρb xb b
b
Xa
b
Xc
b L(y)
ρa xa
b
Xd IsoqL(y)
O X1
It is important to stress that the DF measure does not always coincide with Koopmans’s definition of
efficiency.
The DF measure is necessary but not sufficient to obtain Koopmans’s technical efficiency. For
example, in both Figure 2 and Figure 3, points eb yb and ρb xb satisfy the DF conditions but not the
Koopmans conditions. In fact, in the case of the output approach (Figure 2) it is possible to increase
the production of output Y2 without reducing output Y1 , and similarly in the case of the input approach
(Figure 3) it is possible to reduce the consumption of input X2 without increasing that of X1 . Clearly
points ρa xa and ea ya , instead, satisfy both definitions of technical efficiency.
This problem is related to the possibility of having input or output slacks. It disappears in much
econometric analysis, where the parametric form of the function used to represent production technol-
ogy (e.g. Cobb-Douglas) imposes equality between isoquants and efficient subsets.
Typically the problem may be important in the mathematical programming approach where the
nonparametric form of the frontier used to represent the boundary of production possibilities imposes
slack by assumption as a result of the piece-wise output frontier or isoquant. In this last case a possible
way to solve the problem is to report the DF technical efficiency scores and slacks separately side by
side.
Coelli et al. [2005] observed, however, that the importance of slacks can be overstated since they are
essentially an artefact of the chosen construction method and the use of a finite sample size. Moreover,
it has also been stressed that the problem of slacks can be viewed as allocative efficiency, so when the
focus of the analysis is on technical efficiency it can be ignored without problems.
• the minimum cost function (or cost frontier) c(y, w; β) = min{w′ x : DI (y, x; β) > 1}, where y
is the output.
w′ xe c(y, w; β)
=
w′ xa w′ xa
• technical efficiency corresponds to
ρa xa w′ (ρa xa )
ρa = =
xa w′ xa
1
A detailed analysis, including the output approach case, can be found in Lovell [1993].
Francesco Porcelli - Measurement of Technical Efficiency 8
cost efficiency w′ xe
= ′ (7)
technical efficiency w (ρa xa )
Figure 4 represents the decomposition of the cost efficiency in its three components in relation to
the input combination xa :
• the technical inefficiency component (T ), distance between the red and the green line;
• the slack component (S), distance between the green and the black line;
• the allocative efficiency component (A), distance between the black and the isocost line (the blue
one).
X2
L(y)
Cost efficiency
b
xe
xa
ρa xa b
b b
xb IsoqL(y)
T
S
A
O X1
For example, in Figure 4 the input combination xe lies on the isoquant and on the isocost (the
blue line) and therefore it is efficient from the allocative and consequently also from the technical
point of view. Conversely, the input combination xb , although technically efficient, exhibits some cost
inefficiency that makes it inefficient from the allocative point of view. Then we can see that xa is not
technically efficient, because it is inside the input set, and consequently it does not satisfy conditions
for allocative efficiency.
Therefore, it is clear that technical efficiency is a necessary but not sufficient condition for allocative
efficiency, while allocative efficiency implies technical efficiency because by assumption we are not
observing input combinations on the left of the isoquant.
Moreover we can also see that the correct identification of slacks is much more important in al-
locative efficiency analysis since they are one component of cost inefficiency, while as regards technical
efficiency they are just an additional piece of information.
Of course, the same conclusions would be obtained under the output approach.2
2
The output approach can be found in Lovell [1993].
Francesco Porcelli - Measurement of Technical Efficiency 9
• the econometric approach is stochastic and attempts to distinguish between the effects of noise
and the effects of inefficiency, while the linear programming approach is deterministic and under
the voice inefficiency melt noise and real inefficiency;
• the econometric approach is parametric and as a result suffers from functional form misspecifi-
cation, while the programming approach is non-parametric and so it is immune to any form of
functional misspecification.
B D
eDEA
F ≃ < ≃ eDEA
G
A+B C +D
As a result, the efficiency index of each producer is a number between zero and one, and the closer it is
to one, the more efficient is the producer. Unity will indicate a fully efficient input-output combination.
output
G2 N
M
L b
× b
b
F1
× C
G
y2
I
b
b b
A
H b
F
y1 b b
B
E b
O X1 X2
input
Since in the empirical part we are assuming that regional governments, as producers of health care
services, operate in order to maximise output given a determined amount of input, let us consider in
more detail the output-oriented approach.5
Under these assumptions DEA requires the solution of the following linear programming for each
producer in the sample:
v′ xi
min s.t.
u,v u′ yi
v′ xj (8)
> 1 j = 1, ..., i, ..., N
u′ yj
u, v > 0.
where:
The linear ratio model in (8) can, however, be converted to the following linear programming "multi-
plier" problem:
min v′ xi s.t.
u,v
u′ yi = 1
(9)
v′ xj > u′ yj j = 1, ..., i, ..., N
u, v > 0.
5
The input approach differs very little from the output approach, Ali and Seiford [1993] provides an extensive analysis
of both methodologies.
Francesco Porcelli - Measurement of Technical Efficiency 12
Finally, the linear programming problem in (10), which represents the dual version of the problem in
(9), is solved to derive the technical efficiency score of each producer eDEA
i ∈ [0, 1], i = 1, 2, ..., N .
1
In particular, let eDEA
i = , then φDEA
i solves the linear program:
φDEA
i
max φi s.t.
φ,λ
Xλ 6 xi
(10)
φi yi 6 Yλ
λ>0
where:
• i = 1, 2, ..., N ;
• λ is a (N × 1) intensity vector.
Therefore, our measure of technical efficiency, eDEA i , will correspond to the inverse of the solution to
the problem in (10), φDEA i , that represents the DEA variant of the Debreu-Farrell output measure of
technical efficiency DFo (x, y) considered in section 1.1. In this way we have that edea i ∈ [0, 1] ∀i =
1, 2, ..., N .
eDEA
i = 1 is, however, a necessary but not a sufficient condition for technical efficiency according
to Koopmans’s definition, since the ideal input-output combination (edea i yi , xi ) may contain slacks in
any of its (p + q) dimensions.6 In fact, technical efficiency according to Koopmans, as we have seen in
section 1.1, is reached when:
edea
i = 1, Xλ = xi , yi = Yλ
The problem in (10) is then solved N times in order to obtain N values of eDEA i , and N vectors λi ,
one for each producer. The optimal vector λi is a vector of weights that tell us the position of the ideal
(or projected) input-output combination of the i-th producer on the frontier.
because they are obtained by looking in different directions toward different points on the production
frontier.
Only in the case of constant return to scale there is no difference between the output and input
approach. For example, in Figure 6, under the assumption of constant return to scale the distance from
the frontier of the input-output combination (y1 , x1 ) does not change if it is measured horizontally (AC)
or vertically (AE). Conversely, in the case of variable return to scale the distance from the frontier is
greater if measured vertically (output approach) than horizontally (input approach), i.e. AD > AB.
output
variable return to scale
nonincreasing return to scale
*E
b
*D
(y2 , x2 )
b
b
G *C * (y1 , x1 )
B A
b
F Input
Assuming, for example, variable return to scale, scale efficiency can be evaluated considering the
distance between the VRS frontier and CRS frontier. In Figure 6, only the output-input combination
(y2 , x2 ) is scale-efficient. By using the output approach the scale efficiency of combination (y1 , x1 ) will,
however, correspond to the following ratio:
FD
eSE
1 =
FE
Given that technical efficiency, measured in terms of ratios, corresponds to eV1 RS = FF DA
under variable
FA
return to scale, and to eCRS
1 = FE under constant return to scale, the following result can be derived:7
eCRS
i = eVi RS × eSE
i , ∀i
Therefore, under the assumption of VRS, technical efficiency is computed without considering the
effect of the the scale component. This is the approach that will be followed in the empirical part.
Strong disposability is rarely relaxed and in the empirical part of this work this assumption will
not be relaxed.
Weak disposability is obtained by replacing:
• Xλ 6 xi with αXλ = xi
G2 N
M
L b
× b
C
F1 G
I b
y2 b
× b
H A
b
F
y1 b b
B
E b
O X1 X2
input
Figure 7: Data Envelopment Analysis, FDH one-input, one-output model when we relax the assumption
of convexity of the output or input set.
Francesco Porcelli - Measurement of Technical Efficiency 15
• f (xi ; β) is the deterministic component of the production function, where β is a vector of tech-
nology parameters;
• exp{vi } is the stochastic component of the production function which accounts for the statistical
noise in the production process, and therefore we assume that it has a symmetric distribution
with zero mean;
• finally the possibility of inefficiency is captured by the second error component ui that is assumend
to be distributed independently of vi in such a way as to satisfy the restriction ui > 0.
Following Lovell [1993] the econometric version of the Debreu-Farrell output-oriented technical effi-
ciency measure corresponds to the reciprocal of the DFO (xi , yi ) and is written:
yi
esf
i = = exp{−ui } i = 1, 2, ..., n. (12)
[f (xi ; β)exp{vi }]
where esfi ∈ (0, 1], and unity values indicate fully efficient producers.
These models allow for technical inefficiency controlling for those random shocks the affect output
and at the same time are outside the control of producers. As stressed by Kumbhakar and Lovell
[2000], the great virtue of SFM is that the impact on output (and so also on efficiency) of shocks owing
to variation in inputs can be separated from the effect of environmental variables.
When one of the two error components is taken out of the model we do not have an SFM any
longer and we end up with one of these two other broad families of models:
Francesco Porcelli - Measurement of Technical Efficiency 17
The production model will usually be linear in the logs of the variables, so for estimation purposes
the model in (11) usually becomes:
• Cobb-Douglas
K
xβk k ∀i
Y
f (xi ; β) = β0
(k=1)
• Translog
K K K
X 1 XX
f (xi ; β) = exp β0 + βk log xk + βkl log xk log xl ∀i
2
k=1 k=1 l=1
in both cases β0 is a constant term, and restrictions on the technology parameters are usually
imposed in order to ensure that f (xi ; β) is homogeneous of degree not greater than one in order to
rule out the possibility of increasing return to scale8 .
Figure 8 provides a graphical representation of the SFM. In a deterministic framework the producer
with the (y1 , x1 ) observed output/input combination would have been evaluated as less efficient than
the producer with (y2 , x2 ) output/input observed quantities (consider the distance between the dot
and the asterisk). Conversely, in a stochastic framework we are observing an opposite outcome since
v1 < 0 and v2 > 0 (consider the distance between the the × and the asterisk).
This example provides a powerful theoretical argument against the deterministic approach. As it
will be discussed in more detail in Section ??, however, the stochastic noise can contain some elements
of the environment that affect the output and can be introduced into the model among the inputs in
order to explain the reason behind a bad or good performance instead of being ascribed to a generic
ability or disability. Therefore excluding the possibility of allowing for what we can call "environment
effect", one might risk overestimating or underestimating inefficiency in a stochastic framework also.
1. Corrected ordinary least squares (COLS), first proposed by Winsten [1957] but usually
attributed to Gabrielsen [1975], does not require any assumptions about the functional form of
ui , as it estimates the technology parameters of (13) by OLS and corrects the downward bias in
the estimated OLS intercept by shifting it up until all corrected residuals are non-positive and
at least one is zero. Corrected residuals are then used in Equation (12) to estimate esf
i , that by
sf
construction satisfy the constraint 0 < ei 6 1.
8
A detailed discussion about the functional form of the production function can be found in Coelli et al. [2005]
Francesco Porcelli - Measurement of Technical Efficiency 18
f (x2 , β)exp(v1 ) ×
output
stochastic effect
b
f (x2 , β)
Inefficiency effect
y2 * f (x2 , β)exp(v2 − u2 )
f (x2 , β)
b
stochastic effect
y1 × f (x1 , β)exp(v1 )
Inefficiency effect
* f (x1 , β)exp(v1 − u1 )
O x1 x2 input
2. Modified Ordinary Least Squares (MOLS) was introduced by Richmond [1974] and requires
assumptions about the functional form of ui (half normal, truncated normal, exponential, etc.).
MOLS proceeds by estimating β of Equation (13) by OLS and estimates OLS intercept by shifting
it up by minus the estimated mean of ui , that is extracted from the moments of the OLS residuals.
Finally, the OLS residuals are modified in the opposite direction, and used to estimate esf i in
Equation (12). The pitfall of this technique is that there is no guarantee that the intercept will
be shifted up enough to cover all the observations and it will be possible, for some of them, to
have esf
i > 1, which will be very difficult to justify.
3. Maximum Likelihood Estimator (MLE) was proposed by Afriat [1972] and apparently first
used by Greene [1980] and Stevenson [1980]. This estimator assumes the distribution of ui and
it simultaneously estimates the technology parameters (β) and the moments of the distribution
of ui . The MLE envelops all observations and the residuals are inserted in Equation (12) to
estimate esf sf
i , that by construction will satisfy the constraint 0 < ei 6 1.
Lovell [1993] challenged the assumption of vi = 0 because it combines the bad features of both the
econometric approach and the linear programming approach: it makes the model deterministic and
parametric at the same time. All deviation from the frontier is attributed to technical efficiency since
no accommodation for noise is taken into consideration.
As reported in Figure 9, COLS and MOLS are further deficient because they correct only the
intercept value with respect to OLS, leaving the slope parameters (β) unchanged, and as a result the
structure of the efficient frontier technology is the same as the structure of the technology used by
less efficient producers. In other words, the COLS and MOLS frontier does not necessarily bound
the data frorm above as closely as possible, since it is required to be parallel to the OLS regression.
Therefore MOLS and COLS assign the same efficiency ranking as OLS does, the only difference being
that with MOLS and COLS the magnitude of efficiency scores becomes available. Instead, MLE allows
for structural dissimilarities between OLS and frontier technology.
Consequently, when we make distributional assumptions about the composite error term, MLE is
the most suitable estimator, and COLS seems to be appropriate only when we want to avoid distribu-
tional assumption about ui . To understand why, consider again the model in (11), where the composite
Francesco Porcelli - Measurement of Technical Efficiency 19
ln(output)
MLE
COLS
MOLS
b
b
b
b b
b
b
b
b b
OLS
b
b b b
b b
b
b
b b
b b
b
b b
b
b
b b
b b
b
b b b
b
b
O ln(input)
error (vi − ui ) is asymmetric since we impose ui > 0; then, assuming that vi and ui are distributed
independently of xi , COLS provides consistent estimates only of the slope parameters but not of the
intercept that corresponds to E(vi − ui ) = −E(ui ) 6 0. Consequently, it will not possible to compute
estimates of producer-specific technical efficiency.
Therefore among the three estimators the choice is between COLS and MLE according to the
assumptions one is prepared to make.
The assumption under which these two estimators are consistent and the shape of the likelihood
function will be considered in relation to panel data models, also because, as will be clear later, the
only possibility to estimate SFM by means of COLS is when we have a longitudinal dataset; with cross-
sectional datasets there is very little to compare since MLE is in the end the only possible estimator.
From the previous discussion we understand that when we choose the parametric approach we
need a stochastic approach as well since the assumption of vi 6= 0 is more appropriate for investigating
inefficiency caused by pure chance. On the other hand, the stochastic approach requires a wider set
of assumptions, and unlike the determinist frontier case the resulting residuals obtained from MLE
contains both noise and inefficiency, and must be decomposed in order for the distance from the frontier
measured by ui to be assessed.
The decomposition problem was first solved by Jondrow et al. [1982], who specified the functional
form of the distribution of the one-sided inefficiency component and derived the conditional distribution
(ui |vi − ui ). Either the mode or the mean of this distribution provides a point estimate of ui , which
can be inserted in Equation (12) to obtain estimates of esf i . We are going to describe this procedure
in detail for the case of panel data models.
distinguishd: fixed effect models (FE) and random effect models (RE). For each of them different
estimators can be used to estimate consistently the technology parameters as well as the individual
specific inefficiency terms.
The general model in (11) becomes:
a3 E[vit |vis ] = 0 ∀t 6= s, independence of the idiosyncratic disturbances across time, which implies
a dynamically correctly specified model;
a4 E[yit |xi1 , xi2 , ..., xit , ui ] = f (xit ; β) ∀i which implies that f (xit ; β) is correctly specified and that
the regressors are weakly exogenous;
a6 he asymptotic properties are in principle based on both dimensions of the panel, N → ∞ and
T → ∞.
To simplify the discussion we are assuming that a Cobb-Douglas production function, with non-
neutral technological change,9 correctly specifies the production function, and finally we consider a
linear transformation of the model in (14) using a log-log functional form. As a result the general form
of the empirical model becomes:
K
X K
X
yit = β0 + τ t + δk (tx)k + βk xk + vit − ui (15)
(k=1) (k=1)
PK
where i = 1, 2, ..., I and t = 1, 2, ..., T , variables are expressed in natural logs, and (k=1) βk 61
is imposed in order to rule out the possibility of having increasing return to scale.
where β0i = β0 − ui .
9
The non-neutral technological change is introduced through a linear trend and its interaction with the inputs;
moreover, the Cobb-Douglas specification implicitly assumes that the technological change is constant over time.
Francesco Porcelli - Measurement of Technical Efficiency 21
Then to obtain a consistent estimate of β0i , and also to avoid an incidental parameter problem,
we need to restrict assumption a6, so the asymptotic properties of the model will be based only on
T → ∞ with N fixed.
After the estimation we employ the normalisation
(LSDV ) (LSDV )
β0 = max{β0i } (17)
and ui are estimated from
(LSDV ) (LSDV )
β0 − β0i (18)
that ensure that all ui > 0.
Finally, producer-specific estimates of technical efficiency are then obtained by
As underlined by Schmidt and Sickles [1984] as T → ∞ we can consistently separate the overall
intercept from the one-sided individual effect, which provides only the measure of efficiency relative to
an absolute standard represented by the most efficient producer in the sample that will have eFi E = 1.
The efficiency of the most efficient producer in the sample, however, will approach one, as N → ∞.
Therefore with LSDV we obtain only a consistent "relative" measure of efficiency.
If assumption a6 holds, estimates of ui are fully consistent but the incidental parameter problems
make LSDV infeasible so we have to use WG, which corresponds to COLS applied to the transformation
of the model where all data are expressed in terms of deviation from producers’ means and the ui are
recovered as means of producers’ residuals.
T K K
1X (W G)
X (W G)
X (W G)
u∗i = yit − β0 − δk (tx)k − βk xk
T (20)
t=1 (k=1) (k=1)
ui =max{u∗i } − u∗i
The main advantage of the FE model is its robustness and consistency since we do not need any
assumption about ui and therefore this model can be assumed as a sort of benchmark for the others;
this comes at a price, however: time invariant variables are not allowed in the model; as stressed
by Kumbhakar and Lovell [2000], the fixed effects (the ui ) are capturing not only variation in time-
invariant technical efficiency across producers, but unfortunately will also capture all the time invariant
heterogeneity.10
a7 E[ui |xi1 , xi2 , ..., xiT ] = 0, independence between the inefficiency error term and the inputs;
a8 ui ∼ i.i.d.(µ, σu2 ), we do not need to specify the distribution of the inefficiency error term but we
have to estimate its variance;
a9 E[vit |ui ] = 0 ∀i and ∀i, independence between the two error components.
10
This is, however, a common problem with all panel data SFM. This issue will be discussed in detail in section 3.2.3.
Francesco Porcelli - Measurement of Technical Efficiency 22
Under assumptions a1 - a9, Feasible Generalised Least Squares (FGLS) is a consistent esti-
mator of the technology parameters considering the following transformation of the original model in
(15):
K
X K
X
yit =[β0 − E(ui )] + τ t + δk (tx)k + βk xk + vit − [ui − E(ui )] =
(k=1) (k=1)
(21)
K
X K
X
=β0∗ + τ t + δk (tx)k + βk xk + vit − u∗i
(k=1) (k=1)
We can recover estimates of the specific-producer inefficiency u∗i from the residuals through the
same methodology used in the case of FE:
T K K
1X ∗(F GLS)
X (F GLS)
X (F GLS)
u∗i = yit − β0 − δk (tx)k − βk xk
T t=1 (22)
(k=1) (k=1)
ui =max{u∗i } − u∗i
Under assumptions a1 to a9 and specifying a distribution for ui such that the ui > 0 and a
symmetric distribution for vit , it will be possible to estimate the parameters of the model in (15)
with Maximum Likelihood Estimator (MLE); then the estimates of the technical efficiency of
each producer are obtained in a second step considering the conditional distribution of ui given the
composite error term (vit − ui ) using Jondrow et al. [1982] decomposition approach (JLMS from the
authors of the paper).
Following Pitt and Lee [1981], who first proposed MLE in the context of panel data SFM, in order
to write the complete log-likelihood function, the preliminary step is to assume a distribution g(u) for
the inefficiency error term, then we have to define the composite error structure ǫit = (vit + ui ). At this
stage it is possible to write the joint density of ǫit that corresponds to the "unconditional likelihood
function":
Z T
∞Y
h(ǫi1 , ǫi2 , ..., ǫiT ) = (ǫit − ui )g(ui )dui (23)
0 t=1
Given this density, the complete log-likelihood function is:
N
X
log L(τ, δ, β, σv2 , σu2 ; data) = log h(ǫi1 , ǫi2 , ..., ǫiT ) (24)
i=1
According to the assumption about the distribution of the error terms, at least four different
models can be found in the literature on SFM: the Normal-Half Normal (NHN), Normal-Exponential
(NE), Normal-Truncated Normal (NTN), and Normal-Gamma (NG). Clearly, the distribution of the
idiosyncratic error is always normal whereas the distribution of the inefficiency error term is chosen in
order to compel ui to be non-negative.
As reported by Kumbhakar and Lovell [2000] and Coelli et al. [2005], theoretical consideration
should guide the choice of the distributional specification. For example, the Normal-Half Normal and
Francesco Porcelli - Measurement of Technical Efficiency 23
Normal-Exponential models are appropriate when most inefficiency effects are assumed to be in the
neighbourhood of zero, implying that the modal value of technical efficiency is one with decreasing
values of technical efficiency becoming increasingly less likely. The Normal-Truncated Normal and
Normal-Gamma are more flexible since the modal efficiency value can also be away from one, and for
this reason in most empirical works the NTN is usually preferred to the NHN. Unfortunately, this
sort of flexibility comes at the cost of more computational complexity, and if the probability density
functions of ui and vit are similar it will become more difficult to distinguish inefficiency effects from
noise. Finally, the different distributional assumptions will produce different predictions of technical
efficiency but the rankings of producers on the basis of the predicted technical efficiency have been
found to be quite robust to distributional assumption choice.
Since the NE and the NG models are rarely used in empirical works, and the NHN can be seen as
a particular version of the NTN model, we can consider in more detail this last model.11
In the Normal-Truncated Normal model assumptions a8 and a1 become less general since ui ∼
i.i.d.N + (µ, σu2 ). The density function of the truncated normal variable u is:
(u − µ)2
2
g(u) = exp − , ,u > 0 (25)
(2π)1/2 σu Φ(µ/σu ) 2σu2
After the estimation of the structural parameters through the maximisation of the log-likelihood
function shown in (24), we need a second stage to estimate the producer-specific time-invariant technical
efficiency terms using the JLMS decomposition method. The first step is to derive the conditional
distribution of u given u given ǫ:
QT
t=1 (ǫit − ui )
g(u|ǫ) = R ∞ QT =
0 t=1 (ǫit − ui )g(ui )dui
" # (26)
1 (µ − µ∗ )2
= exp −
(2π)1/2 σ∗ [1 − Φ(−µ∗ /σ∗ )] 2σ∗2
µσv2 − t ǫt σu2
P
∗
• µ = ;
σv2 + T σu2
σu2 σv2
• σ∗2 =
σv2
+ T σu2
As a result, either the mean or the mode of this distribution can be considered as a point estimate of
the specific-producer inefficiency ui .
As reported in Kumbhakar and Lovell [2000], the mean is used more frequently than the mode.
Finally, once point estimates of ui are obtained, estimates of the technical efficiency score of each
producer can be derived as in the case of FE models:
eRE
i = exp(−ui ) (29)
Considering other distributions for the inefficiency error term involves the same processes to obtain
ui , and of course the NTN requires more computations, essentially the price we pay for its flexibility.
The main advantage, and in a certain sense also the main reason, of considering RE instead of
FE models is the possibility of having time invariant inputs among the regressors. Moreover, under
the assumptions a1 to a9 RE models are more efficient than FE models but, as always stressed in the
classical panel data literature, if some of the three extra assumptions (a7-a9 ) required by RE models
will not hold, either FGLS or MLE is not consistent.
The assumption of independence between the inefficiency error term and the inputs is indeed very
strong and should be tested through the Hausman test or its robust version in case we do not have
conditional homoscedasticity and/or correct dynamic specification. Similarly, the joint hypothesis that
the effects are uncorrelated with the regressors and that the distributional assumptions are correct can
be tested through an Hausman test based on the difference between MLE and WG estimators.
In general the nature of our dataset should guide us in the choice of the most suitable estima-
tor among WG/LSDV, FGLS and MLE. In general, in long panels, where T → ∞ and N is fixed,
WG/LSDV is the better choice, but in short panels, where N → ∞ and T is fixed, FGLS is clearly
preferred, provided that independence between the inefficiency error term and the regressor holds.
Then, at the cost of more computations, if we are prepared to make distributional assumptions,
MLE is more efficient (and consistent either over N or over T ). A strong argument against MLE in the
case of short panels, however, is that as N increases the variance of the conditional mean or mode of
g(ui |ǫi ) for each producer does not go to zero, and technical efficiency of producers cannot be estimated
consistently.
WG and LSDV are still unbiased and either the technology parameters or the specific-producers
intercepts can be consistently estimated under the same set of assumptions from a2 to a5. In this case,
robust estimators for the variance-covariance matrix like White Covariance Estimator are required for
hypothesis testing.
An alternative to the WG/LSDV approach that accounts for heteroscedasticity in v is the FGLS
approach, which does not require distributional assumptions for the error components and is more
efficient. Of course, we need to be sure that assumptions a7-a9 hold.
In the case of RE models we can also have heteroscedasticity in both error components when both
assumptions a1 and a7 are not satisfied. In this case FGLS in is not practical, as well as MLE, when N
is large because in that case we have to estimate N variances causing an incidental parameter problem.
A possible solution, as suggested by Kumbhakar and Lovell [2000], is to use the method of moments
estimator.
Therefore, in the case of FE models the failure of the conditional homoscedasticity assumption is
much less problematic than in RE models since it is related only to the idiosyncratic error component
and for this reason one can cope with it easily. This is, obviously, another point in favour of FE models.
Francesco Porcelli - Measurement of Technical Efficiency 25
The failure of weak exogeneity assumption in case of endogenous inputs In the case of FE
models, assumption a4 fails when the deterministic component of the production function is not cor-
rectly specified because there are omitted inputs or some measurement errors in the inputs. Collection
of better data is the first way to solve the problem. If data cannot be improved then the only solution
is to use some instrumental variables for the endogenous inputs.
In the case of RE models, the problem is twofold. In fact, apart from the failure of assumption a4,
assumption a7 also may not hold when independence between the inefficiency error component and
inputs is not verified. To re-establish assumption a4 one can use the same methods valid in the case
of FE models, and to re-establish assumption a7 one can use the Mundlak [1978] and Chamberlain
[1980] approach or, under more strict assumptions, it is possible to use the Hausman and Taylor [1981]
method.
Again, FE models seem to be less problematic than RE models.
ui = δ0 + d′i δ (31)
The incidental parameter problem and the possible overspecification of the model are the two
main problems in this approach. We need T → ∞. As reported by Greene [2003], the incidental
parameter problem is a persistent bias that arises in nonlinear fixed effects models when the number
of periods is small (five is small). This bias has been widely documented for binary choice models
but not systematically examined in SFM. Greene [2005a] uses a Monte Carlo simulation to show that
the incidental parameter problem is much less serious that one might expect: the coefficient appears
biased but far less so than in the binary choice models, and also the estimated inefficiencies appear to
be only slightly biased.
There is, however, another big problem related to this new formulation of the FE model. Now all
factors that are time invariant will not affect the inefficiency component of the model at all. This is
an extreme solution that generates, in some senses, a diametrically opposed new problem. Therefore
an effective solution to disentangle inefficiency from unobserved heterogeneity still requires further
research.
Francesco Porcelli - Measurement of Technical Efficiency 26
References
Afriat, S. N. (1972). Efficiency estimation of production functions. International Economic Review,
13(3):568–598.
Aigner, D. L., Lovell, C. K., and Schmidt, P. (1977). Formulation and estimation of stochastic frontier
production function models. Journal of Econometrics, 6(1):21–37.
Ali, A. I. and Seiford, L. M. (1993). The mathematical programming approach to efficiency analysis. In
Fried, A. O., Lovell, A. K., and Schmidt, S. S., editors, The Measurement of Productive Efficiency,
chapter 3, pages 120 – 159. Oxford University Press.
Chamberlain, G. (1980). Analysis of covariance with qualitative data. Review of Economic Studies,
47:225–238.
Charnes, A., Cooper, W. W., and Rhodes, E. (1978). Measuring the efficiency of decision-making
units. European Journal of Operational Research, 2(6):429–444.
Coelli, T. J., Rao, D. S. P., O’Donnell, C. J., and Battese, G. E. (2005). An Introduction to Efficiency
and Productivity Analysis. Spinger, second edition.
Deprins, D., Simar, L., and Tulkens, H. (1984). Measuring labor-efficiency in post office. In North-
Holland, editor, The Performance of Public Enterprises. M. Marchand and P. Pestieau and H.
Tulkens, Amsterdam.
Farrell, M. J. (1957). The measurement of productive efficiency. Journal of the Royal Statistical Society,
120(3):253–281.
Gabrielsen, A. (1975). On estimating efficient production functions. Chr. Michelsen Institute, Depart-
ment of Humanities and Social Science, (Working Paper No. A-35).
Greene, W. H. (1997). Frontier production functions. In Pesaran, M. H. and Shmidt, P., editors,
Handbook of Applied Econometrics, volume II: Microeconometrics. Blackwell Publishers Ltd.
Greene, W. H. (2005a). Fixed and random effects in stochastic frontier models. Journal of Productivity
Analysis, 23:7–32.
Greene, W. H. (2005b). Reconsidering heterogeneity in panel data estimators of the stochastic frontier
model. Journal of Econometrics, 126:269–303.
Hausman, J. A. and Taylor, W. E. (1981). Panel data and unobservable individual effects. Economet-
rica, 49(1377-1399).
ISSiRFA-CNR (1982-2008). La finanza regionale - anni vari. In Buglione, E., editor, Osservatorio
Finanziario Regionale. Franco Angeli, Milano.
ISTAT (2006). Conti ed aggregati economici delle Amministrazioni pubbliche Serie SEC95 - anni
1980-2005. Roma. www.istat.it.
Francesco Porcelli - Measurement of Technical Efficiency 27
Jondrow, J., Lovell, C. K., Materov, I. S., and Shmidt, P. (1982). On the estimation of technical inef-
ficiency in the stochastic frontier production function model. Journal of Econometrics, 19(2/3):233
– 238.
Kumbhakar, S. C. and Lovell, C. K. (2000). Stochastic Frontier Analysis. Cmbridge University Press.
Lovell, C. K. (1993). Production frontiers and productive efficiency. In Fried, A. O., Lovell, A. K., and
Schmidt, S. S., editors, The Measurement of Productive Efficiency, chapter 1, pages 3 – 67. Oxford
University Press.
Meeusen, W. and van den Broeck, J. (1977). Efficiency Estimation from Cobb-Douglas Production
Functions with Composed Error. International Economic Review, 18(2):435–444.
Mundlak, Y. (1978). On the pooling of time-series and cross-section data. Econometrica, 46:69–86.
Pestieau, P. and Tulkens, H. (1990). Assessing the performance of public sector activities: Some recent
evidence form the productive efficiency viewpoint. Discussion Paper No. 9060, CORE, Universitè
Catholique de Louvain.
Pitt, M. M. and Lee, L.-F. (1981). The Measurement and Sources of Technical Inefficiency in the
Indonesian Weaving Industry. Journal of Development Economics, 9:43–64.
Schmidt, P. and Sickles, R. C. (1984). Production frontiers and panel data. Journal of Business
Economic Statistics, 2(367-374).
Shephard, R. W. (1970). Theroy of Cost and Prodution Functions. Princeton University Press.
Stevenson, R. E. (1980). Likelihood functions for generalized stochastic frontier estimations. Journal
of Econometrics, 13(1):58–66.
Winsten, C. B. (1957). Discussion on Mr. Farrell’s Paper. Journal of the Royal Statistical Society,
Series A(120):282 – 284.