Statistical Estimation Methods in Hydrological Engineering
Statistical Estimation Methods in Hydrological Engineering
Introduction
In designing civil engineering structures use is made of probabilistic calculation
methods. Stress and load parameters are described by statistical distribution functions.
The parameters of these distribution functions can be estimated by various methods.
An extensive comparison of these different estimation methods is given in this paper.
The main point of interest is the behaviour of each method for predicting p-quantiles
(the value which is exceeded by the random variable with probability p), where p«1.
The estimation of extreme quantiles corresponding to a small probability of
exceedance is commonly required in the risk analysis of hydraulic structures. Such
extreme quantiles may represent design values of environmental loads (wind, waves,
snow, earthquake), river discharges, and flood levels specified by design codes and
regulations (TAW, 1990). In this paper the performance of the parameter estimation
methods with respect to its small sample behaviour is analyzed with Monte Carlo
simulations, added with mathematical proofs.
In civil engineering practice many parameter estimation methods for
probability distribution functions are in circulation. Well known methods are for
example:
- the method of moments (Johan Bernoulli, 1667-1748),
- the method of maximum likelihood (Daniel Bernoulli, 1700-1782),
- the method of least squares (on the original or on the linearized data), (Gauss, 1777-
1855),
- the method of Bayesian estimation (Bayes, 1763),
- the method of minimum cross entropy (Shannon, 1949),
- the method of probability weighted moments (Greenwood et al., 1979),
- the method of L-moments (Hosking, 1990).
Textbooks, such as Benjamin and Cornell (1970), Berger (1980), treat the
traditional methods in detail. The methods will be briefly reviewed in this paper.
Many attempts (for instance, Goda and Kobune, (1990), Burcharth and Liu, (1994),
Yamaguchi, (1997), and Van Gelder and Vrijling, (1997a)), have been made to find
out which estimation method is preferable for the parameter estimation of a particular
probability distribution in order to obtain a reliable estimate of the p-quantiles. In this
paper, we will in particularly investigate the performance of the parameter estimation
method with respect to three different criteria; (i) based on the relative bias and root
mean squared error (RMSE), (ii) based on the over- and underdesign.
It is desirable that the quantile estimate be unbiased, that is, its expected value
should be equal to the true value. It is also desirable that an unbiased estimate be
efficient, i.e., its variance should be as small as possible. The problem of unbiased
and efficient estimation of extreme quantiles from small samples is commonly
encountered in the civil engineering practice. For example, annual flood discharge
data may be available for past 50 to 100 years and on that basis one may have to
estimate a design flood level corresponding to a 1,000 to 10,000 years return period
(Van Gelder et al., 1995).
The first step in quantile estimation involves fitting an analytical probability
distribution to represent adequately the sample observations. To achieve this, the
distribution type should be judged from data and then parameters of the selected
distribution should be estimated. Since the bias and efficiency of quantile estimates
are sensitive to the distribution type, the development of simple and robust criteria for
fitting a representative distribution to small samples of observations has been an
active area of research. In this paper three different methods for the selection of the
distribution type will be reviewed, extended and tested. The first method is based on
Bayesian statistics, the second one on linear regression, and the third one on L-
moments. Certain linear combinations of expectations of order statistics, also referred
to as L-moments by Hosking (1990), have been shown to be very useful in statistical
parameter estimation. Being a linear combination of data, they are less influenced by
outliers, and the bias of their small sample estimates remains fairly small. A measure
of kurtosis derived from L-moments, referred to as L-kurtosis, was suggested as a
useful indicator of distribution shape (Hosking, 1992).
Hosking (1997) proposed a simple but effective approach to fit 3-parameter
distributions. The approach involves the computation of three L-moments from a
given sample. By matching the three L-moments, a set of 3-parameter distributions
can be fitted to the sample data. In this paper, a distribution type selection which is
based on the the 4th L-moment is suggested to be the most representative distribution,
which should be used for quantile estimation. In essence, the L-kurtosis, which is
related to the 4th L-moment, can be interpreted as a measure of resemblance between
two distributions having common values of the first three L-moments.
The concept of probabilistic distance or discrimination between two
distributions is discussed in great detail in modern information theory (Kullback 1959,
Jumarie 1990). Mathematically sound measure of probabilistic distance, namely, the
divergence, has been used to establish resemblance between two distributions or
conversely to select the closest possible posterior distribution given an assumed prior
distribution. The divergence is a comprehensive measure of probabilistic distance,
since it involves the computation of departure of a distribution from the reference
parent distribution over an entire range of the random variable. Apart from the
performance of estimation methods based on bias and RMSE, and the performance
based on under- and overdesign, is suggested in this paper.
Furthermore, this paper will focus on evaluating the robustness of the L-
kurtosis measure in the distribution selection and extreme quantile estimation from
small samples. The robustness is evaluated against the benchmark estimates obtained
from the information theoretic measure, namely, the divergence. For this purpose, a
series of Monte Carlo simulation experiments were designed in which probability
distributions were fitted to the sample observations based on L-kurtosis and
divergence based criteria, and the accuracies of quantile estimates were compared.
The simulation study revealed that the L-kurtosis measure is fairly effective in
quantile estimation.
Finally, this paper shows some analytical considerations concerning statistical
estimation methods and probability distribution functions. The paper ends with a
discussion.
2
Classical estimation methods
To make statements about the population on the basis of a sample, it is important to
understand in what way the sample relates to the population. In most cases the
following assumptions will be made:
1. Every sample observation x is the outcome of a random variable X which has an
identical distribution (either discrete or continuous) for every member of the
population;
2. The random variables X1, X2, ..., Xn corresponding to the different members of the
sample are independent.
E(T(X)) = 2 (0.1)
3
and var(T) = E(T2(X)) – E2(T(X)) = E(T2(X)) – (B(T) + 2)2 = E(T2(X)) – B2(T) -
22B(T) - 22 . This proves the equality.
It is difficult to trace back who introduced the MOM, but Johan Bernoulli (1667-
1748) was one of the first who used the method in his work. With the MOM, the
moments of a distribution function in terms of its parameters are set equal to the
moments of the observed sample. Analytical expressions can be derived quite easily
(Appendix), but the estimators can be biased and not efficient. The moment estimators
however, can be very well used as a starting estimation in an iteration process.
z
The central moments of a distribution are given by:
µ r = E ( X − µ ) r = ( x − µ ) r f X ( x ) dx
Variance: σ = µ 2
2
µ3 (0.5)
Skewness: β 1 =
µ 32/ 2
µ4
Kurtosis: β 2 =
µ 22
The sample moments are given by:
x = n −1Σxi
(0.6)
mr = n −1Σ( xi − x )r
The sample mean x is a natural estimator for :. The higher sample moments mr are
reasonable estimators of the :r , but they are not unbiased. Unbiased estimators are
4
often used. In particular F2, :3 and the fourth cumulant 64=:4 - 3:22 are unbiasedly
estimated by:
s2 = ( n − 1)−1 Σ( xi − x )2
n2
m3* = m3 (0.7)
( n − 1)( n − 2 )
n2 RSn +1 UV
k4* =
T
( n − 2 )( n − 3) n − 1
m4 − 3m22
W
The sample standard deviation s = s2 is an estimator of F but is not unbiased. The
sample estimators of CV (Coefficient of Variation), skewness and kurtosis are
respectively:
C$ v = s / x
g = m3* / s3 (0.8)
k = k4* / s 4 + 3
Finding theoretical moments as a function of 2 is not easy for all probability
distributions. The method is difficult to generalize to more complex situations
(dependent data, covariates, non-identically distributed data). Sample covariances
may be used to estimate parameters that determine dependence. For some
distributions (such as Cauchy), moments may not exist. In the Appendix moments are
given for most familiar PDFs.
Also with the MML it is difficult to say who discovered the method, although Daniel
Bernoulli (1700-1782) was one of the first who reported about it (Kendall, 1961). The
likelihood function gives the relative likelihood of the obtained observations, as a
function of the parameters 2:
L(2,x)=Af(xi , 2) (0.9)
With this method one chooses that value of 2 for which the likelihood function is
maximized. In the Appendix an overview is given of the likelihood functions of some
familiar distribution functions. The ML-method gives asymptotically unbiased
parameter estimations and of all the unbiased estimators it has the smallest mean
squared error. The variances approach asymptotically to:
Furthermore these estimators are invariant, consistent and sufficient. For the
definitions we refer to Hald (1952). Analytical expressions for the parameter
estimators are sometimes difficult to derive. In those cases, numerical optimization
routines have to be used to determine the maximum of the likelihood function, which
can also be quite difficult since the optimum of the likelihood function can be
extremely flat for large sample sizes. Optimization of the likelihood function may also
be hampered by the presence of local maxima. Furthermore:
- MML is (usually) straightforward to implement,
- Maximum likelihood estimators (MLEs) may not exist, and when they do, they may
not be unique or give a bias error (Koch, 1991),
5
- MLE may give inadmissible results (Lundgren, 1988),
- The likelihood function can be used for much more than just finding the MLE:
values close to the MLE are more plausible than those further away, for example. This
argument can be used to obtain an interval [pL,pU] which comprises a plausible range
of values for 2,
- ML is adaptable to more complex modeling situations, because the MLE satisfies a
very convenient invariance property: If q=h(2) where h is a bijective function then
qML=h(2ML). For example, if q=1/2, then qML=1/2ML. So, having found the MLE for
one parameterization, the MLEs for other parameterizations are immediate,
- The maximum likelihood estimator is unbiased, fully efficient (in that it achieves the
Cramér-Rao bound under regularity conditions), and normally distributed; all of them
in asymptotical sense. Regularity conditions are not fulfilled if the range of the
random variable X depends on unknown parameters as is the case for many
distributions in the present paper.
The MML is extremely useful since it is often quite straightforward to evaluate from
the MLE and the observed information. Nonetheless it is an approximation, and
should only be trusted for large values of n (though the quality of the approximation
will vary from model to model).
If the available sample sizes are large, there seems little doubt that the
maximum-likelihood estimator is a good choice. It should be emphasized, however,
that the properties above are asymptotic (large n), and better estimators may be
available when sample sizes are small.
Least Squares were introduced by Gauss (1777-1855). Given the observations x=(x1,
x2, ..., xn) and y=(y1, y2, ..., yn), a regression model can be fitted. For the general case:
with F2 the assumed constant variance of Y around its regression line, the parameter
estimates are:
"* = mY - $* mX
$* = sXY / sX2 (0.12)
F2* = n/(n-2) (1-rXY2)sY2
in which mX, mY, sX, sY, sXY and rXY are defined as:
mX = 1/n Ei=1..nxi
mY = 1/n Ei=1..nyi
sX2 =1/n Ei=1..n(xi-mX)2 (0.13)
sY2 =1/n Ei=1..n(yi-mY)2
sXY = 1/n Ei=1..n(xi-mX)(yi-mY)
rXY = sXY/sXsY
The estimators "* and $* are linear functions of the Yi ‘s and they are unbiased. Their
variances are:
6
FA2 = F2/n (1+mX2/sX2)
FB2 = F2 / nsX2 (0.14)
in which A and B are respectively, "* and $* treated as random variables.
With the above regression techniques, the MLS can be defined. Assume that
we have n observations in sorted order given by x1:n, x2:n, ..., xn:n. Define the plotting
position pi = i/(n+1) of the i-th observation. We want to estimate the optimal value 2
of the distribution function F(x|2) in least squares sense. The following options are
available:
Assume that we can linearize the probability distribution F(x) under consideration
(with scale parameter B and location parameter A). So we can find a function g s.t.:
For instance, for the Exponential distribution g is given by gE(.) = -ln(1-.) and for the
Gumbel distribution, we have gG(.) = -ln(-ln(.)) (see Appendix for a complete
overview of linearization functions).
In Van Gelder (1996b), several LS-methods are applied to wave data at the location of
Pozzallo in Southern Italy. It can be shown that the parameter estimation of the scale
of various 2-parameter distributions with LS always leads to larger p-quantiles than
with the method of moments. Furthermore, note that under the assumption that the
errors in a regression model are independent and normally distributed with zero mean
and a fixed standard deviation, the ML-estimators of the error-distribution are exactly
the same as the LS-estimators (Gauss, 1777-1855). Problems with LS-estimators are
described by McCuen et al. (1990), in which they showed that logarithmic
transformations may lead to a biased model.
Method of L-Moments
Hosking, (1990), introduced the L-moments. They have become popular tools for
solving various problems related to parameter estimation, distribution identification,
and regionalization. It can be shown that L-moments are linear functions of
probability weighted moments (PWM's) and hence for certain applications, such as
the estimation of distribution parameters, serve identical purposes (Hosking, 1986). In
other situations, however, L-moments have significant advantages over PWM's,
notably their ability to summarize a statistical distribution in a more meaningful way.
Since L-moment estimators are linear functions of the ordered data values, they are
7
virtually unbiased and have relatively small sampling variance. L-moment ratio
estimators also have small bias and variance, especially in comparison with the
classical coefficients of skewness and kurtosis. Moreover, estimators of L-moments
are relatively insensitive to outliers. These often-heard arguments in favor of
estimation of distribution parameters by L-moments (or PWM's) should, nevertheless,
not be accepted blindly. In for instance a wave height frequency analysis, the interest
is the estimation of a given quantile, not in the L-moments themselves. Although the
latter may have desirable sampling properties, the same does not necessarily apply to
a function of them, such as a quantile estimator. In fact, several simulation studies
have demonstrated that for some distributions, other estimation methods may be
superior in terms of mean square errors of quantile estimators (Hosking and Wallis,
1987; Rosbjerg et al., 1992). As compared with for example the classical method of
moments, the robustness vis-à-vis sample outliers is clearly a characteristic of
L-moment estimators. However, estimators can be “too robust” in the sense that large
(or small) sample values reflecting important information on the tail of the parent
distribution are given too little weight in the estimation. Hosking (1990) assessed that
L-moments weigh each element of a sample according to its relative importance.
In this section first the theory of L-Moments will be briefly described,
followed with an overview of papers with applications of L-moments. The literature
review has shown that the theory of L-moments have mostly been applied in a
regionalized setting combining data from more than one site. However, in univariate
settings the method of L-moments has not been investigated so much. Therefore
furtheron in this paper a Monte Carlo experiment is designed in a univariate setting in
order to compare the L-moments method with the classical parameter estimation
methods (MOM, MML, and MLS). The performance of these methods will also be
analyzed w.r.t. inhomogeneous data.
Finally, L-moments are in fact nothing else than summary statistics for probability
distributions and data samples. They are analogous to ordinary moments -- they
provide measures of location, dispersion, skewness, kurtosis, and other aspects of the
shape of probability distributions or data samples -- but are computed from linear
combinations of the ordered data values (hence the prefix L). Hosking and Wallis
(1997) give an excellent overview on the whole theory of L-Moments.
n
b0 = n −1 ∑ x j:n
j =1
br = n −1 b j − 1gb j − 2g...b j − r g
n
∑ b n − 1gb n − 2g...bn − r g x j :n
(0.16)
j = r +1
8
x1:3 , x2:3 and x3:3. The difference between the two observations x2:3 – x1:3 and the
difference between the two observations x3:3 – x2:3 can be subtracted from eachother to
have a measure of the skewness of the distribution. This leads to: x3:3 – x2:3 - (x2:3 –
x1:3) = x3:3 – 2x2:3 + x1:3 . A sample of size 4 contains four observations in ascending
order x1:4 , x2:4 , x3:4 and x4:4. A measure for the kurtosis of the distribution is given by:
x4:4 – x1:4 - 3 (x3:4 – x2:4). In short: the above linear combinations of the elements of
the ordered sample contain information about the location, scale, skewness and
kurtosis of the distribution from which the sample was drawn. A natural way to
generalize the above approach to samples of size n, is to take all possible sub-samples
of size 2 and to average the differences (x2:2 – x1:2)/2:
FG IJ
1 n
−1
∑ ∑ ( xi:n − x j:n )
l2 =
2 2HK i> j
(0.17)
1 F nI
−1
= G J ∑ ∑ ∑ ∑(x
4 H 4K
l4 i:n − 3x j:n + 3xk :n − xl:n ) (0.19)
i> j > k >l
Hosking (1990) showed that the first few L-moments follow from PWMs via:
l1 = b0
l2 = 2b1 − b0
(0.20)
l3 = 6b2 − 6b1 + b0
l4 = 20b3 − 30b2 + 12b1 − b0
The coefficients in Eqn. (0.20) are those of the shifted Legendre polynomials. The
first L-moment is the sample mean, a measure of location. The second L-moment is (a
multiple of) Gini's mean difference statistic (Johnson et al., 1994), a measure of the
dispersion of the data values about their mean. By dividing the higher-order L-
moments by the dispersion measure, we obtain the L-moment ratios,
tr = lr/l2 (0.21)
t=l2/l1 (0.22)
9
For a probability distribution with cumulative distribution function F(x), probability
z
weighted moments are defined by:
β r = x{F ( x )}r dF ( x ), r = 0,1,2,... (0.23)
L-moments are defined in terms of probability weighted moments, analogously to the
sample L-moments:
λ1 = β 0
λ 2 = 2β 1 − β 0
(0.24)
λ 3 = 6 β 2 − 6β 1 + β 0
λ 4 = 20β 3 − 30β 2 + 12β 1 − β 0
Jr=8r/82 (0.25)
The L-moment analogue of the coefficient of variation, is the L-CV, defined by:
J=82/81 (0.26)
Consider a sample consisting of n observations {x1, x2, . . ., xn} randomly drawn from
a statistical population. If the sample values are rearranged in a non-decreasing order
of magnitude, x1:n # x2:n # . . . # xn:n, then the r-th member (xr:n) of this new sequence
is called the r-th order statistic of the sample (Harter, 1969). When all the sample
values come from a common parent population with cumulative distribution function
F(x), the probability distribution (CDF) of the r-th order statistic, i.e., Prob[Xr:n ≤ x ],
means that at least r observations in a sample of n do not exceed a fixed value, x.
10
n
n
F(r)(x) = ∑ k F k
( x)[1 − F ( x)]( n − k ) (0.29)
k =r
The incomplete Beta function Ix(a,b) (Kendall and Stuart 1977) is defined via the Beta
function B(a,b) as:
Ix (a, b) =
1
B( a , b ) z x
0
t a −1 (1 − t ) b −1 dt
in which: (0.30)
B( a , b ) = z 1
0
t a −1 (1 − t ) b−1 dt =
( a − 1)!( b − 1)!
( a + b − 1)!
if a , b > 0
So, the expression (0.29) can be written in terms of an incomplete Beta function as:
F ( x)
n
F(r)(x) = r ∫ u r −1 (1 − u ) n − r du = IF(x)(r,n-r+1) (0.31)
r 0
1 n! n FG IJ
Indeed; note that =
B( r , n − r + 1) ( r − 1)!( n − r )!
=r
r
.
HK
The probability density function of Xr:n is given by the first derivative of Eqn. (0.31):
n r −1 n− r
f(r)(x) = r F ( x )[1 − F ( x )] f ( x) (0.32)
r
Now, the expected value of r-th order statistics can be obtained as
∞
E[Xr:n] = ∫ xf
−∞
(r) ( x)dx (0.33)
Comparing eqns.(0.37) and (0.38), it can be seen that αk and βk, respectively, are
related to the expectations of the minimum and maximum in a sample of size k
11
1 1
αk = E[ X 1:k +1 ] , βk = E[ X k +1:k +1 ] (k ≥ 1)
k +1 k +1
(0.39)
In essence, PWM’s are the normalized expectations of maximum/minimum of k
random observations; the normalization is done by the sample size (k) itself. From
z z
Eqn. (0.39), we notice that E(Xn:n) = n$n-1 and from Eqn. (0.23) we have
β n−1 = xF n−1 ( x ) f ( x )dx . So E ( X n:n ) = xnf ( x ) F n−1 ( x )dx . On the other hand, using
n
Eqn. (0.16), we have bn−1 = n −1 ∑ x j = n −1xn:n . From this it indeed follows that bn-1 is
j= n
z FG IJ
n r n
j 1
H K
β r = x{F ( x )} dF ( x ) ≈ ∑ x j:n = n −1 ∑ li xi:n
r
(0.40)
j =1 n +1 n i =1
Notice that the expression looks almost the same as Eqn. (0.16) by writing:
br = n −1 ∑
n
b gb g b g
j − 1 j − 2 ... j − r n
x j:n =n −1 ∑ ki xi:n
b gb g b g
j = r +1 n − 1 n − 2 ... n − r i = r +1
(0.41)
The terms li and ki in Eqn. (0.40) and (0.41) are compared in Figure 1 and we notice
indeed a very close similarity. Reiss (1989) derived more approximate distributions of
order statistics and Durrans (1992a) derived distributions of fractional order statistics.
12
where p r*,k represents the coefficients of shifted Legendre polynomials (Hosking
1990). The following normalized form of higher order L-moments is convenient to
work with:
λr
τr = , r = 3,4,. . . and |τr| < 1 (0.43)
λ2
The normalized fourth order L-moment, τ4, is referred to as the L-kurtosis of a
distribution. Hosking and Wallis (1997) showed that L-moments are very efficient in
estimating parameters of a wide range of distributions from small samples. The
required computation is fairly limited as compared with other traditional techniques,
such as maximum likelihood and least squares. In the Appendix the L-moment
formulae are given for a selection of PDFs. Apart from the well-known moments
diagrams, also L-Moment diagrams exist in which L-skewness and L-kurtosis are
plotted against eachother. However, the L-Moment diagrams do not form a complete
class; that is to say points in the diagram may correspond to more than one probability
distribution. This is in contrast to the ordinary moment diagram and also to the *1-*2
diagram of Halphen distributions (Bobee et al., 1993).
The Bayesian method
An introduction
Where x=(x1, x2, ...., xn). If, for instance, we assume a normal distribution for the
location parameter p(8G) = N(8G |µ G,FG) then, the posterior distribution of 8 becomes:
in which C a normalisation constant. Figure 2 shows the prior and posterior of the 8-
parameter when the following values are given: 8G = 20, *G = 20, (µ 8,F8) = (17, 2),
n=150.
P ri o r a nd P o s te ri o r
0 .0 6
0 .0 5
0 .0 4 P o s te ri o r
P robability D ensity
0 .0 3
0 .0 2 P ri o r
0 .0 1
13
0
5 10 15 20 25 30
L bd
Figure 2: Prior and posterior of the 8-parameter of the Gumbel likelihood model
Notice that the updating process of 150 observations has led to a more peaked and
shifted distribution function of the 8-parameter.
The way in which an engineer applies his information about a parameter 1
depends on the objectives in analyzing the data. If he is involved for instance in
calculating the frequency of wave heights H, then the inferences he makes on H
should reflect the uncertainty in 1. In the Bayesian framework we are interested in the
so-called predictive probability density function:
Provided they are not in direct conflict with our prior beliefs, and provided such a
family can be found, the simplicity obtained by using a conjugate prior is very large.
But in which situations can a conjugate family be obtained?
14
It appears that the only case where conjugates can be easily obtained is for
data models within the exponential family (Bernardo and Smith, 1994). That is,
z z
f ( x|θ )dx = g(θ ) h( x )et ( x )c (θ ) dx = 1 (0.50)
This might seem restrictive, but in fact Eqn. (0.49) includes the exponential
distribution, the Poisson distribution, the one-parameter Gamma distribution, the
Binomial distribution, the Normal distribution (with known variance), and the
extreme value distributions. For instance, the Gumbel distribution can be written as:
∑ t ( xi ) c ( θ )
∝ f (θ ) g(θ )n e i =1
Thus if we choose:
f (θ | x ) ∝ g(θ ) n+ d L n
O
exp{c(θ )M∑ t ( x ) + bP} = g(θ ) e
~
d
~
b c (θ )
N i =1Q i (0.54)
giving a posterior in the same family as the prior, but with modified parameters
~ ~ n
d = n + d , b = ∑ t ( xi ) + b. Indeed, for the Gumbel likelihood model and conjugate
i =1
prior with parameters d and b, we obtain:
15
ξα −α ( xi −ξ ) ξα −Σe−α ( xi −ξ )
f (ξ| x) = edξαe−be Παe−α( xi −ξ )e−e = α nedξα−Σα( xi −ξ )e−be =
ξα αξ Σe−αxi αξ ( b+Σe−αxi )
= α nedξα−αΣxi +nαξe−be e−e = α ne−αΣxi edξα+nξαe−e =
−eαξ ( b+Σe−αxi )
∝ eξα(n+d )e
(0.55)
The use of conjugate priors should be seen for what it is: a convenient mathematical
device. However, expression of one's prior beliefs as a parametric distribution is
always an approximation. In many situations the richness of the conjugate family is
great enough for a conjugate prior to be found which is sufficiently close to one's
beliefs (see also Pannullo et al., 1993).
The conjugate priors for the scale parameters of other PDFs can be easily
found with the techniques from this section, and they are summarized in the
Appendix.
- Statistical uncertainty in the parameter values which have been estimated on the
basis of past data; and
- Inherent uncertainty due to the fact that any future value is itself a random event.
In classical statistics it is usual to fit a model to the past data, and then make
predictions of future values on the assumption that this model is correct, the so-called
estimative approach. That is, only the second source of uncertainty is included in the
analysis, leading to estimates which are believed to be more precise than they really
are. There is no completely satisfactory way around this problem in the classical
framework since parameters are not thought of as being random.
Within Bayesian inference it is straightforward to allow for both sources of
uncertainty by simply averaging over the uncertainty in the parameter estimates (the
information of which is completely contained in the posterior distribution).
So, suppose we have past observations of a variable with density function (or
likelihood) f(x|2) and we wish to make inferences about the distribution of a future
value y from this process. With a prior distribution f(2), Bayes' theorem leads to a
posterior distribution f(2|x). Then the predictive density function of y given the data x
is:
16
Thus, the predictive density is the integral of the likelihood (of a single observation)
times the posterior. Notice that this definition is simply constructed from the usual
laws of probability manipulation, and the definition itself has a straightforward
interpretation itself in terms of probabilities.
The corresponding approach in classical statistics would be, for example, to
obtain the maximum-likelihood estimate 2* of 2 and to base inference on the
distribution f(y|2*), the estimative distribution.
To emphasize again, this makes no allowance for the variability incurred as a
result of estimating 2, and so gives a false sense of precision (the predictive density
f(y|x) is more variable by averaging across the posterior distribution for 2).
For data models within the exponential family, we obtain from (0.49):
z
f ( y| x) = f ( y|θ) f (θ| x)dθ = h( y)g(θ)e
Note:
z t ( y )c(θ ) ~ ~
z ~ ~
z θυ−1e−µθ dθ =
Γ(υ)
µυ
(0.57)
The latter expression can be used to solve the integral in case of simple h, g, c and t
functions. In case of the Gumbel likelihood model, we can derive (by applying the
substitution u=e"2):
z z
~
~ ~
d +1 − u ( e − αy + b ) du 1 d~ − u( e− αy + b~ ) Γ( d + 1)
f ( y| x ) ∝ h( y ) u e = h( y ) ue du = e−αy
αu α ~ ~
( e −αy + b)d +1 (0.58)
The normalising constant follows from If(y|x)dy = 1. The posterior functions of the
location and shape parameters of the exponential family (except for the normal
distribution) cannot be expressed in explicit form. Numerical integration has to be
performed. Uniform prior distributions can be used for the scale and location
parameters. For a large selection of models, the following information is given in the
Appendix: likelihood model, non-informative prior (and corresponding posterior), the
conjugate prior, the conjugate posterior, and the conjugate posterior predictive.
Some comments on the Empirical Bayes Method and the Bayes Linear
Estimation
17
Nonparametric methods
The basic idea of nonparametric density estimation is to relax the parametric
assumptions about the data, typically replacing these assumptions with ones about the
smoothness of the density. The most common and familiar nonparametric estimator is
the histogram. Here the assumption is that the density is fairly smooth (as determined
by the bin widths) and an estimate is made by binning the data and displaying the
proportion of points in each bin (producing a necessarily non-differentiable, but still
useful estimate).
The kernel density estimator is related to the histogram, but produces a smooth
(differentiable) estimate of the density. It has been studied widely since its
introduction in Rosenblatt (1956). Given i.i.d. data x1,...,xn drawn from the unknown
density ", the standard kernel estimator (SKE) is the single bandwidth estimator:
1 n FG
x − xi IJ
α$ ( x ) = ∑
nh i =1
K
H h K (0.59)
The bandwidth h determines the amount of smoothing produced by the estimator. See
the recent books by Silverman 1986, Scott 1992 and the bibliographies contained
therein, for a good introduction to kernel estimators. Much work has been done on
α ( x ) = pϕ ( x ,− µ , σ 12 ) + (1 − p )ϕ ( x, µ , σ 22 ) (0.60)
18
Suppose we wished to use the standard (single bandwidth) kernel estimator if
possible. It seems reasonable that near the left-hand mode one would wish to use a
bandwidth appropriate to that normal, and similarly on the right. So a reasonable
approach might be to filter the data into two distinct data sets, one from the right
component and one from the left, and estimate these two components separately.
One way to do this (approximately) is to define
ρ1 ( x ) = χ {x >0} ( x )
(0.61)
ρ 2 ( x ) = χ {x ≤0} ( x )
and define our estimator to be:
F
1 n ρ1 ( x i ) FG
x − xi ρ (x ) IJ FG
x − xi IJ I
α$ ( x ) = ∑
n i =1 h1GH K
h1 H+ 2 i K
h2 Kh2 H K JK (0.62)
This allows us to use the bandwidths appropriate to the different components in the
different regions where they are supported. Equation (0.61) is not quite right,
however, for as we move the two components closer together, the overlapping region
becomes more and more significant. What we really want to do is use the posteriors
for each component as our D functions. This is the motivation of the filtered kernel
estimator as proposed by Marchette (1995).
With the above example in mind, suppose we wish to have a small number of
bandwidths where each bandwidth is associated with a region of the support of the
density. To each bandwidth we associate a function which "filters" the data, as in
(0.61). Basically, the filter will define the extent to which each local bandwidth is to
be used for any particular data point. We can then construct a kernel estimator which
is a combination of the kernel estimators constructed using each bandwidth, with the
data filtered by the filtering functions.
Another method of nonparametric estimations is the following. Suppose we
have independent random variables X1 ; X2 ; ... , all of them with distribution function
F . Suppose that F is in the domain of attraction of some extreme value distribution
G(. In other words: suppose 1-F is regularly varying at infinity, i.e.,
1 − F ( tx )
lim t →∞ = x −1/γ (0.63)
1 − F (t )
γ$ M = M (1) 1
+1− 1−
RS
( M n(1) )2 UV −1
(0.65)
n
2 T
M n( 2 ) W
in which:
1 k −1
M n( i ) = ∑ (log X n− i:n − log X n− k :n )i (0.66)
k i=0
19
The estimators are consistent for any (. Under appropriate smoothness conditions
on F and a further bound on the rate of increase of k(n) the estimators for ( are
asymptotically normal after normalization (see e.g. Dekkers and de Haan, 1989).
Variability of estimators
As written earlier in the paper we prefer estimators which have low variability. But
the question arises as to whether there is any limit to the accuracy that an estimator
can achieve (Beard, 1994). Intuitively, we would expect that there is, since the
variation from one sample to another means that any estimator is bound to have some
degree of variation. However, is it possible to bound the accuracy which an estimator
can achieve? In fact, provided we restrict ourselves to unbiased estimators there is a
remarkable theorem which provides a bound on the variance of any (unbiased)
estimator (Barnett, 1973).
The quantity :
I(p)=-1/nE(M2logL(p,x)/Mp2) (0.67)
20
The result is important for both theoretical and practical reasons; it formulates
in a precise way the limits that can be achieved in statistical inference due to random
variation in the population. Though the Cramér-Rao bound puts a lower limit on the
variance of unbiased estimators, the bound may not be achievable.
So far we have used the likelihood function only to determine an estimate of
an unknown parameter 2. We have noted several times, however, that values of 2
with relatively high likelihood are more plausible than those with low likelihood, and
that it should be possible to exploit this knowledge to construct a credible range for 2.
This is achieved formally by looking at the sampling properties of the likelihood
function: quantifying the variation in L(2) from sample to sample.
In most situations exact calculation is very difficult and we will be forced to
use (asymptotic) approximations that assume large sample sizes. In the next Sub-
section we present some methods which are useful in deriving the sampling properties
of the likelihood function. It will be shown that these methods can be used to derive
the sampling distributions of the ML and LM estimators of the exponential
distribution.
Mood et al. (1974) present methods for deriving the sample distribution of estimators.
In some cases it is possible to establish the exact sampling distribution of an estimator
and use this as the basis for confidence interval estimation. More generally this is not
possible but approximations can be obtained. The following relations can be used in
deriving the exact sampling distributions:
fZ(z) = If(x)g(z-x) dx
(0.69)
fU(u) = If(u+y)g(y) dy
The proof of these formulae is given in Mood et al. (1974). In textbooks on statistics
the following relation is proven:
Furthermore it is possible to derive the following property for the product of random
variables:
21
Var(V) = Var(X)Var(Y) + E2(X)Var(Y) + E2(Y)Var(X)
(0.71)
If exact calculations are not possible, the following approximation rules can be used
(using Taylor’s formula):
dg ( x ) ( X − mX )2 d 2 g ( x )
g ( X ) = g ( mX ) + ( X − mX ) + +...
dx |x = mX 2 dx 2 |x = mX (0.72)
Var ( X )
E ( X ) ≈ E ( X ), Var ( X ) ≈
4E( X )
(0.75)
−1 1 Var ( X )
E( X ) ≈ , Var ( X −1 ) ≈ 4
E( X ) E (X)
FG 1 IJ σ( X )
F 1 I H X K ≈ E ( X ) = σ( X ) = CV ( X )
CV G J =
σ 2
From Eqn. (0.75) the approximation
H X K EFG 1 IJ 1 E( X )
H X K E( X )
follows directly. This approximation was shown to be very useful in Sec. 2.6.
The second equation in (0.75) is useful in deriving the sampling distribution of
the standard devation. It is well known that Var(s2) = 2F4/N in which F is the
population standard deviation of the normal distributed Xi and all Xi’s are i.i.d.
Therefore:
2σ 4
σ ( s) = N = σ (0.76)
4σ 2 2N
Exact calculations for the estimations of the 2-parameter Exponential distribution will
be given in the next sub-section.
The techniques from the previous section will shown to be succesful in deriving the
exact sampling distributions for the parameter estimations of the MML and MLM of
the exponential distribution.
The expressions of the MML and MLM for the location and scale parameter of
the exponential distribution are given in the Appendix. Suppose that X1, X2, ..., Xn ~
22
Exp(>,") i.i.d. This is the usual exponential model parameterized in such a way that
1 n
>+" is the population mean. The sampling distribution of mX = ∑ X i is the sum of
n i=1
n exponential distributions divided by n. According to Johnson et al. (1994), a
summation of n independent exponential distributions, each with scale parameter " is
gamma distributed with parameters (n, "), and a constant k times a gamma
distribution with parameters (",$) is also gamma distributed with parameters (",$/k).
Therefore mX is > + gamma distributed with parameters (n,"n).
Furthermore P(min(X1, X2, ..., Xn) < x) = 1 – P(min(X1, X2, ..., Xn) > x) =
= 1 – P (Xi > x) = 1 – (1 – P(Xi < x))n. So the minimum of n exponential distributions,
n
each with scale parameter ", is also exponentially distributed with scale parameter
"/n. Finally, Sukhatme (1937) showed that 2n/"(mX – min((X1, X2, ..., Xn)) has a Chi-
square distribution with 2(n-1) degrees of freedom.
With the above considerations and with Eqn. (0.70) we can derive that the
following properties for the sampling distribution of the MLEs:
E("ML) = "(1-1/n)
Var("ML) = "2(n-1 - n-2) (0.77)
E(>ML) = > + "/n
Var(>ML) = "2/n2
In order to derive the sampling distribution for the MLM, we use the following
properties of the order statistics of the standard exponential distribution (Johnson et
al., 1994):
i
1
E( X( i) ) = ∑
j =1 n − j + 1
i
1
Var ( X ( i ) ) = ∑ (0.78)
j =1 ( n − j + 1)
2
i
1
Cov( X ( i ) , X ( k ) ) = ∑ = Var ( X ( i ) )
j =1 ( n − j + 1)
2
n j
j −1
∑ ∑ n − k + 1 = c( n )
j = 2 k =1
(0.79)
n
2 j −1− n 1
∑
j = 2 n( n − 1)
= 1+
n
( n − 1)n n−1
The function c(n) can be derived to c( n ) = Ψ( n ) − ∑ ( n − j )Ψ( j ) , in which
2 j =1
Γ '( x ) n
1
Q is the Psi-function Ψ( x ) = , having the property ∑ = Ψ( n ) − Ψ( k )
Γ( x ) i= k i
2
(Abromowitz and Stegun, 1965). The function c(n) behaves as n for large values of n.
We write c(n)=O(n2).
23
For n=2 the L-Moments estimator of " reads "LM = x2:2 – x1:2. The
difference x2:2 – x1:2 has an exponential distribution with parameter ". Consequently
E("LM) = " (unbiased) and Var("LM) = "2 for n=2. The Eqn. (0.79) and the asymptotic
covariance matrix of the sample L-moments of Hosking (1986) are necessary to
derive the following sampling properties of MLM:
E("LM) = "(4n-1(n-1)-1c(n)-2)
Var("LM) = 4"2 / 3n + O(n-2)
E(>LM) = > + "(4n-1(n-1)-1c(n)-3) (0.80)
Var(>LM) = "2 /3n + O(n-2)
Note that the variance of "LM is slightly larger than the variance of "ML. The variance
of >LM decreases with O(n-1) instead of O(n-2) in case of the maximum likelihood
method. Furthermore, the expressions 3- 4n-1(n-1)-1c(n) and n-1 can be plotted in
Figure 3 for a graphical comparison of the biases:
0.9
0.8
0.7
0.6
B ias
0.5
0.4
0.3
L-Moments
0.2
ML
0.1
0
0 10 20 30 40 50 60
S ample size
Notice that the ML-method has a slightly better performance for the parameter
estimation of the exponential distribution than the L-Moments method; its bias and
RMSE is lower than the MLM-equivalents.
With Monte Carlo simulation studies, datasets can be generated from a beforehand
known probability distribution function (and known p-quantile). Different parameter
estimation methods can be applied on these datasets and compared with respect to
their estimates of the p-quantiles. The estimation method with the smallest bias and/or
variance is then considered to be the best method for that particular distribution
function. In order to familiarize the reader with the concept, in the following we start
with a simulation analysis of six estimation methods for the scale parameter of a one-
parameter exponential distribution F(x)=1-e-8x. In Figure 4 we see the performances
of the six methods for one simulation of 20 values the exponential distribution with
24
scale parameter 8=1.5. The different methods cause a very wide range of frequency
lines. The 100 year event (or the 0.99 quantile) of the theoretical distribution is 3.07m.
The above simulation analysis is repeated 200 times and from each simulation we
store per estimation method the results of the 1/100 year prediction. This excersise is
also repeated for other sample sizes. Apart from n=20 values we look at n=3, 6, 9,...,
60 values. In the next Figure 5 we have plotted the mean and standard deviations of
each estimation method.
25
Figure 5: Performance of six estimation methods as function of sample size (source: Van
Gelder and Vrijling (1997a))
26
WEIB Yamaguchi, 97 Yamaguchi, 97 Yamaguchi, 97 Yamaguchi, 97 Smith, 1987 Singh, 1990
Abernethy, 83 Smith, 1987
GPA Moharran, 93 Moharran, 93 Moharran, 93 Moharram, 93 Coles, 1996 Singh, 1997
Hosking 1987 Hosking 1987 Hosking 1987
Castillo 1997 Castillo 1997 Castillo 1997
GEV Takara, 1989 Takara, 1989 Takara, 1989 Yamaguchi, 97 Fill, 1998 Singh, 1992
Yamaguchi, 97 Yamaguchi, 97 Yamaguchi, 97 Coles, 1996 Jowitt, 1979
Sank., 1999 Lu, 1992b Hosking, 1985 Preumont, 88 Lind, 1991
Wang. 1998
LN Stedinger,1980 Takara,1990 Stedinger,1980 Takara, 1990 Corbyn 1988 Singh, 1987
Hoshi, 1986 Sank., 1999 Takara, 1990 Lechner, 1991 Lye, 1988
Yamaguchi, 97 Yamaguchi, 97 Yamaguchi, 97
Goda, 1992 Takeuchi 1988 Takeuchi 1988
GAM Hoshi, 1986 Rasmussen, 94 Hoshi, 1986 Ashkar, 1998 vNoortwijk, 99 Singh 1985
Bobee, 1991 Wu, 1991 Hu, 1987 Ribeiro 1993
Durrans, 92b Stacy, 1965
27
estimator uses the normalized quantile regression estimator to define a prior
distribution which is employed with the normalized 2P-quantile estimator. Monte
Carlo simulations indicate that this empirical Bayes estimator does essentially as well
as or better than the simpler normalized quantile regression estimator at sites with
short records, and performs as well as or better than the 2P-estimator at sites with
longer records or smaller L-CV.
In Wang (1998), approximate goodness-of-fit tests of fitted generalized
extreme value (GEV) distributions using LH moments are formulated on the basis of
comparison of sample LH kurtosis estimates and theoretical LH kurtosis values of the
fitted distributions. Their tests are different from those that have been derived for
testing the GEV distributions of which parameter values are known a priori. The tests
are intended to answer the following questions: Does a fitted GEV distribution
describe adequately a given data series? If not, can the GEV distribution function
describe adequately the larger events in that data series for use for high quantile
estimation? If so, what degree of emphasis on the larger events is needed in order that
the GEV distribution becomes acceptable? The use of the GEV distribution in
conjunction with the LH moment estimation method and the formulated tests should
alleviate the need for finding the "correct" distribution. The tests are evaluated by
Monte Carlo simulations using generated samples of both GEV and Wakeby
distributions.
Takeuchi and Tsuchiya (1988) derive PWM solutions for Normal and 3-parameter
Lognormal distributions. Their paper presents their relative accuracy in comparison
with other parameter estimation procedures such as Moment, Maximum-likelihood,
Quantile and Sextile methods through Monte Carlo simulation experiments.
Simulation results revealed that PWM estimates of quantiles are unbiased for the
Normal distribution and less biased than those of the Moment method for Lognormal
distribution with a large coefficient of skewness. It was also revealed that the RMSE
of PWM estimates of quantiles is as small as that of the Moment method for the
Normal distribution but larger for the Lognormal distribution.
In Lechner (1991) three common estimators for the parameters of the
lognormal distribution are evaluated. Correction factors which eliminate essentially
all the bias, and formulas for the standard deviations of the estimators, are presented.
It is reported that the Persson-Rootzen estimators are about as good as the maximum-
likelihood estimators, without the penalty of requiring iterative optimization. Also, the
estimators resulting from (least squares) fitting a line to the plot of log lifetimes on
normal (Gaussian) probability paper are reasonably good. Formulas are given for
obtaining these latter estimators without actually plotting the points. Lechner (1991)
simulated 5k to 30k samples (more samples for smaller N for each case) and
calculated the following: the means, standard deviations, and third moments of each
estimator; correlations between the two members of each pair; comparisons between
the estimators; and simple corrections to improve the performance of the estimators.
In Corbyn (1988), methods are developed for the determination of the
posterior distribution of the first moment of the lognormal distribution with
exponential and other prior distributions. Bayesian methods of statistical inference are
compared with the more generally used method of inference based on confidence
limits. The general problem of Bayesian estimation of the mean of a correlated
random variable is discussed.
Sankarasubramanian et al. (1999) deals with fitting of regression equations for
the sampling properties, variance of L-standard deviation, and bias and variance of L-
skewness, based on Monte-Carlo simulation results, for generalised Normal
28
(Lognormal-3) and Pearson-3 distributions. These fitted equations will be useful in
formulating goodness-of-fit test statistics in regional frequency analysis. The second
part of their paper presents a comparison of the sampling properties between L-
moments and conventional product moments for generalised Normal, generalised
Extreme Value, generalised Pareto and Pearson-3 distributions, in a relative form. The
comparison reveals that the bias in L-skewness is found to be insignificant up to a
skewness of about 1.0, even for small samples. In case of higher skewness, for a
reasonable sample size of 30, L-skewness is found to be nearly unbiased. However,
the conventional skewness is found to be significantly biased, even for a low
skewness of 0.5 and a reasonable sample size of 30. The overall performance
evaluation in terms of "Relative-RMSE in third moment ratio" reveals that
conventional moments are preferable at lower skewness, particularly for smaller
samples, while L-moments are preferable at higher skewness, for all sample sizes.
Corsini et al. (1995) analyze the Maximum likelihood (ML) algorithms and
Cramer-Rao (CR) bounds for the location and scale parameters of the Gumbel
distribution are discussed. First they consider the case in which the scale parameter is
known, obtaining the estimator of the location parameter by solving the likelihood
equation and then evaluating its performance. They also consider the case where both
the location parameter and the scale parameter are unknown and need to be estimated
simultaneously from the reference samples. For this case, performance is analyzed by
means of Monte-Carlo simulation and compared with the asymptotic CR bound.
Also results of a Monte Carlo study are presented in Guo and Cunnane (1991)
comparing different simulation procedures and assessing the value of historical floods
for at-site flood frequency analysis on the assumption of a Gumbel distribution.
In Wu et al. (1991), a new procedure, the method of lower-bound (MLB), is
proposed for determining the design quantile Xp. Their basic concept is first to
determine an estimate of the location parameter using the probability weighted
moment (PWM) method, and then to transform the variable X from the original (X)
space to a new (Y) space. The variable Y is considered to have a two parameter
gamma distribution. In Y-space, the two parameters are estimated by the PWM or an
autocovariance method, then transformed from the Y-space to the X-space. Results
from the Monte Carlo experiments show that the MLB estimates are less biased than
comparable moment estimates and maximum likelihood estimates, and more efficient
than those of PWM for design quantiles xp.
In Hu (1987a), the determination of confidence intervals for design floods
using the Pearson Type III distribution with a known skewness is analyzed. Tables for
the confidence factors based on moment and curve-fitting estimates were developed
by Monte Carlo simulation technique and were used to construct the confidence
intervals for frequency curves. The performance of methods using tables presented
herein, the method of B values based on the curve-fitting method are evaluated.
In Naghavi and Yu (1996), it is shown that the quantile prediction accuracy of
the log-Pearson type III (LP3) distribution depends largely on the accuracy of the
parameter-estimation method used. The performance of a parameter-estimation
method, on the other hand, depends on both the individual population chosen from the
LP3 family and the sample size. In this study Monte Carlo experiments were
conducted to evaluate four parameter-estimation methods that are frequently used in
hydrological analysis. The four methods tested are the method of indirect moments
(MMI), the method of mixed moments (MIX), the method of direct moments (MMD),
and a modification of MMI using optimization techniques (MMO). A quantile ratio
index (QRI) was devised to identify the limits (sample size and LP3 population
29
subset) within which each of these methods will perform best. This study
suggested that when QRI is less than or equivalent to 1.14, MMI or MMO should be
used for sample size N less than equivalent to 30, MIX for 30 less than N less than
100, and any of the four methods for N greater than equivalent to 100. When QRI
greater than 1.14, MMO is recommended for N less than equivalent to 30, MIX for 30
less than N less than 100, and MIX, MMO, or MMI for N greater than equivalent to
100.
In Pilon and Adamowski (1993), maximum likelihood and censored sample
theory are applied for flood frequency analysis purposes to the log Pearson Type III
(LP3) distribution. The logarithmic likelihood functions are developed and solved in
terms of fully specified floods, historical information, and parameters to be estimated.
The asymptotic standard error of estimate of the T-year flood is obtained using the
general equation for the variance of estimate of a function. The variances and
covariances of the parameters are obtained through inversion of Fisher's information
matrix. Monte Carlo studies to verify the accuracy of the derived asymptotic
expression for the standard errors of the 10, 50, 100, and 500 year floods, indicate that
these are accurate for both Type I and Type II censored samples, while the bias is less
than 2.5%. Subsequently, the Type II censored data were subjected to a random,
multiplicative error. Results indicate that historical information contributes greatly to
the accuracy of estimation of the quantities even when the error of its measurement
becomes excessive. (Type I: The threshold is fixed and the number of censored values
is a random variable. Type II: The number of censored values is fixed and the
threshold is a random variable).
In Lye et al. (1988), the three-parameter lognormal distribution is studied with
Bayesian estimates of the parameters and of the T-year quantile and their posterior
variances of the estimates are obtained by using Lindley's Bayesian approximation
procedure. These estimates are compared to estimates obtained by the method of
maximum likelihood. In all cases the posterior variances of the Bayes estimates of the
T-year flood events are less than the corresponding variances of their maximum
likelihood counterparts.
In Lye et al. (1993), ML estimators were compared with Bayesian estimators
for the reliability functions of the extreme value distributions.
In Yamaguchi (1997), Monte Carlo simulations have been performed to
determine a preferable parameter estimation method for each of eight distribution
functions. It was also shown that a jackknife method is beneficial to correct the bias
and RMSE irrespective of the estimation method used.
In Takara and Stedinger (1994) the use of two-parameter distributions is
recommended from the viewpoint of quantile estimation accuracy for datasets having
sample skewness greater than 0.38 and less than 1.8. They found that the quantile
lower bound estimators are likely to provide more accurate quantile estimators than
other procedures.
Singh and Guo (1997) and Singh and Singh (1985, 1987) showed that the
Method of Entropy yielded parameter estimates for the Generalized Pareto, the
Gamma and the lognormal distributions which were comparable or better within
certain ranges of sample size and coefficient of variation in comparison with MOM,
PWM and ML.
30
In Van Gelder (1996b) it was suggested to measure the performance of statistical
estimation methods with respect to over- and underdesign. In fact there is a strong
relation with the performance based on relative bias and RMSE from the previous
section. The relative bias and RMSE can give a first indication how much the under-
or overdesign is, however, in case of very skewed distributions of the quantile, this
indication might give a false impression of the amount of under- or overdesign.
Under- or overestimation of the p-quantiles have an important meaning in civil
engineering practice as well. Underestimation may give rise to unsafe structures
whereas overestimation may lead to conservatism or too expensive structures.
Therefore it is very useful to study the probabilities of under- and overdesign of a
certain estimation method.
As a typical result from Van Gelder (1996b), in this section, we will in particularly
concentrate on the under- and overestimation of the p-quantile of an Exponential and
Gumbel distribution with a ML- and LS-parameter estimation method. Different
sample sizes are considered (n=10, 30 and 100) for the same quantile of interest x100
such that P(x>x100 | Data) = 1/100. Typical results are shown in the next Tables 3 and
4.
Data from n ML LS
Exponential
Fitted by Gumbel 10 0.85 0.50
30 0.96 0.55
100 0.99 0.67
Fitted by Exponential 10 0.59 0.37
30 0.54 0.37
100 0.52 0.37
Table 4: Probabilities of underdesign pu
The probabilities of overdesign follow from the relation po=1-pu. From the Tables 3
and 4, it follows that the least squares method usually gives lower probabilities of
underdesign than the maximum likelihood method. That’s why a least squares method
is so popular under engineers. If we define assymetric loss-functions, in which we can
model the risk aversion of a designer towards underdesign, we can determine optimal
31
choices for distribution type and estimation method. For example if we penalize
underdesign with a factor 4 more than overdesign we get the following Table 5.
n Gumbel Exponential
f* EM* f* EM*
10 Exp ML Exp LS
30 G LS Exp LS
100 G LS Exp ML
Table 5: Optimal choices for distribution type f* and estimation method EM*
From this table, we indeed notice a preference for the least squares method, except for
large sample sizes from an exponential distribution where a ML-method is prefered
and for small sample sizes from a Gumbel distribution which are better modeled by an
exponential distribution with a ML-method for risk-averse engineers.
Discussion
32
Discussion
33
Figure 6: Effect of changes in the power parameter of Chezy’s equation on the flood
frequency curve (source: Chick et al., 1996)
References
Abernethy, R.B., 1983. Weibull Analysis Handbook, Pratt & Whitney Aircraft, November 1983.
Abramowitz, M., and Stegun, I., 1965. Handbook of Mathematical Functions, Dover Publications Inc.,
New York.
Anctil, F., Martel, N., and Van Diem Hoang. 1998. Analyse régionale des crues journalières de la
province de Québec. Canadian Journal of Civil Engineering, Pages 360-369 Volume 25,
Number 2, April 1998 ISSN 1208-6029.
Ashkar, F., and Ouarda, T.B.M.J., 1998. Approximate confidence intervals for quantiles of gamma and
generalized gamma distributions. Journal-of-Hydrologic-Engineering. v 3 n 1 Jan 1998, p 43-51.
Bardsley W.E., 1994. Against objective statistical-analysis of hydrological extremes , Journal of
hydrology, 162: (3-4) 429-431 NOV 1994.
Barnett, V., 1973. Comparitive Statistical Inference, John Wiley and Sons, Inc.
Bayes, T., 1763. An essay towards solving a probleme in the doctrine of chances, Reprinted from
Philos. Trans. Roy. Soc. London. Vol. 53: 370-418.
Beard, L.R., 1994. Anatomy of the best estimate. Journal of Hydraulic Engineering, Vol. 120, No. 6,
June, 1994.
Benjamin, J.R., and Cornell, C.A., 1970. Probability, Statistics and Decision for Civil Engineers,
McGraw-Hill, Inc.
Ben-Zvi, A., and Azmon, B., 1997. Joint use of L-moment diagram and goodness-of-fit test: A case
study of diverse series, Journal of Hydrology. v 198 n 1-4 Nov 1997. p 245-259.
Berger, J.O., 1980. Statistical Decision Theory, Foundations, Concepts, and Methods, Springer-Verlag.
Bernardo, J.M., and Smith, A.F.M., 1994. Bayesian Theory, John Wiley & Sons, Inc., Wiley Series in
Probability and Mathematical Statistics.
Bernoulli Family, 1999. The MacTutor History of Mathematics archive, http://www-history.mcs.st-
and.ac.uk/history/
34
Bobée, B., and F. Ashkar, 1991. The Gamma Family and Derived Distributions Applied in Hydrology,
203 pp., Water ResourcesPublication, Littleton, CO.
Bobée, B., Ashkar, F., and Perreault, L., 1993a. Two kinds of moment ratio diagrams and their
applications in hydrology, Stochastic Hydrol. Hydraul., 7, 41-65.
Bobée, B., Cavadias, G., Ashkar, F., Bernier, J., and Rasmussen, P.F., 1993b. Towards a systematic
approach to comparing distributions used in flood frequency analysis, J. Hydrol., 142, 121-136.
Box, G.E.P., and Tiao, G.C., 1973. Bayesian Inference in Statistical Analysis, Addison-Wesley.
Burcharth, H.F., and Liu, Z., 1994. On the extreme wave height analysis, In Port and Harbour Research
Institute, Ministry of Transport, editor, International Conference on Hydro-Technical
Engineering for Port and Harbor Construction (Hydro-Port), Yokosuka, Japan, 1994, pages 123-
142, Yokosuka: Coastal Development Institute of Technology.
Capehart, B.L., Mahoney, J.F., and Sivazlian, B.D., 1998. Technological risk assessment: The
statistical evaluation of low frequency events.Publ by Inderscience Enterprises Ltd, Geneva,
Switz p 749-755. Technology Management 1: Proceedings of the First International Conference
on Technology Management. Special Publication of the International Journal of Technology
Management. Miami, FL, USA
Carlin, B.P., and Louis, T.A., 1996. Bayes and Empirical Bayes Methods for Data Analysis, Chapman
& Hall.
Carter, D.J.T., and Challenor, P.G., 1983. Methods of Fitting the Fisher-Tippett Type 1 Extreme Value
Distribution, Ocean Engineering, Vol.10, No. 3, pp.191-199.
Castillo, E., and Hadi, A.S., 1997. Fitting the generalized Pareto distribution to data, Journal-of-the-
American-Statistical-Association. DEC 1997; 92 (440) : 1609-1620.
Chick, S., Shortle, J., Van Gelder, P., and Mendel, M., 1995. A Physics-Based Approach to Predicting
the Frequency of Extreme River Levels, Engineering Probabilistic Design and Maintenance for
Flood Protection, pp.89-107. Discussion: pp.109-138.
Chick, S., Shortle, J., Van Gelder, P., and Mendel, M., 1996. A Model for the Frequency of Extreme
River Levels Based on River Dynamics, Structural Safety, Vol. 18, Nr. 4, pp.261-276.
Chow, K.C.A., and Watt, W.E., 1990. A knowledge-based expert system for flood frequency analysis,
Can. J. Civ. Eng., 17, 597-609.
Coles, S.G., and Powell, E.A., 1996. Bayesian methods in extreme value modelling: A review and new
developments, International Statistical Review, 64: (1) 119-136 APR 1996.
Coles, S.G., and Tawn, J.A., 1996. A Bayesian analysis of extreme rainfall data, applied statistics-
journal of the royal statistical society series C, 45: (4) 463-478 1996.
Corbyn, J.A., 1988. Statistical analysis of samples from lognormal distribution by Bayesian methods
with minerals industry applications.Transactions-of-the-Institution-of-Mining-and-Metallurgy,-
Section-A, v 97 Jul 1988, p 118-124.
Corsini, G., Gini, F., Greco, M.V., and Verrazzani, L., 1995. Cramer-Rao bounds and estimation of the
parameters of the Gumbel distribution, IEEE-Transactions-on-Aerospace-and-Electronic-
Systems. V 31 n 3 Jul 1995, p 1202-1204.
Cramer, H., 1946. Mathematical methods of statistics, Princeton University Press.
Cunnane, C., 1987. Review of statistical methods for flood frequency estimation, in Hydrologic
Frequency Modeling, edited by V.P. Singh, pp. 49-95, D. Reidel, Dordrecht.
Cunnane, C., 1988. Methods and merits of regional flood frequency analysis, J. Hydrol., 100, p. 269-
290.
35
Dekkers, A.L.M., and De Haan, L., 1989. On the estimation of the extreme-value index and large
quantile estimation. Annals of Statistics 17, 1795-1832.
Demetracopoulos, A.C., 1994. nonlinear-regression applied to hydrologic data, journal of irrigation and
drainage engineering-ASCE, 120: (3) 652-659 May-June 1994.
Demuth, S., and Kuells, C., 1997. Probability analysis and regional aspects of droughts in southern
Germany. Symp 1: Sustainability of Water Resources under Increasing Uncertainty IAHS
Publication (International Association of Hydrological Sciences). n 240.
Duan, J., Selker, J., and Grant, G.E., 1998. Evaluation of probability density functions in precipitation
models for the Pacific Northwest, Journal of the American Water Resources Association. v 34 n
3 Jun 1998. p 617-627.
Durrans, S.R., 1992a. Distributions of fractional order statistics in hydrology, Water Resour. Res.,
28(6), 1649-1655.
Durrans, S.R., 1992b. Parameter estimation for the Pearson type 3 distribution using order statistics, J.
Hydrol., 133, 215-232.
Fill, H.D., and Stedinger, J.R., 1995. Homogeneity tests based upon Gumbel distribution and a critical
appraisal of Dalrymple’s test, Journal of Hydrology. v 166 p 81-105.
Fill, H.D., and Stedinger, J.R., 1998. Using regional regression within index flood procedures and an
empirical Bayesian estimator, Journal-of-Hydrology. SEP 1998; 210 (1-4) : 128-145.
Fisher, R.A., 1934. Statistical methods for research workers, Publisher Edinburgh, Biological
monographs and manuals. no. 5.
Gauss, C.F., 1901. Sechs Beweise des Fundamentaltheorems über quadratische Reste (Six proofs of the
fundamental theorems on quadratic residuals; in German), Publisher Leipzig : Engelmann,
HRSG von E. Netto. 1901, 111 pages, Series Ostwald's Klassiker der exakten Wissenschaften
Nr. 122.
Gingras, D., and Adamowski, K., 1994. Performance of L-moments and nonparametric flood frequency
analysis. Canadian Journal of Civil Engineering. v 21 n 5 Oct 1994. p 856-862.
Goda, Y., and Kobune, K., 1990. Distribution Function Fitting for Storm Wave Data, Proceedings of
the International Conference on Coastal Engineering, Delft, The Netherlands.
Goda, Y., Hawkes, P.J., Mansard, E., Martin, M.J., Mathiesen, M., Peltier, E., Thompson, E.F., and
Van Vledder, G., 1993. Intercomparison of extremal wave analysis methods using numerically
simulated data, Ocean wave measurement and analysis, pp. 963-977.
Greenwood, J.A., Landwehr, J.M., Matalas, N.C., and Wallis, J.R., 1979. Probability weighted
moments: Definition and relation to parameters of several distributions expressable in inverse
form. Water Resources Research, 15(5), 1049-1054.
Guo, S.L., and Cunnane, C., 1991. Evaluation of the usefulness of historical and palaeological floods in
quantile estimation, J. Hydrol., 129, 245-262.
Haimes, Y.Y., Barry, T., and Lambert, J.H., 1994. When and how can you specify a probability
distribution when you don't know much, Risk-Based Decision Making in Water Resources VI.
Y.Y. Haimes, D.A. Moser, and E.Z. Stakhiv, Eds., New York: American Society of Civil
Engineers, pp. 19-25.
Hald, A., 1952. Statistical Theory with Engineering Applications, John Wiley and Sons, Inc., New
York.
Harter, H.L., 1969. Order Statistics and Their Use in Testing and Estimation. Volume 2, Aerospace
Research Laboratories, U.S. Air Force, Washington DC, USA.
36
Hoshi, K., and Leeyavanija, U., 1986. A new approach to parameter estimations of gamma-type
distributions, Journal of Hydroscience and Hydraulic Engineering, JSCE, 4 (2), 79-95.
Hosking, J.R.M., 1986. The theory of probability weighted moments, Research Rep. RC 12210, 160
pp., IBM Research Division, Yorktown Heights, NY.
Hosking, J.R.M., 1990. L-moments: Analysis and estimation of distributions using linear combinations
of order statistics. J. R. Stat. Soc. Ser. B., vol. 52, 105-124.
Hosking, J.R.M., 1992. Moments or L moments? An example comparing two measures of distribution
shape. The American Statistician, 46(3), 186-189.
Hosking, J.R.M., 1997. Fortran Routines for Use with the Method of L-Moments. IBM Research
Report, RC20525, Yorktown Heights, NY.
Hosking, J.R.M., and Wallis, J.R., 1987. Parameter and quantile estimation for the generalized Pareto
distribution, Technometrics, 29(3), 339-349.
Hosking, J.R.M., and Wallis, J.R., 1997. Regional Frequency Analysis: An Approach based on L-
Moments. Cambridge University Press, Cambridge, UK.
Hosking, J.R.M., Wallis, J.R., and Wood, E.F., 1985. Estimation of the Generalized Extreme Value
Distribution by the Method of Probability Weighted Moments, Technometrics, August 1985,
Vol. 27, No.3, pp.251-261.
Hu, S., 1987a. Determination of confidence intervals for design floods. J-Hydrol. v 96 n 1-4 Dec 15
1987, p 201-213.
Hu, S., 1987b. Problems with outlier test methods in flood frequency analysis. J-Hydrol. v 96 n 1-4
Dec 15 1987, p 375-383.
Johnson, N.L., Kotz, S., and Balakrishnan, N., 1994. Continuous univariate distributions. Vol. 1
Publisher New York : Wiley, 1994 ISBN 0-471-58495-9, 756 pages.
Jowitt, P.W., 1979. The extreme-value type-1 distribution and the principle of maximum entropy,
Journal of Hydrology, 42, pp.23-28.
Jumarie, G., 1990. Relative Information: Theories and Applications. Springer-Verlag, Berlin,
Germany.
Kendall, M.G., 1961. Daniel Bernoulli on maximum likelihood, Biometrika 48, 1-18.
Koch, S.P., 1991. Bias error in maximum likelihood estimation, J. Hydrol., 122, p. 289-300.
Kuczera, G., 1994. Comprehensive Bayesian at-site flood frequency inference, Proceedings of the
Water Down Under 1994 Conference. Part 3 (of 3), IE Aust, Crows Nest, NSW, Aust. p 205-
210.
Kullback, S., 1959. Information Theory and Statistics. John Wiley & Sons Inc., New York, USA.
Lall, U., Moon, Y., and Bosworth, K., 1993. Kernel Flood Frequency Estimators: Bandwidth Selection
and Kernel Choice. Water Resources Research , April 1993, pp. 1003-1016.
Lambert, J.H., Matalas, N.C., Ling, C.W., Haimes, Y.Y., and Li, D., 1994. Selection of Probability
Distributions in Characterizing Risk of Extreme Events, Risk Analysis, Vol.14, No.5.
Landwehr, J.M., Matalas, N.C., and Wallis, J.R., 1979. Probability weighted moments compared with
some traditional techniques in estimating Gumbel parameters and Quantiles, Water Resources
Research, Vol. 15., No.5, October 1979, pp.1055-1064.
Lechner, J.A., 1991. Estimators for type-II censored (log)normal samples. IEEE-Trans-Reliab. v 40 n 5
Dec 1991, p 547-552.
Lin, B., and Vogel, J.L., 1993. Comparison of L-moments with method of moments Proceedings of the
Symposium on Engineering Hydrology Proc Symp Eng Hydrol. Publ by ASCE, New York,
NY, USA. 1993. p 443-448.
37
Lind, N.C., and Hong, H.P., 1991. Entropy estimation of hydrological extremes, Stochastic
Hydrol. Hydraul., 5, 77-87.
Lu, L.-H., and Stedinger, J.R., 1992b. Variance of two- and three-parameter GEV/PWM quantile
estimators: formulae, confidence intervals, and a comparison, J. Hydrol., 138, 247-267.
Lundgren, H., 1988. Variational criteria in spectrum analysis and image reconstruction, Signal
Processing IV, Elsevier, B157pp. 1573-1576.
Lye, L.M., Hapuarachchi, K.P., and Ryan, S., 1993. Bayes estimation of the extreme-value reliability
function, IEEE Transactions on Reliability, Vol. 42, No. 4, 1993 December.
Lye, L.M., Sinha, S.K., and Booy, C., 1988. Bayesian analysis of the tau-year events for flood data
fitted by a three-prameter lognormal distribution. Civ-Eng-Syst. v 5 n 2 Jun 1988, p 81-86.
Marchette, D.J., 1995. An Investigation of Misspecification of Terms in Mixture Densities, Technical
Report, Computational Statistics at George Mason University
Marchette, D.J., Priebe, C.E., Rogers G.W., and Solka, J.L., 1994. The Filtered Kernel Estimator,
Center for Computational Statistics, GMU, Tech Rpt No 104, October 1994.
McCuen, R.H., Leahy, R.B., and Johnson, P.A., 1990. Problems with logarithmic transformations in
regression , journal of hydraulic engineering-ASCE, 116: (3) 414-428 MAR 1990.
Mendel, M.B., and Chick, S.E., 1993. The Geometry and Calculus of Engineering Probability, UC-
Berkeley, Lecture notes, December 1993.
Moharram, S.H., Gosain, A.K., and Kapoor, P.N., 1993. A comparative study for the estimators of the
generalized Pareto distribution, J. Hydrol., 150, 169-185.
Mood, A.M., Graybill, F.A., and Boes, D.C., 1974. Introduction to the theory of statistics. 3rd ed.
Publisher New York : McGraw-Hill, 1974 Series in probability and statistics.
Moon, Y., and Lall, U., 1994. A Kernel Quantile Function Estimator For Flood Frequency Analysis, in
Extreme Values: Floods and Droughts, ed. K. Hipel, Kluwer.
Moon, Y., Lall, U., and Bosworth, K., 1993. A Comparison of Tail Probability Estimators for Flood
Frequency Analysis. Journal of Hydrology , V. 151, 1993, pp. 343-363.
Naghavi, B., and Yu, F.-X., 1996. Selection of parameter-estimation method for LP3 distribution.
Journal-of-Irrigation-and-Drainage-Engineering. v 122 n 1 Jan-Feb 1996, p 24-30.
Pannullo, J.E., Li, D., and Haimes, Y.Y., 1993. Posterior analysis in assessing risk of extreme events: a
conjugate family approach, Proceedings-of-the-IEEE-International-Conference-on-Systems,-
Man-and-Cybernetics. v 1 1993, Publ by IEEE, IEEE Service Center, Piscataway, NJ,
USA,CH3242-5. p 477-482.
Pearson, C.P., McKerchar, A.I., and Woods, R.A., 1991. Regional flood frequency analysis of
Western Australian data using L-moments. International Hydrology and Water Resources
Symposium 1991 Part 2. v 2 n 91 pt 22. p 631-632.
Perreault, L., Bobee, B., and Perron, H., 1994. Ajuste II: hydrological frequency analysis software part
1: theoretical aspects, International-Conference-on-Hydraulic-Engineering-Software,-
Hydrosoft,-Proceedings-2. 1994, Computational Mechanics Publ, Southampton, Engl. p 243-
250
Perron, H., Bobee, B., Perreault, L., and Roy. R., 1994. Ajuste II: hydrological frequency analysis
software part 2: software aspects and applications, International-Conference-on-Hydraulic-
Engineering-Software,-Hydrosoft,-Proceedings-2. 1994, Computational Mechanics Publ,
Southampton, Engl. p 251-258.
Pickands, J., 1975. Statistical inference using extreme order statistics, The Annals of Statistics, 3(1),
119-131.
38
Pilon, P.J., and Adamowski, K., 1993. Asymptotic variance of flood quantile in log Pearson Type III
distribution with historical information, J. Hydrol., 143, 481-503.
Preumont, A., 1988. Bayesian analysis of the extreme value of a time history. Reliab-Eng-Syst-Saf. v
20 n 3 1988, p 165-172.
Rao, A.R., and Hamed, K.H., 1997. Regional frequency analysis of Wabash River flood data by L-
moments, Journal of Hydrologic Engineering. v 2 n 4 Oct 1997. p 169-179.
Rasmussen, P.F., Bobée, B., and Bernier, J., 1994. Parameter estimation for the Pearson type 3
distribution using order statistics (Durrans, 1992)---Comment, J. Hydrol., 153, 417-424.
Reiss, R.-D., 1989. Approximate Distributions of Order Statistics, With Applications to Nonparametric
Statistics, Spinger Verlag Series in Statistics.
Ribeiro-Correa, J., and Rousselle, J., 1993. A hierarchical and empirical Bayes approach for the
regional Pearson type III distribution, Water Resour. Res., 29(2), p. 435-444.
Rosbjerg, D., Madsen, H., and Rasmussen, P.F., 1992. Prediction in partial duration series with
generalized Pareto-distributed exceedances, Water Resour. Res., 28(11), p. 3001-3010.
Rosenblatt, M., 1956, Remarks on some Nonparametric Estimates of a Density Function, Ann. Math.
Statist., 27, p. 832-835.
Ruprecht, J.K., and Karafilis, D.W., 1994. Regional flood frequency - caution needed Proceedings of
the Water Down Under 1994 Conference. Part 3 (of 3). Adelaide, Aust. 19941121-19941125.
Sankarasubramanian, A., and Srinivasan, K., 1999. Investigation and comparison of sampling
properties of L-moments and conventional moments. Journal-of-Hydrology. May 10 1999; 218
(1-2) : 13-34.
Scott, D.W., 1992. Multivariate Density Estimation, New York: John Wiley.
Seiler, F.A., and Alvarez, J.L., 1996. On the selection of distributions for stochastic variables, Risk
Analysis, Vol. 16, No. 1, pp. 5-18.
Shannon, C.E., 1949. The Mathematical Theory of Communication. The University of Illinois Press,
Urbana, IL, USA.
Silverman, B.W., 1986. Density Estimation for Statistics and Data Analysis, New York: Chapman and
Hall.
Singh V.P., 1997. The use of entropy in hydrology and water resources, hydrological processes, 11: (6)
587-626 MAY 1997.
Singh V.P., and Guo H., 1997. Parameter estimation for 2-parameter generalized pareto distribution by
Pome, Stochastic hydrology and hydraulics, 11: (3) 211-227 JUN 1997.
Singh, V.P., and Fiorentino, M., 1992. A Historical perspective of entropy applications in water
resources, In Singh, V.P., and Fiorentino, M. (eds.), Entropy and Energy Dissipation in Water
Resources, pp.21-61, Kluwer Academic Publishers.
Singh, V.P., and Singh, K., 1985. Derivation of the Gamma distribution by using the principle of
maximum entropy (POME). Water Resources Bulletin, Vol. 21, No. 6, pp.941-952.
Singh, V.P., and Singh, K., 1987. Parameter estimation for TPLN distribution for flood frequency
analysis. Water Resources Bulletin, vol. 23, No. 6, pp.1185-1992.
Singh, V.P., Cruise, J.F., and Ma, M., 1990. A Comparative-evaluation of the estimators of the 3-
parameter lognormal-distribution by Monte-Carlo Simulation, Computational statistics & data
analysis, 10: (1) 71-85 AUG 1990.
Smith, R.L., and Naylor, J.C., 1987. A comparison of Maximum Likelihood and Bayesian Estimators
for the Three-parameter Weibull Distribution, Appl. Statist. 36, No. 3, pp. 358-369.
39
Smith, R.L., and Weissman, I., 1987. Large deviations of tail estimators based on the Pareto
approximation. J. Appl. Prob. Vol. 24, pp. 619-630.
Stacy, E.W., and Mihram, G.A., 1965. Parameter estimation for a generalized gamma distribution,
Technometrics, Vol. 7, No. 3, pp. 349-358, August 1965.
Stedinger, J.R., 1980. Fitting log normal distributions to hydrologic data, Water Resources Research,
16 (3), 481-490.
Sukhatme, P.V., 1937. Tests of significance for samples of the Chi-squared population with two
degrees of freedom, Ann. Eugen., Vol. 8, pp. 52-56.
Takara, K., and Takasao, T., 1990. Comparison for parameter estimation methods for hydrologic
frequency analysis models, Proc. Hydraulic Engineering, JSCE, 34, 7-12 (in Japanese).
Takara, K., Takasao, T., and Shimizu, A., 1989. Comparison of parameter estimation methods for
extreme value distributions. Annuals, DPRI, Kyoto University, 32 B-2, 455-469 (in Japanese).
Takara, K.T., and Stedinger, J.R., 1994. Recent Japanese contributions to frequency analysis and
quantile lower bound estimators, In: K.W. Hipel (ed.), Stochastic and Statistical methods in
hydrology and environmental engineering, Vol. 1, p. 217-234, Kluwer Academic Publishers.
Takeuchi, K., and Tsuchiya, K., 1988. On relative accuracy of PWM estimates of normal and 3-
parameter lognormal distributions, Proc. of JSCE, Vol.393/ll-9, May,pp. 103-112.
TAW: Technical Advisory Committee on Water Defences, 1990. Probabilistic Design of Flood
Defences. Report 141, Centre for Civil Engineering Research and Codes, The Netherlands.
Van Gelder, P.H.A.J.M., 1999b. Risks and safety of flood protection structures in the Netherlands,
Proceedings of the Participation of Young Scientists in the Forum Engelberg 1999 on Risk and
Safety of Technical Systems - in View of Profound Changes, pp.55-60, p.93.
Van Gelder, P.H.A.J.M., and Neykov, N.M., 1998. Regional frequency analysis of extreme water
levels along the Dutch coast using L-moments: A preliminary study, In: Stochastic models of
hydrological processes and their applications to problems of environmental preservation, pp.14-
20.
Van Gelder, P.H.A.J.M., and Vrijling, J.K., 1997a. A comparative study of different parameter
estimation methods for statistical distribution functions in civil engineering applications,
Structural Safety and Reliability, Vol. 1, pp.665-668.
Van Gelder, P.H.A.J.M., Pandey, M.D., and Vrijling, J.K., 1999b. The use of L-Kurtosis in the
Estimation of Extreme Floods, 9th Annual Conference Risk Analysis: Facing the New
Millennium, Rotterdam, The Netherlands, October 10 - 13, 1999.
Van Gelder, P.H.A.J.M., 2000. Statistical methods for the risk-based design of civil structures,
Publisher TU Delft, Descr. 248 p, Series Communications on Hydraulic and Geotechnical
Engineering ISSN:0169-6548 00-1.
Van Noortwijk, J.M., 1999. Quantiles of Generalised Gamma Distributions from a Bayesian Point of
View, HKV-report.
Vrijling, J.K., and Van Gelder, P.H.A.J.M., 1995. Probabilistic design of berm breakwaters, in:
Engineering Probabilistic Design and Maintenance for Flood Protection, pp. 181-198,
Discussion: pp. 199-213.
Vrijling, J.K., and Van Gelder, P.H.A.J.M., 1997a. Societal risk and the concept of risk aversion,
Advances in Safety and Reliability, Vol. 1, pp. 45-52.
Wang, Q.J., 1998. Approximate goodness-of-fit tests of fitted generalized extreme value distributions
using LH moments, Water Resources Research, 34: (12) 3497-3502 DEC 1998.
40
WMO, 1989. Statistical distributions for flood frequency analysis, Operational Hydrology Report No.
33, World Meteorological Organization---No. 718, Geneva, Switzerland.
Wu B., Hou Y., and Ding J., 1991. Method of lower-bound to estimate the parameters of a Pearson
type III distribution. Hydrol-Sci-J. v 36 n 3 Jun 1991, p 271-280.
Yamaguchi, M., 1997. Intercomparison of parameter estimation methods in extremal wave analysis. In
B.L. Edge, editor, 25th International Conference on Coastal Engineering, Orlando, Florida,
U.S.A., 1996, pages 900-913, New York: American Society of Civil Engineers (ASCE).
41