egamma {EnvStats}  R Documentation 
Estimate the shape and scale parameters (or the mean and coefficient of variation) of a Gamma distribution.
egamma(x, method = "mle", ci = FALSE,
ci.type = "twosided", ci.method = "normal.approx",
normal.approx.transform = "kulkarni.powar", conf.level = 0.95)
egammaAlt(x, method = "mle", ci = FALSE,
ci.type = "twosided", ci.method = "normal.approx",
normal.approx.transform = "kulkarni.powar", conf.level = 0.95)
x 
numeric vector of nonnegative observations.
Missing ( 
method 
character string specifying the method of estimation. The possible values are: 
ci 
logical scalar indicating whether to compute a confidence interval for the mean.
The default value is 
ci.type 
character string indicating what kind of confidence interval to compute.
The possible values are

ci.method 
character string indicating which method to use to construct the confidence interval.
Possible values are 
normal.approx.transform 
character string indicating which power transformation to use when 
conf.level 
a scalar between 0 and 1 indicating the confidence level of the confidence interval. The default
value is 
If x
contains any missing (NA
), undefined (NaN
) or
infinite (Inf
, Inf
) values, they will be removed prior to
performing the estimation.
Let \underline{x} = x_1, x_2, \ldots, x_n
denote a random sample of
n
observations from a gamma distribution
with parameters shape=
\kappa
and scale=
\theta
.
The relationship between these parameters and the mean (mean=
\mu
)
and coefficient of variation (cv=
\tau
) of this distribution is given by:
\kappa = \tau^{2} \;\;\;\;\;\; (1)
\theta = \mu/\kappa \;\;\;\;\;\; (2)
\mu = \kappa \; \theta \;\;\;\;\;\; (3)
\tau = \kappa^{1/2} \;\;\;\;\;\; (4)
The function egamma
returns estimates of the shape and scale parameters.
The function egammaAlt
returns estimates of the mean (\mu
) and
coefficient of variation (cv
) based on the estimates of the shape and
scale parameters.
Estimation
Maximum Likelihood Estimation (method="mle"
)
The maximum likelihood estimators (mle's) of the shape and scale parameters
\kappa
and \theta
are solutions of the simultaneous equations:
\hat{\kappa}_{mle} = \frac{1}{n}\sum_{i=1}^n log(x_i)  log(\bar{x}) = \psi(\hat{\kappa}_{mle})  log(\hat{\kappa}_{mle}) \ \;\;\;\;\;\; (5)
\hat{\theta}_{mle} = \bar{x} / \hat{\kappa}_{mle} \;\;\;\;\;\; (6)
where \psi
denotes the digamma function
,
and \bar{x}
denotes the sample mean:
\bar{x} = \frac{1}{n}\sum_{i=1}^n x_i \;\;\;\;\;\; (7)
(Forbes et al., 2011, chapter 22; Johnson et al., 1994, chapter 17).
BiasCorrected Maximum Likelihood Estimation (method="bcmle"
)
The “biascorrected” maximum likelihood estimator of
the shape parameter is based on the suggestion of Anderson and Ray (1975;
see also Johnon et al., 1994, p.366 and Singh et al., 2010b, p.48), who noted that
the bias of the maximum likelihood estimator of the shape parameter can be
considerable when the sample size is small. This estimator is given by:
\hat{\kappa}_{bcmle} = \frac{n3}{n}\hat{\kappa}_{mle} + \frac{2}{3n} \;\;\;\;\;\; (8)
When method="bcmle"
, Equation (6) above is modified so that the estimate of the
scale paramter is based on the “biascorrected” maximum likelihood estimator
of the shape parameter:
\hat{\theta}_{bcmle} = \bar{x} / \hat{\kappa}_{bcmle} \;\;\;\;\;\; (9)
Method of Moments Estimation (method="mme"
)
The method of moments estimators (mme's) of the shape and scale parameters
\kappa
and \theta
are:
\hat{\kappa}_{mme} = (\bar{x}/s_m)^2 \;\;\;\;\;\; (10)
\hat{\theta}_{mme} = s_m^2 / \bar{x} \;\;\;\;\;\; (11)
where s_m^2
denotes the method of moments estimator of variance:
s_m^2 = \frac{1}{n} \sum_{i=1}^n (x_i  \bar{x})^2 \;\;\;\;\;\; (12)
Method of Moments Estimation Based on the Unbiased Estimator of Variance (method="mmue"
)
The method of moments estimators based on the unbiased estimator of variance
are exactly the same as the method of moments estimators,
except that the method of moments estimator of variance is replaced with the
unbiased estimator of variance:
\hat{\kappa}_{mmue} = (\bar{x}/s)^2 \;\;\;\;\;\; (13)
\hat{\theta}_{mmue} = s^2 / \bar{x} \;\;\;\;\;\; (14)
where s^2
denotes the unbiased estimator of variance:
s^2 = \frac{1}{n1} \sum_{i=1}^n (x_i  \bar{x})^2 \;\;\;\;\;\; (15)
Confidence Intervals
This section discusses how confidence intervals for the mean \mu
are computed.
Normal Approximation (ci.method="normal.approx"
)
The normal approximation method is based on the method of Kulkarni and Powar (2010),
who use a power transformation of the the original data to approximate a sample
from a normal distribuiton, compute the confidence interval for the mean on the
transformed scale using the usual formula for a confidence interval for the
mean of a normal distribuiton, and then tranform the limits back to the original
space using equations based on the expected value of a gamma random variable
raised to a power.
The particular power used for the normal approximation is defined by the argument
normal.approx.transform
. The value
normal.approx.transform="cube.root"
uses the cube root transformation
suggested by Wilson and Hilferty (1931), and the value
"fourth.root"
uses the fourth root transformation suggested
by Hawkins and Wixley (1986). The default value "kulkarni.powar"
uses the “Optimum Power Normal Approximation Method” of Kulkarni and Powar
(2010), who show this method performs the best in terms of maintining coverage
and minimizing confidence interval width compared to eight other methods.
The “optimum” power p
is determined by:
p = 0.0705  0.178\hat{\kappa} + 0.475 \sqrt{\hat{\kappa}}  if \hat{\kappa} \le 1.5  
p = 0.246  if \hat{\kappa} > 1.5  (16) 
where \hat{\kappa}
denotes the estimate of the shape parameter.
Kulkarni and Powar (2010)
derived this equation by determining what power transformation yields a skew closest to 0 and
a kurtosis closest to 3 for a gamma random variable with a given shape parameter.
Although Kulkarni and Powar (2010) use the maximum likelihood estimate of shape to
determine the power to use to induce approximate normality, for the functions
egamma
and egammaAlt
the power is based on whatever estimate of
shape is used (e.g., method="mle"
, method="bcmle"
, etc.).
Likelihood Profile (ci.method="profile.likelihood"
)
This method was proposed by Cox (1970, p.88), and Venzon and Moolgavkar (1988)
introduced an efficient method of computation. This method is also discussed by
Stryhn and Christensen (2003) and Royston (2007).
The idea behind this method is to invert the likelihoodratio test to obtain a
confidence interval for the mean \mu
while treating the coefficient of
variation \tau
as a nuisance parameter.
The likelihood function is given by:
L(\mu, \tau  \underline{x}) = \prod_{i=1}^n \frac{x_i^{\kappa1} e^{x_i/\theta}}{\theta^\kappa \Gamma(\kappa)} \;\;\;\;\;\; (17)
where \kappa
, \theta
, \mu
, and \tau
are defined in
Equations (1)(4) above, and
\Gamma(t)
denotes the Gamma function evaluated at t
.
Following Stryhn and Christensen (2003), denote the maximum likelihood estimates
of the mean and coefficient of variation by (\mu^*, \tau^*)
.
The likelihood ratio test statistic (G^2
) of the hypothesis
H_0: \mu = \mu_0
(where \mu_0
is a fixed value) equals the
drop in 2 log(L)
between the “full” model and the reduced model with
\mu
fixed at \mu_0
, i.e.,
G^2 = 2 \{log[L(\mu^*, \tau^*)]  log[L(\mu_0, \tau_0^*)]\} \;\;\;\;\;\; (18)
where \tau_0^*
is the maximum likelihood estimate of \tau
for the
reduced model (i.e., when \mu = \mu_0
). Under the null hypothesis,
the test statistic G^2
follows a
chisquared distribution with 1 degree of freedom.
Alternatively, we may
express the test statistic in terms of the profile likelihood function L_1
for the mean \mu
, which is obtained from the usual likelihood function by
maximizing over the parameter \tau
, i.e.,
L_1(\mu) = max_{\tau} L(\mu, \tau) \;\;\;\;\;\; (19)
Then we have
G^2 = 2 \{log[L_1(\mu^*)]  log[L_1(\mu_0)]\} \;\;\;\;\;\; (20)
A twosided (1\alpha)100\%
confidence interval for the mean \mu
consists of all values of \mu_0
for which the test is not significant at
level alpha
:
\mu_0: G^2 \le \chi^2_{1, {1\alpha}} \;\;\;\;\;\; (21)
where \chi^2_{\nu, p}
denotes the p
'th quantile of the
chisquared distribution with \nu
degrees of freedom.
Onesided lower and onesided upper confidence intervals are computed in a similar
fashion, except that the quantity 1\alpha
in Equation (21) is replaced with
12\alpha
.
ChiSquare Approximation (ci.method="chisq.approx"
)
This method is based on the relationship between the sample mean of the gamma
distribution and the chisquared distribution (Grice and Bain, 1980):
\frac{2n\bar{x}}{\theta} \sim \chi^2_{2n\kappa} \;\;\;\;\;\; (22)
Therefore, an exact onesided upper (1\alpha)100\%
confidence interval
for the mean \mu
is given by:
[0, \, \frac{2n\bar{x}\kappa}{\chi^2_{2n\kappa, \alpha}}] \;\;\;\;\;\; (23)
an exact onesided lower (1\alpha)100\%
confidence interval
is given by:
[\frac{2n\bar{x}\kappa}{\chi^2_{2n\kappa, 1\alpha}}, \, \infty] \;\;\;\;\;\; (24)
and a twosided (1\alpha)100\%
confidence interval is given by:
[\frac{2n\bar{x}\kappa}{\chi^2_{2n\kappa, 1\alpha/2}},
\, \frac{2n\bar{x}\kappa}{\chi^2_{2n\kappa, \alpha/2}}] \;\;\;\;\;\; (25)
Because this method is exact only when the shape parameter \kappa
is known, the
method used here is called the “chisquare approximation” method
because the estimate of the shape parameter, \hat{\kappa}
, is used in place
of \kappa
in Equations (23)(25) above. The ChiSquare Approximation method
is not the method proposed by Grice and Bain (1980) in which the
confidence interval is adjusted based on adjusting for the fact that the shape
parameter \kappa
is estimated (see the explanation of the
ChiSquare Adjusted method below). The ChiSquare Approximation method used
by egamma
and egammaAlt
is equivalent to the “approximate gamma”
method of ProUCL (USEPA, 2015, equation (234), p.62).
ChiSquare Adjusted (ci.method="chisq.adj"
)
This is the same method as the ChiSquare Approximation method discussed above,
execpt that the value of \alpha
is adjusted to account for the fact that
the shape parameter \kappa
is estimated rather than known. Grice and Bain (1980)
performed Monte Carlo simulations to determine how to adjust \alpha
and the
values in their Table 2 are given in the matrix Grice.Bain.80.mat
.
This method requires that the sample size n
is at least 5 and the confidence level
is between 75% and 99.5% (except when n=5
, in which case the confidence level
must be less than 99%). For values of the sample size n
and/or \alpha
that are not listed in the table, linear interpolation is used (when the sample
size n
is greater than 40, linear interpolation on 1/n
is used, as
recommended by Grice and Bain (1980)). The ChiSquare Adjusted method used
by egamma
and egammaAlt
is equivalent to the “adjusted gamma”
method of ProUCL (USEPA, 2015, equation (235), p.63).
a list of class "estimate"
containing the estimated parameters and other information.
See
estimate.object
for details.
When ci=TRUE
and ci.method="normal.approx"
, it is possible for the
lower confidence limit based on the transformed data to be less than 0.
In this case, the lower confidence limit on the original scale is set to 0 and a warning is
issued stating that the normal approximation is not accurate in this case.
The gamma distribution takes values on the positive real line. Special cases of the gamma are the exponential distribution and the chisquare distributions. Applications of the gamma include life testing, statistical ecology, queuing theory, inventory control, and precipitation processes. A gamma distribution starts to resemble a normal distribution as the shape parameter a tends to infinity.
Some EPA guidance documents (e.g., Singh et al., 2002; Singh et al., 2010a,b) strongly recommend against using a lognormal model for environmental data and recommend trying a gamma distribuiton instead.
Steven P. Millard (EnvStats@ProbStatInfo.com)
Anderson, C.W., and W.D. Ray. (1975). Improved Maximum Likelihood Estimators for the Gamma Distribution. Communications in Statistics, 4, 437–448.
Forbes, C., M. Evans, N. Hastings, and B. Peacock. (2011). Statistical Distributions, Fourth Edition. John Wiley and Sons, Hoboken, NJ.
Grice, J.V., and L.J. Bain. (1980). Inferences Concerning the Mean of the Gamma Distribution. Journal of the American Statistician, 75, 929933.
Hawkins, D. M., and R.A.J. Wixley. (1986). A Note on the Transformation of ChiSquared Variables to Normality. The American Statistician, 40, 296–298.
Johnson, N.L., S. Kotz, and N. Balakrishnan. (1994). Continuous Univariate Distributions, Volume 1. Second Edition. John Wiley and Sons, New York, Chapter 17.
Kulkarni, H.V., and S.K. Powar. (2010). A New Method for Interval Estimation of the Mean of the Gamma Distribution. Lifetime Data Analysis, 16, 431–447.
Singh, A., A.K. Singh, and R.J. Iaci. (2002). Estimation of the Exposure Point Concentration Term Using a Gamma Distribution. EPA/600/R02/084. October 2002. Technology Support Center for Monitoring and Site Characterization, Office of Research and Development, Office of Solid Waste and Emergency Response, U.S. Environmental Protection Agency, Washington, D.C.
USEPA. (2015). ProUCL Version 5.1.002 Technical Guide. EPA/600/R07/041, October 2015. Office of Research and Development. U.S. Environmental Protection Agency, Washington, D.C.
Wilson, E.B., and M.M. Hilferty. (1931). The Distribution of ChiSquares. Proceedings of the National Academy of Sciences, 17, 684–688.
GammaDist
, estimate.object
, eqgamma
,
predIntGamma
, tolIntGamma
.
# Generate 20 observations from a gamma distribution with parameters
# shape=3 and scale=2, then estimate the parameters.
# (Note: the call to set.seed simply allows you to reproduce this
# example.)
set.seed(250)
dat < rgamma(20, shape = 3, scale = 2)
egamma(dat, ci = TRUE)
#Results of Distribution Parameter Estimation
#
#
#Assumed Distribution: Gamma
#
#Estimated Parameter(s): shape = 2.203862
# scale = 2.174928
#
#Estimation Method: mle
#
#Data: dat
#
#Sample Size: 20
#
#Confidence Interval for: mean
#
#Confidence Interval Method: Optimum Power Normal Approximation
# of Kulkarni & Powar (2010)
# using mle of 'shape'
#
#Normal Transform Power: 0.246
#
#Confidence Interval Type: twosided
#
#Confidence Level: 95%
#
#Confidence Interval: LCL = 3.361652
# UCL = 6.746794
# Clean up
rm(dat)
#====================================================================
# Using the reference area TcCB data in EPA.94b.tccb.df, assume a
# gamma distribution, estimate the parameters based on the
# biascorrected mle of shape, and compute a onesided upper 90%
# confidence interval for the mean.
#
# First test to see whether the data appear to follow a gamma
# distribution.
with(EPA.94b.tccb.df,
gofTest(TcCB[Area == "Reference"], dist = "gamma",
est.arg.list = list(method = "bcmle"))
)
#Results of GoodnessofFit Test
#
#
#Test Method: ShapiroWilk GOF Based on
# Chen & Balakrisnan (1995)
#
#Hypothesized Distribution: Gamma
#
#Estimated Parameter(s): shape = 4.5695247
# scale = 0.1309788
#
#Estimation Method: bcmle
#
#Data: TcCB[Area == "Reference"]
#
#Sample Size: 47
#
#Test Statistic: W = 0.9703827
#
#Test Statistic Parameter: n = 47
#
#Pvalue: 0.2739512
#
#Alternative Hypothesis: True cdf does not equal the
# Gamma Distribution.
#
# Now estimate the paramters and compute the upper confidence limit.
with(EPA.94b.tccb.df,
egamma(TcCB[Area == "Reference"], method = "bcmle", ci = TRUE,
ci.type = "upper", conf.level = 0.9)
)
#Results of Distribution Parameter Estimation
#
#
#Assumed Distribution: Gamma
#
#Estimated Parameter(s): shape = 4.5695247
# scale = 0.1309788
#
#Estimation Method: bcmle
#
#Data: TcCB[Area == "Reference"]
#
#Sample Size: 47
#
#Confidence Interval for: mean
#
#Confidence Interval Method: Optimum Power Normal Approximation
# of Kulkarni & Powar (2010)
# using bcmle of 'shape'
#
#Normal Transform Power: 0.246
#
#Confidence Interval Type: upper
#
#Confidence Level: 90%
#
#Confidence Interval: LCL = 0.0000000
# UCL = 0.6561838
#
# Repeat the above example but use the alternative parameterization.
with(EPA.94b.tccb.df,
egammaAlt(TcCB[Area == "Reference"], method = "bcmle", ci = TRUE,
ci.type = "upper", conf.level = 0.9)
)
#Results of Distribution Parameter Estimation
#
#
#Assumed Distribution: Gamma
#
#Estimated Parameter(s): mean = 0.5985106
# cv = 0.4678046
#
#Estimation Method: bcmle of 'shape'
#
#Data: TcCB[Area == "Reference"]
#
#Sample Size: 47
#
#Confidence Interval for: mean
#
#Confidence Interval Method: Optimum Power Normal Approximation
# of Kulkarni & Powar (2010)
# using bcmle of 'shape'
#
#Normal Transform Power: 0.246
#
#Confidence Interval Type: upper
#
#Confidence Level: 90%
#
#Confidence Interval: LCL = 0.0000000
# UCL = 0.6561838
#
# Compare the upper confidence limit based on
# 1) the default method:
# normal approximation method based on Kulkarni and Powar (2010)
# 2) Profile Likelihood
# 3) ChiSquare Approximation
# 4) ChiSquare Adjusted
# Default Method
#
with(EPA.94b.tccb.df,
egamma(TcCB[Area == "Reference"], method = "bcmle", ci = TRUE,
ci.type = "upper", conf.level = 0.9)$interval$limits["UCL"]
)
# UCL
#0.6561838
# Profile Likelihood
#
with(EPA.94b.tccb.df,
egamma(TcCB[Area == "Reference"], method = "mle", ci = TRUE,
ci.type = "upper", conf.level = 0.9,
ci.method = "profile.likelihood")$interval$limits["UCL"]
)
# UCL
#0.6527009
# ChiSquare Approximation
#
with(EPA.94b.tccb.df,
egamma(TcCB[Area == "Reference"], method = "mle", ci = TRUE,
ci.type = "upper", conf.level = 0.9,
ci.method = "chisq.approx")$interval$limits["UCL"]
)
# UCL
#0.6532188
# ChiSquare Adjusted
#
with(EPA.94b.tccb.df,
egamma(TcCB[Area == "Reference"], method = "mle", ci = TRUE,
ci.type = "upper", conf.level = 0.9,
ci.method = "chisq.adj")$interval$limits["UCL"]
)
# UCL
#0.65467