Title: | Actuarial Functions and Heavy Tailed Distributions |
---|---|
Description: | Functions and data sets for actuarial science: modeling of loss distributions; risk theory and ruin theory; simulation of compound models, discrete mixtures and compound hierarchical models; credibility theory. Support for many additional probability distributions to model insurance loss size and frequency: 23 continuous heavy tailed distributions; the Poisson-inverse Gaussian discrete distribution; zero-truncated and zero-modified extensions of the standard discrete distributions. Support for phase-type distributions commonly used to compute ruin probabilities. Main reference: <doi:10.18637/jss.v025.i07>. Implementation of the Feller-Pareto family of distributions: <doi:10.18637/jss.v103.i06>. |
Authors: | Vincent Goulet [cre, aut], Sébastien Auclair [ctb], Christophe Dutang [aut], Walter Garcia-Fontes [ctb], Nicholas Langevin [ctb], Xavier Milhaud [ctb], Tommy Ouellet [ctb], Alexandre Parent [ctb], Mathieu Pigeon [aut], Louis-Philippe Pouliot [ctb], Jeffrey A. Ryan [aut] (Package API), Robert Gentleman [aut] (Parts of the R to C interface), Ross Ihaka [aut] (Parts of the R to C interface), R Core Team [aut] (Parts of the R to C interface), R Foundation [aut] (Parts of the R to C interface) |
Maintainer: | Vincent Goulet <[email protected]> |
License: | GPL (>= 2) |
Version: | 3.3-4 |
Built: | 2024-12-02 06:43:41 UTC |
Source: | CRAN |
Functions and data sets for actuarial science: modeling of loss distributions; risk theory and ruin theory; simulation of compound models, discrete mixtures and compound hierarchical models; credibility theory. Support for many additional probability distributions to model insurance loss size and frequency: 23 continuous heavy tailed distributions; the Poisson-inverse Gaussian discrete distribution; zero-truncated and zero-modified extensions of the standard discrete distributions. Support for phase-type distributions commonly used to compute ruin probabilities. Main reference: <doi:10.18637/jss.v025.i07>. Implementation of the Feller-Pareto family of distributions: <doi:10.18637/jss.v103.i06>.
actuar provides additional actuarial science functionality and support for heavy tailed distributions to the R statistical system.
The current feature set of the package can be split into five main categories.
Additional probability distributions: 23 continuous heavy tailed distributions from the Feller-Pareto and Transformed Gamma families, the loggamma, the Gumbel, the inverse Gaussian and the generalized beta; phase-type distributions; the Poisson-inverse Gaussian discrete distribution; zero-truncated and zero-modified extensions of the standard discrete distributions; computation of raw moments, limited moments and the moment generating function (when it exists) of continuous distributions. See the “distributions” package vignette for details.
Loss distributions modeling: extensive support of grouped data; functions to compute empirical raw and limited moments; support for minimum distance estimation using three different measures; treatment of coverage modifications (deductibles, limits, inflation, coinsurance). See the “modeling” and “coverage” package vignettes for details.
Risk and ruin theory: discretization of the claim amount distribution; calculation of the aggregate claim amount distribution; calculation of the adjustment coefficient; calculation of the probability of ruin, including using phase-type distributions. See the “risk” package vignette for details.
Simulation of discrete mixtures, compound models (including the compound Poisson), and compound hierarchical models. See the “simulation” package vignette for details.
Credibility theory: function cm
fits hierarchical
(including Bühlmann, Bühlmann-Straub), regression and linear Bayes
credibility models. See the “credibility” package vignette
for details.
Christophe Dutang, Vincent Goulet, Mathieu Pigeon and many other
contributors; use packageDescription("actuar")
for the complete
list.
Maintainer: Vincent Goulet.
Dutang, C., Goulet, V. and Pigeon, M. (2008). actuar: An R Package for Actuarial Science. Journal of Statistical Software, 25(7), 1–37. doi:10.18637/jss.v025.i07.
Dutang, C., Goulet, V., Langevin, N. (2022). Feller-Pareto and Related Distributions: Numerical Implementation and Actuarial Applications. Journal of Statistical Software, 103(6), 1–22. doi:10.18637/jss.v103.i06.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
For probability distributions support functions, use as starting
points:
FellerPareto
,
TransformedGamma
,
Loggamma
,
Gumbel
,
InverseGaussian
,
PhaseType
,
PoissonInverseGaussian
and, e.g.,
ZeroTruncatedPoisson
,
ZeroModifiedPoisson
.
For loss modeling support functions:
grouped.data
,
ogive
,
emm
,
elev
,
mde
,
coverage
.
For risk and ruin theory functions:
discretize
,
aggregateDist
,
adjCoef
,
ruin
.
For credibility theory functions and datasets:
cm
,
hachemeister
.
## The package comes with extensive demonstration scripts; ## use the following command to obtain the list. ## Not run: demo(package = "actuar")
## The package comes with extensive demonstration scripts; ## use the following command to obtain the list. ## Not run: demo(package = "actuar")
Compute the adjustment coefficient in ruin theory, or return a function to compute the adjustment coefficient for various reinsurance retentions.
adjCoef(mgf.claim, mgf.wait = mgfexp, premium.rate, upper.bound, h, reinsurance = c("none", "proportional", "excess-of-loss"), from, to, n = 101) ## S3 method for class 'adjCoef' plot(x, xlab = "x", ylab = "R(x)", main = "Adjustment Coefficient", sub = comment(x), type = "l", add = FALSE, ...)
adjCoef(mgf.claim, mgf.wait = mgfexp, premium.rate, upper.bound, h, reinsurance = c("none", "proportional", "excess-of-loss"), from, to, n = 101) ## S3 method for class 'adjCoef' plot(x, xlab = "x", ylab = "R(x)", main = "Adjustment Coefficient", sub = comment(x), type = "l", add = FALSE, ...)
mgf.claim |
an expression written as a function of |
mgf.wait |
an expression written as a function of |
premium.rate |
if |
upper.bound |
numeric; an upper bound for the coefficient, usually the upper bound of the support of the claim severity mgf. |
h |
an expression written as a function of |
reinsurance |
the type of reinsurance for the portfolio; can be abbreviated. |
from , to
|
the range over which the adjustment coefficient will be calculated. |
n |
integer; the number of values at which to evaluate the adjustment coefficient. |
x |
an object of class |
xlab , ylab
|
label of the x and y axes, respectively. |
main |
main title. |
sub |
subtitle, defaulting to the type of reinsurance. |
type |
1-character string giving the type of plot desired; see
|
add |
logical; if |
... |
In the typical case reinsurance = "none"
, the coefficient of
determination is the smallest (strictly) positive root of the Lundberg
equation
on , where
upper.bound
, is the
claim severity random variable,
is the claim interarrival
(or wait) time random variable and
premium.rate
. The
premium rate must satisfy the positive safety loading constraint
.
With reinsurance = "proportional"
, the equation becomes
where is the retention rate and
is the premium rate
function.
With reinsurance = "excess-of-loss"
, the equation becomes
where is the retention limit and
is the premium rate
function.
One can use argument h
as an alternative way to provide
function or
. This is necessary in cases where
random variables
and
are not independent.
The root of is found by minimizing
.
If reinsurance = "none"
, a numeric vector of length one.
Otherwise, a function of class "adjCoef"
inheriting from the
"function"
class.
Christophe Dutang, Vincent Goulet [email protected]
Bowers, N. J. J., Gerber, H. U., Hickman, J., Jones, D. and Nesbitt, C. (1986), Actuarial Mathematics, Society of Actuaries.
Centeno, M. d. L. (2002), Measuring the effects of reinsurance by the adjustment coefficient in the Sparre-Anderson model, Insurance: Mathematics and Economics 30, 37–49.
Gerber, H. U. (1979), An Introduction to Mathematical Risk Theory, Huebner Foundation.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2008), Loss Models, From Data to Decisions, Third Edition, Wiley.
## Basic example: no reinsurance, exponential claim severity and wait ## times, premium rate computed with expected value principle and ## safety loading of 20%. adjCoef(mgfexp, premium = 1.2, upper = 1) ## Same thing, giving function h. h <- function(x) 1/((1 - x) * (1 + 1.2 * x)) adjCoef(h = h, upper = 1) ## Example 11.4 of Klugman et al. (2008) mgfx <- function(x) 0.6 * exp(x) + 0.4 * exp(2 * x) adjCoef(mgfx(x), mgfexp(x, 4), prem = 7, upper = 0.3182) ## Proportional reinsurance, same assumptions as above, reinsurer's ## safety loading of 30%. mgfx <- function(x, y) mgfexp(x * y) p <- function(x) 1.3 * x - 0.1 h <- function(x, a) 1/((1 - a * x) * (1 + x * p(a))) R1 <- adjCoef(mgfx, premium = p, upper = 1, reins = "proportional", from = 0, to = 1, n = 11) R2 <- adjCoef(h = h, upper = 1, reins = "p", from = 0, to = 1, n = 101) R1(seq(0, 1, length = 10)) # evaluation for various retention rates R2(seq(0, 1, length = 10)) # same plot(R1) # graphical representation plot(R2, col = "green", add = TRUE) # smoother function ## Excess-of-loss reinsurance p <- function(x) 1.3 * levgamma(x, 2, 2) - 0.1 mgfx <- function(x, l) mgfgamma(x, 2, 2) * pgamma(l, 2, 2 - x) + exp(x * l) * pgamma(l, 2, 2, lower = FALSE) h <- function(x, l) mgfx(x, l) * mgfexp(-x * p(l)) R1 <- adjCoef(mgfx, upper = 1, premium = p, reins = "excess-of-loss", from = 0, to = 10, n = 11) R2 <- adjCoef(h = h, upper = 1, reins = "e", from = 0, to = 10, n = 101) plot(R1) plot(R2, col = "green", add = TRUE)
## Basic example: no reinsurance, exponential claim severity and wait ## times, premium rate computed with expected value principle and ## safety loading of 20%. adjCoef(mgfexp, premium = 1.2, upper = 1) ## Same thing, giving function h. h <- function(x) 1/((1 - x) * (1 + 1.2 * x)) adjCoef(h = h, upper = 1) ## Example 11.4 of Klugman et al. (2008) mgfx <- function(x) 0.6 * exp(x) + 0.4 * exp(2 * x) adjCoef(mgfx(x), mgfexp(x, 4), prem = 7, upper = 0.3182) ## Proportional reinsurance, same assumptions as above, reinsurer's ## safety loading of 30%. mgfx <- function(x, y) mgfexp(x * y) p <- function(x) 1.3 * x - 0.1 h <- function(x, a) 1/((1 - a * x) * (1 + x * p(a))) R1 <- adjCoef(mgfx, premium = p, upper = 1, reins = "proportional", from = 0, to = 1, n = 11) R2 <- adjCoef(h = h, upper = 1, reins = "p", from = 0, to = 1, n = 101) R1(seq(0, 1, length = 10)) # evaluation for various retention rates R2(seq(0, 1, length = 10)) # same plot(R1) # graphical representation plot(R2, col = "green", add = TRUE) # smoother function ## Excess-of-loss reinsurance p <- function(x) 1.3 * levgamma(x, 2, 2) - 0.1 mgfx <- function(x, l) mgfgamma(x, 2, 2) * pgamma(l, 2, 2 - x) + exp(x * l) * pgamma(l, 2, 2, lower = FALSE) h <- function(x, l) mgfx(x, l) * mgfexp(-x * p(l)) R1 <- adjCoef(mgfx, upper = 1, premium = p, reins = "excess-of-loss", from = 0, to = 10, n = 11) R2 <- adjCoef(h = h, upper = 1, reins = "e", from = 0, to = 10, n = 101) plot(R1) plot(R2, col = "green", add = TRUE)
Compute the aggregate claim amount cumulative distribution function of a portfolio over a period using one of five methods.
aggregateDist(method = c("recursive", "convolution", "normal", "npower", "simulation"), model.freq = NULL, model.sev = NULL, p0 = NULL, x.scale = 1, convolve = 0, moments, nb.simul, ..., tol = 1e-06, maxit = 500, echo = FALSE) ## S3 method for class 'aggregateDist' print(x, ...) ## S3 method for class 'aggregateDist' plot(x, xlim, ylab = expression(F[S](x)), main = "Aggregate Claim Amount Distribution", sub = comment(x), ...) ## S3 method for class 'aggregateDist' summary(object, ...) ## S3 method for class 'aggregateDist' mean(x, ...) ## S3 method for class 'aggregateDist' diff(x, ...)
aggregateDist(method = c("recursive", "convolution", "normal", "npower", "simulation"), model.freq = NULL, model.sev = NULL, p0 = NULL, x.scale = 1, convolve = 0, moments, nb.simul, ..., tol = 1e-06, maxit = 500, echo = FALSE) ## S3 method for class 'aggregateDist' print(x, ...) ## S3 method for class 'aggregateDist' plot(x, xlim, ylab = expression(F[S](x)), main = "Aggregate Claim Amount Distribution", sub = comment(x), ...) ## S3 method for class 'aggregateDist' summary(object, ...) ## S3 method for class 'aggregateDist' mean(x, ...) ## S3 method for class 'aggregateDist' diff(x, ...)
method |
method to be used |
model.freq |
for |
model.sev |
for |
p0 |
arbitrary probability at zero for the frequency
distribution. Creates a zero-modified or zero-truncated
distribution if not |
x.scale |
value of an amount of 1 in the severity model (monetary
unit). Used only with |
convolve |
number of times to convolve the resulting distribution
with itself. Used only with |
moments |
vector of the true moments of the aggregate claim
amount distribution; required only by the |
nb.simul |
number of simulations for the |
... |
parameters of the frequency distribution for the
|
tol |
the resulting cumulative distribution in the
|
maxit |
maximum number of recursions in the |
echo |
logical; echo the recursions to screen in the
|
x , object
|
an object of class |
xlim |
numeric of length 2; the |
ylab |
label of the y axis. |
main |
main title. |
sub |
subtitle, defaulting to the calculation method. |
aggregateDist
returns a function to compute the cumulative
distribution function (cdf) of the aggregate claim amount distribution
in any point.
The "recursive"
method computes the cdf using the Panjer
algorithm; the "convolution"
method using convolutions; the
"normal"
method using a normal approximation; the
"npower"
method using the Normal Power 2 approximation; the
"simulation"
method using simulations. More details follow.
A function of class "aggregateDist"
, inheriting from the
"function"
class when using normal and Normal Power
approximations and additionally inheriting from the "ecdf"
and
"stepfun"
classes when other methods are used.
There are methods available to summarize (summary
), represent
(print
), plot (plot
), compute quantiles
(quantile
) and compute the mean (mean
) of
"aggregateDist"
objects.
For the diff
method: a numeric vector of probabilities
corresponding to the probability mass function evaluated
at the knots of the distribution.
The frequency distribution must be a member of the or
families of discrete distributions.
To use a distribution from the family,
model.freq
must be one of
"binomial"
,
"geometric"
,
"negative binomial"
or
"poisson"
,
and p0
must be NULL
.
To use a zero-truncated distribution from the family,
model.freq
may be one of the strings above together with
p0 = 0
. As a shortcut, model.freq
may also be one of
"zero-truncated binomial"
,
"zero-truncated geometric"
,
"zero-truncated negative binomial"
,
"zero-truncated poisson"
or
"logarithmic"
,
and p0
is then ignored (with a warning if non NULL
).
(Note: since the logarithmic distribution is always zero-truncated.
model.freq = "logarithmic"
may be used with either p0 =
NULL
or p0 = 0
.)
To use a zero-modified distribution from the family,
model.freq
may be one of standard frequency distributions
mentioned above with p0
set to some probability that the
distribution takes the value . It is equivalent, but more
explicit, to set
model.freq
to one of
"zero-modified binomial"
,
"zero-modified geometric"
,
"zero-modified negative binomial"
,
"zero-modified poisson"
or
"zero-modified logarithmic"
.
The parameters of the frequency distribution must be specified using
names identical to the arguments of the appropriate function
dbinom
, dgeom
, dnbinom
,
dpois
or dlogarithmic
. In the latter case,
do take note that the parametrization of dlogarithmic
is
different from Appendix B of Klugman et al. (2012).
If the length of p0
is greater than one, only the first element
is used, with a warning.
model.sev
is a vector of the (discretized) claim amount
distribution ; the first element must be
.
The recursion will fail to start if the expected number of claims is
too large. One may divide the appropriate parameter of the frequency
distribution by and convolve the resulting distribution
convolve
times.
Failure to obtain a cumulative distribution function less than
tol
away from 1 within maxit
iterations is often due
to too coarse a discretization of the severity distribution.
The cumulative distribution function (cdf) of the
aggregate claim amount of a portfolio in the collective risk model is
for ;
is
the frequency probability mass function and
is the cdf of the
th convolution of
the (discrete) claim amount random variable.
model.freq
is vector of the number of claims
probabilities; the first element must be
.
model.sev
is vector of the (discretized)
claim amount distribution; the first element must be
.
The Normal approximation of a cumulative distribution function (cdf)
with mean
and standard deviation
is
The Normal Power 2 approximation of a cumulative distribution function (cdf)
with mean
, standard deviation
and skewness
is
This formula is valid only for the right-hand tail of the distribution and skewness should not exceed unity.
This methods returns the empirical distribution function of a sample
of size nb.simul
of the aggregate claim amount distribution
specified by model.freq
and
model.sev
. rcomphierarc
is used for the simulation of
claim amounts, hence both the frequency and severity models can be
mixtures of distributions.
Vincent Goulet [email protected] and Louis-Philippe Pouliot
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
Daykin, C.D., Pentikäinen, T. and Pesonen, M. (1994), Practical Risk Theory for Actuaries, Chapman & Hall.
discretize
to discretize a severity distribution;
mean.aggregateDist
to compute the mean of the
distribution;
quantile.aggregateDist
to compute the quantiles or the
Value-at-Risk;
CTE.aggregateDist
to compute the Conditional Tail
Expectation (or Tail Value-at-Risk);
rcomphierarc
.
## Convolution method (example 9.5 of Klugman et al. (2012)) fx <- c(0, 0.15, 0.2, 0.25, 0.125, 0.075, 0.05, 0.05, 0.05, 0.025, 0.025) pn <- c(0.05, 0.1, 0.15, 0.2, 0.25, 0.15, 0.06, 0.03, 0.01) Fs <- aggregateDist("convolution", model.freq = pn, model.sev = fx, x.scale = 25) summary(Fs) c(Fs(0), diff(Fs(25 * 0:21))) # probability mass function plot(Fs) ## Recursive method (example 9.10 of Klugman et al. (2012)) fx <- c(0, crossprod(c(2, 1)/3, matrix(c(0.6, 0.7, 0.4, 0, 0, 0.3), 2, 3))) Fs <- aggregateDist("recursive", model.freq = "poisson", model.sev = fx, lambda = 3) plot(Fs) Fs(knots(Fs)) # cdf evaluated at its knots diff(Fs) # probability mass function ## Recursive method (high frequency) fx <- c(0, 0.15, 0.2, 0.25, 0.125, 0.075, 0.05, 0.05, 0.05, 0.025, 0.025) ## Not run: Fs <- aggregateDist("recursive", model.freq = "poisson", model.sev = fx, lambda = 1000) ## End(Not run) Fs <- aggregateDist("recursive", model.freq = "poisson", model.sev = fx, lambda = 250, convolve = 2, maxit = 1500) plot(Fs) ## Recursive method (zero-modified distribution; example 9.11 of ## Klugman et al. (2012)) Fn <- aggregateDist("recursive", model.freq = "binomial", model.sev = c(0.3, 0.5, 0.2), x.scale = 50, p0 = 0.4, size = 3, prob = 0.3) diff(Fn) ## Equivalent but more explicit call aggregateDist("recursive", model.freq = "zero-modified binomial", model.sev = c(0.3, 0.5, 0.2), x.scale = 50, p0 = 0.4, size = 3, prob = 0.3) ## Recursive method (zero-truncated distribution). Using 'fx' above ## would mean that both Pr[N = 0] = 0 and Pr[X = 0] = 0, therefore ## Pr[S = 0] = 0 and recursions would not start. fx <- discretize(pexp(x, 1), from = 0, to = 100, method = "upper") fx[1L] # non zero aggregateDist("recursive", model.freq = "zero-truncated poisson", model.sev = fx, lambda = 3, x.scale = 25, echo=TRUE) ## Normal Power approximation Fs <- aggregateDist("npower", moments = c(200, 200, 0.5)) Fs(210) ## Simulation method model.freq <- expression(data = rpois(3)) model.sev <- expression(data = rgamma(100, 2)) Fs <- aggregateDist("simulation", nb.simul = 1000, model.freq, model.sev) mean(Fs) plot(Fs) ## Evaluation of ruin probabilities using Beekman's formula with ## Exponential(1) claim severity, Poisson(1) frequency and premium rate ## c = 1.2. fx <- discretize(pexp(x, 1), from = 0, to = 100, method = "lower") phi0 <- 0.2/1.2 Fs <- aggregateDist(method = "recursive", model.freq = "geometric", model.sev = fx, prob = phi0) 1 - Fs(400) # approximate ruin probability u <- 0:100 plot(u, 1 - Fs(u), type = "l", main = "Ruin probability")
## Convolution method (example 9.5 of Klugman et al. (2012)) fx <- c(0, 0.15, 0.2, 0.25, 0.125, 0.075, 0.05, 0.05, 0.05, 0.025, 0.025) pn <- c(0.05, 0.1, 0.15, 0.2, 0.25, 0.15, 0.06, 0.03, 0.01) Fs <- aggregateDist("convolution", model.freq = pn, model.sev = fx, x.scale = 25) summary(Fs) c(Fs(0), diff(Fs(25 * 0:21))) # probability mass function plot(Fs) ## Recursive method (example 9.10 of Klugman et al. (2012)) fx <- c(0, crossprod(c(2, 1)/3, matrix(c(0.6, 0.7, 0.4, 0, 0, 0.3), 2, 3))) Fs <- aggregateDist("recursive", model.freq = "poisson", model.sev = fx, lambda = 3) plot(Fs) Fs(knots(Fs)) # cdf evaluated at its knots diff(Fs) # probability mass function ## Recursive method (high frequency) fx <- c(0, 0.15, 0.2, 0.25, 0.125, 0.075, 0.05, 0.05, 0.05, 0.025, 0.025) ## Not run: Fs <- aggregateDist("recursive", model.freq = "poisson", model.sev = fx, lambda = 1000) ## End(Not run) Fs <- aggregateDist("recursive", model.freq = "poisson", model.sev = fx, lambda = 250, convolve = 2, maxit = 1500) plot(Fs) ## Recursive method (zero-modified distribution; example 9.11 of ## Klugman et al. (2012)) Fn <- aggregateDist("recursive", model.freq = "binomial", model.sev = c(0.3, 0.5, 0.2), x.scale = 50, p0 = 0.4, size = 3, prob = 0.3) diff(Fn) ## Equivalent but more explicit call aggregateDist("recursive", model.freq = "zero-modified binomial", model.sev = c(0.3, 0.5, 0.2), x.scale = 50, p0 = 0.4, size = 3, prob = 0.3) ## Recursive method (zero-truncated distribution). Using 'fx' above ## would mean that both Pr[N = 0] = 0 and Pr[X = 0] = 0, therefore ## Pr[S = 0] = 0 and recursions would not start. fx <- discretize(pexp(x, 1), from = 0, to = 100, method = "upper") fx[1L] # non zero aggregateDist("recursive", model.freq = "zero-truncated poisson", model.sev = fx, lambda = 3, x.scale = 25, echo=TRUE) ## Normal Power approximation Fs <- aggregateDist("npower", moments = c(200, 200, 0.5)) Fs(210) ## Simulation method model.freq <- expression(data = rpois(3)) model.sev <- expression(data = rgamma(100, 2)) Fs <- aggregateDist("simulation", nb.simul = 1000, model.freq, model.sev) mean(Fs) plot(Fs) ## Evaluation of ruin probabilities using Beekman's formula with ## Exponential(1) claim severity, Poisson(1) frequency and premium rate ## c = 1.2. fx <- discretize(pexp(x, 1), from = 0, to = 100, method = "lower") phi0 <- 0.2/1.2 Fs <- aggregateDist(method = "recursive", model.freq = "geometric", model.sev = fx, prob = phi0) 1 - Fs(400) # approximate ruin probability u <- 0:100 plot(u, 1 - Fs(u), type = "l", main = "Ruin probability")
The “beta integral” which is just a multiple of the non regularized incomplete beta function. This function merely provides an R interface to the C level routine. It is not exported by the package.
betaint(x, a, b)
betaint(x, a, b)
x |
vector of quantiles. |
a , b
|
parameters. See Details for admissible values. |
Function betaint
computes the “beta integral”
for ,
and
.
(Here
is the function implemented
by R's
gamma()
and defined in its help.)
When ,
where is
pbeta(x, a, b)
. When ,
, and
,
where .
This function is used (at the C level) to compute the
limited expected value for distributions of the transformed beta
family; see, for example, levtrbeta
.
The value of the integral.
Invalid arguments will result in return value NaN
, with a warning.
The need for this function in the package is well explained in the introduction of Appendix A of Klugman et al. (2012). See also chapter 6 and 15 of Abramowitz and Stegun (1972) for definitions and relations to the hypergeometric series.
Vincent Goulet [email protected]
Abramowitz, M. and Stegun, I. A. (1972), Handbook of Mathematical Functions, Dover.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
x <- 0.3 a <- 7 ## case with b > 0 b <- 2 actuar:::betaint(x, a, b) gamma(a) * gamma(b) * pbeta(x, a, b) # same ## case with b < 0 b <- -2.2 r <- floor(-b) # r = 2 actuar:::betaint(x, a, b) ## "manual" calculation s <- (x^(a-1) * (1-x)^b)/b + ((a-1) * x^(a-2) * (1-x)^(b+1))/(b * (b+1)) + ((a-1) * (a-2) * x^(a-3) * (1-x)^(b+2))/(b * (b+1) * (b+2)) -gamma(a+b) * s + (a-1)*(a-2)*(a-3) * gamma(a-r-1)/(b*(b+1)*(b+2)) * gamma(b+r+1)*pbeta(x, a-r-1, b+r+1)
x <- 0.3 a <- 7 ## case with b > 0 b <- 2 actuar:::betaint(x, a, b) gamma(a) * gamma(b) * pbeta(x, a, b) # same ## case with b < 0 b <- -2.2 r <- floor(-b) # r = 2 actuar:::betaint(x, a, b) ## "manual" calculation s <- (x^(a-1) * (1-x)^b)/b + ((a-1) * x^(a-2) * (1-x)^(b+1))/(b * (b+1)) + ((a-1) * (a-2) * x^(a-3) * (1-x)^(b+2))/(b * (b+1) * (b+2)) -gamma(a+b) * s + (a-1)*(a-2)*(a-3) * gamma(a-r-1)/(b*(b+1)*(b+2)) * gamma(b+r+1)*pbeta(x, a-r-1, b+r+1)
Raw moments and limited moments for the (central) Beta distribution
with parameters shape1
and shape2
.
mbeta(order, shape1, shape2) levbeta(limit, shape1, shape2, order = 1)
mbeta(order, shape1, shape2) levbeta(limit, shape1, shape2, order = 1)
order |
order of the moment. |
limit |
limit of the loss variable. |
shape1 , shape2
|
positive parameters of the Beta distribution. |
The th raw moment of the random variable
is
and the
th limited moment at some limit
is
,
.
The noncentral beta distribution is not supported.
mbeta
gives the th raw moment and
levbeta
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a
warning.
Vincent Goulet [email protected] and Mathieu Pigeon
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
Beta
for details on the beta distribution and
functions [dpqr]beta
.
mbeta(2, 3, 4) - mbeta(1, 3, 4)^2 levbeta(10, 3, 4, order = 2)
mbeta(2, 3, 4) - mbeta(1, 3, 4)^2 levbeta(10, 3, 4, order = 2)
Density function, distribution function, quantile function, random generation,
raw moments and limited moments for the Burr distribution with
parameters shape1
, shape2
and scale
.
dburr(x, shape1, shape2, rate = 1, scale = 1/rate, log = FALSE) pburr(q, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qburr(p, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rburr(n, shape1, shape2, rate = 1, scale = 1/rate) mburr(order, shape1, shape2, rate = 1, scale = 1/rate) levburr(limit, shape1, shape2, rate = 1, scale = 1/rate, order = 1)
dburr(x, shape1, shape2, rate = 1, scale = 1/rate, log = FALSE) pburr(q, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qburr(p, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rburr(n, shape1, shape2, rate = 1, scale = 1/rate) mburr(order, shape1, shape2, rate = 1, scale = 1/rate) levburr(limit, shape1, shape2, rate = 1, scale = 1/rate, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
shape1 , shape2 , scale
|
parameters. Must be strictly positive. |
rate |
an alternative way to specify the scale. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The Burr distribution with parameters shape1
,
shape2
and
scale
has density:
for ,
,
and
.
The Burr is the distribution of the random variable
where has a beta distribution with parameters
and
.
The Burr distribution has the following special cases:
A Loglogistic distribution when shape1
== 1
;
A Paralogistic distribution when
shape2 == shape1
;
A Pareto distribution when shape2 ==
1
.
The th raw moment of the random variable
is
,
.
The th limited moment at some limit
is
,
and
not a negative integer.
dburr
gives the density,
pburr
gives the distribution function,
qburr
gives the quantile function,
rburr
generates random deviates,
mburr
gives the th raw moment, and
levburr
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
levburr
computes the limited expected value using
betaint
.
Distribution also known as the Burr Type XII or Singh-Maddala distribution. See also Kleiber and Kotz (2003) for alternative names and parametrizations.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected] and Mathieu Pigeon
Kleiber, C. and Kotz, S. (2003), Statistical Size Distributions in Economics and Actuarial Sciences, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dpareto4
for an equivalent distribution with a location
parameter.
exp(dburr(1, 2, 3, log = TRUE)) p <- (1:10)/10 pburr(qburr(p, 2, 3, 2), 2, 3, 2) ## variance mburr(2, 2, 3, 1) - mburr(1, 2, 3, 1) ^ 2 ## case with shape1 - order/shape2 > 0 levburr(10, 2, 3, 1, order = 2) ## case with shape1 - order/shape2 < 0 levburr(10, 1.5, 0.5, 1, order = 2)
exp(dburr(1, 2, 3, log = TRUE)) p <- (1:10)/10 pburr(qburr(p, 2, 3, 2), 2, 3, 2) ## variance mburr(2, 2, 3, 1) - mburr(1, 2, 3, 1) ^ 2 ## case with shape1 - order/shape2 > 0 levburr(10, 2, 3, 1, order = 2) ## case with shape1 - order/shape2 < 0 levburr(10, 1.5, 0.5, 1, order = 2)
Raw moments, limited moments and moment generating function for the
chi-squared () distribution with
df
degrees
of freedom and optional non-centrality parameter ncp
.
mchisq(order, df, ncp = 0) levchisq(limit, df, ncp = 0, order = 1) mgfchisq(t, df, ncp = 0, log= FALSE)
mchisq(order, df, ncp = 0) levchisq(limit, df, ncp = 0, order = 1) mgfchisq(t, df, ncp = 0, log= FALSE)
order |
order of the moment. |
limit |
limit of the loss variable. |
df |
degrees of freedom (non-negative, but can be non-integer). |
ncp |
non-centrality parameter (non-negative). |
t |
numeric vector. |
log |
logical; if |
The th raw moment of the random variable
is
, the
th limited moment at some limit
is
and the moment generating
function is
.
Only integer moments are supported for the non central Chi-square
distribution (ncp > 0
).
The limited expected value is supported for the centered Chi-square
distribution (ncp = 0
).
mchisq
gives the th raw moment,
levchisq
gives the th moment of the limited loss
variable, and
mgfchisq
gives the moment generating function in t
.
Invalid arguments will result in return value NaN
, with a warning.
Christophe Dutang, Vincent Goulet [email protected]
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
Johnson, N. L. and Kotz, S. (1970), Continuous Univariate Distributions, Volume 1, Wiley.
mchisq(2, 3, 4) levchisq(10, 3, order = 2) mgfchisq(0.25, 3, 2)
mchisq(2, 3, 4) levchisq(10, 3, order = 2) mgfchisq(0.25, 3, 2)
Fit the following credibility models: Bühlmann, Bühlmann-Straub, hierarchical, regression (Hachemeister) or linear Bayes.
cm(formula, data, ratios, weights, subset, regformula = NULL, regdata, adj.intercept = FALSE, method = c("Buhlmann-Gisler", "Ohlsson", "iterative"), likelihood, ..., tol = sqrt(.Machine$double.eps), maxit = 100, echo = FALSE) ## S3 method for class 'cm' print(x, ...) ## S3 method for class 'cm' predict(object, levels = NULL, newdata, ...) ## S3 method for class 'cm' summary(object, levels = NULL, newdata, ...) ## S3 method for class 'summary.cm' print(x, ...)
cm(formula, data, ratios, weights, subset, regformula = NULL, regdata, adj.intercept = FALSE, method = c("Buhlmann-Gisler", "Ohlsson", "iterative"), likelihood, ..., tol = sqrt(.Machine$double.eps), maxit = 100, echo = FALSE) ## S3 method for class 'cm' print(x, ...) ## S3 method for class 'cm' predict(object, levels = NULL, newdata, ...) ## S3 method for class 'cm' summary(object, levels = NULL, newdata, ...) ## S3 method for class 'summary.cm' print(x, ...)
formula |
character string |
data |
a matrix or a data frame containing the portfolio structure, the ratios or claim amounts and their associated weights, if any. |
ratios |
expression indicating the columns of |
weights |
expression indicating the columns of |
subset |
an optional logical expression indicating a subset of observations to be used in the modeling process. All observations are included by default. |
regformula |
an object of class |
regdata |
an optional data frame, list or environment (or object
coercible by |
adj.intercept |
if |
method |
estimation method for the variance components of the model; see Details. |
likelihood |
a character string giving the name of the likelihood function in one of the supported linear Bayes cases; see Details. |
tol |
tolerance level for the stopping criteria for iterative estimation method. |
maxit |
maximum number of iterations in iterative estimation method. |
echo |
logical; whether to echo the iterative procedure or not. |
x , object
|
an object of class |
levels |
character vector indicating the levels to predict or to
include in the summary; if |
newdata |
data frame containing the variables used to predict credibility regression models. |
... |
parameters of the prior distribution for |
cm
is the unified front end for credibility models fitting. The
function supports hierarchical models with any number of levels (with
Bühlmann and Bühlmann-Straub models as
special cases) and the regression model of Hachemeister. Usage of
cm
is similar to lm
for these cases.
cm
can also fit linear Bayes models, in which case usage is
much simplified; see the section on linear Bayes below.
When not "bayes"
, the formula
argument symbolically
describes the structure of the portfolio in the form .
Each term is an interaction between risk factors contributing to the
total variance of the portfolio data. Terms are separated by
+
operators and interactions within each term by :
. For a
portfolio divided first into sectors, then units and finally
contracts, formula
would be ~ sector + sector:unit +
sector:unit:contract
, where sector
, unit
and
contract
are column names in data
. In general, the
formula should be of the form ~ a + a:b + a:b:c + a:b:c:d +
...
.
If argument regformula
is not NULL
, the regression model
of Hachemeister is fit to the data. The response is usually time. By
default, the intercept of the model is located at time origin. If
argument adj.intercept
is TRUE
, the intercept is moved
to the (collective) barycenter of time, by orthogonalization of the
design matrix. Note that the regression coefficients may be difficult
to interpret in this case.
Arguments ratios
, weights
and subset
are used
like arguments select
, select
and subset
,
respectively, of function subset
.
Data does not have to be sorted by level. Nodes with no data (complete
lines of NA
except for the portfolio structure) are allowed,
with the restriction mentioned above.
Function cm
computes the structure parameters estimators of the
model specified in formula
. The value returned is an object of
class cm
.
An object of class "cm"
is a list with at least the following
components:
means |
a list containing, for each level, the vector of linearly sufficient statistics. |
weights |
a list containing, for each level, the vector of total weights. |
unbiased |
a vector containing the unbiased variance components
estimators, or |
iterative |
a vector containing the iterative variance components
estimators, or |
cred |
for multi-level hierarchical models: a list containing, the vector of credibility factors for each level. For one-level models: an array or vector of credibility factors. |
nodes |
a list containing, for each level, the vector of the number of nodes in the level. |
classification |
the columns of |
ordering |
a list containing, for each level, the affiliation of a node to the node of the level above. |
Regression fits have in addition the following components:
adj.models |
a list containing, for each node, the credibility
adjusted regression model as obtained with
|
transition |
if |
terms |
the |
The method of predict
for objects of class "cm"
computes
the credibility premiums for the nodes of every level included in
argument levels
(all by default). Result is a list the same
length as levels
or the number of levels in formula
, or
an atomic vector for one-level models.
The credibility premium at one level is a convex combination between
the linearly sufficient statistic of a node and the credibility
premium of the level above. (For the first level, the complement of
credibility is given to the collective premium.) The linearly
sufficient statistic of a node is the credibility weighted average of
the data of the node, except at the last level, where natural weights
are used. The credibility factor of node is equal to
where is the weight of the node used in the linearly
sufficient statistic,
is the average within node variance and
is the average between node variance.
The credibility premium of node is equal to
where is a matrix created from
newdata
and
is the vector of credibility adjusted regression
coefficients of node
. The latter is given by
where is the vector of regression coefficients based
on data of node
only,
is the vector of collective
regression coefficients,
is the credibility matrix and
is the identity matrix. The credibility matrix of node
is equal to
where is the unscaled regression covariance matrix of
the node,
is the average within node variance and
is the within node covariance matrix.
If the intercept is positioned at the barycenter of time, matrices
and
(and hence
) are diagonal.
This amounts to use Bühlmann-Straub models for each
regression coefficient.
Argument newdata
provides the “future” value of the
regressors for prediction purposes. It should be given as specified in
predict.lm
.
For hierarchical models, two sets of estimators of the variance components (other than the within node variance) are available: unbiased estimators and iterative estimators.
Unbiased estimators are based on sums of squares of the form
and constants of the form
where is the linearly sufficient statistic of
level
;
is the weighted average of
the latter using weights
;
;
is the effective number of
nodes at level
;
is the within variance of this
level. Weights
are the natural weights at the
lowest level, the sum of the natural weights the next level and the
sum of the credibility factors for all upper levels.
The Bühlmann-Gisler estimators (method =
"Buhlmann-Gisler"
) are given by
that is the average of the per node variance estimators truncated at 0.
The Ohlsson estimators (method = "Ohlsson"
) are given by
that is the weighted average of the per node variance estimators without any truncation. Note that negative estimates will be truncated to zero for credibility factor calculations.
In the Bühlmann-Straub model, these estimators are equivalent.
Iterative estimators method = "iterative"
are pseudo-estimators
of the form
where is the linearly sufficient statistic of one
level,
is the linearly sufficient statistic of
the level above and
is the effective number of nodes at one
level minus the effective number of nodes of the level above. The
Ohlsson estimators are used as starting values.
For regression models, with the intercept at time origin, only
iterative estimators are available. If method
is different from
"iterative"
, a warning is issued. With the intercept at the
barycenter of time, the choice of estimators is the same as in the
Bühlmann-Straub model.
When formula
is "bayes"
, the function computes pure
Bayesian premiums for the following combinations of distributions
where they are linear credibility premiums:
and
;
and
;
and
;
and
;
and
;
and
;
and
.
and
.
The following combination is also supported:
and
. In this case, the Bayesian estimator not of
the risk premium, but rather of parameter
is linear with
a “credibility” factor that is not restricted to
.
Argument likelihood
identifies the distribution of as one of
"poisson"
,
"exponential"
,
"gamma"
,
"normal"
,
"bernoulli"
,
"binomial"
,
"geometric"
,
"negative binomial"
or
"pareto"
.
The parameters of the distributions of (when
needed) and
are set in
...
using the argument
names (and default values) of dgamma
,
dnorm
, dbeta
,
dbinom
, dnbinom
or
dpareto1
, as appropriate. For the Gamma/Gamma case, use
shape.lik
for the shape parameter of the Gamma
likelihood. For the Normal/Normal case, use
sd.lik
for the
standard error of the Normal likelihood.
Data for the linear Bayes case may be a matrix or data frame as usual;
an atomic vector to fit the model to a single contract; missing or
NULL
to fit the prior model. Arguments ratios
,
weights
and subset
are ignored.
Vincent Goulet [email protected], Xavier Milhaud, Tommy Ouellet, Louis-Philippe Pouliot
Bühlmann, H. and Gisler, A. (2005), A Course in Credibility Theory and its Applications, Springer.
Belhadj, H., Goulet, V. and Ouellet, T. (2009), On parameter estimation in hierarchical credibility, Astin Bulletin 39.
Goulet, V. (1998), Principles and application of credibility theory, Journal of Actuarial Practice 6, ISSN 1064-6647.
Goovaerts, M. J. and Hoogstad, W. J. (1987), Credibility Theory, Surveys of Actuarial Studies, No. 4, Nationale-Nederlanden N.V.
subset
, formula
,
lm
, predict.lm
.
data(hachemeister) ## Buhlmann-Straub model fit <- cm(~state, hachemeister, ratios = ratio.1:ratio.12, weights = weight.1:weight.12) fit # print method predict(fit) # credibility premiums summary(fit) # more details ## Two-level hierarchical model. Notice that data does not have ## to be sorted by level X <- data.frame(unit = c("A", "B", "A", "B", "B"), hachemeister) fit <- cm(~unit + unit:state, X, ratio.1:ratio.12, weight.1:weight.12) predict(fit) predict(fit, levels = "unit") # unit credibility premiums only summary(fit) summary(fit, levels = "unit") # unit summaries only ## Regression model with intercept at time origin fit <- cm(~state, hachemeister, regformula = ~time, regdata = data.frame(time = 12:1), ratios = ratio.1:ratio.12, weights = weight.1:weight.12) fit predict(fit, newdata = data.frame(time = 0)) summary(fit, newdata = data.frame(time = 0)) ## Same regression model, with intercept at barycenter of time fit <- cm(~state, hachemeister, adj.intercept = TRUE, regformula = ~time, regdata = data.frame(time = 12:1), ratios = ratio.1:ratio.12, weights = weight.1:weight.12) fit predict(fit, newdata = data.frame(time = 0)) summary(fit, newdata = data.frame(time = 0)) ## Poisson/Gamma pure Bayesian model fit <- cm("bayes", data = c(5, 3, 0, 1, 1), likelihood = "poisson", shape = 3, rate = 3) fit predict(fit) summary(fit) ## Normal/Normal pure Bayesian model cm("bayes", data = c(5, 3, 0, 1, 1), likelihood = "normal", sd.lik = 2, mean = 2, sd = 1)
data(hachemeister) ## Buhlmann-Straub model fit <- cm(~state, hachemeister, ratios = ratio.1:ratio.12, weights = weight.1:weight.12) fit # print method predict(fit) # credibility premiums summary(fit) # more details ## Two-level hierarchical model. Notice that data does not have ## to be sorted by level X <- data.frame(unit = c("A", "B", "A", "B", "B"), hachemeister) fit <- cm(~unit + unit:state, X, ratio.1:ratio.12, weight.1:weight.12) predict(fit) predict(fit, levels = "unit") # unit credibility premiums only summary(fit) summary(fit, levels = "unit") # unit summaries only ## Regression model with intercept at time origin fit <- cm(~state, hachemeister, regformula = ~time, regdata = data.frame(time = 12:1), ratios = ratio.1:ratio.12, weights = weight.1:weight.12) fit predict(fit, newdata = data.frame(time = 0)) summary(fit, newdata = data.frame(time = 0)) ## Same regression model, with intercept at barycenter of time fit <- cm(~state, hachemeister, adj.intercept = TRUE, regformula = ~time, regdata = data.frame(time = 12:1), ratios = ratio.1:ratio.12, weights = weight.1:weight.12) fit predict(fit, newdata = data.frame(time = 0)) summary(fit, newdata = data.frame(time = 0)) ## Poisson/Gamma pure Bayesian model fit <- cm("bayes", data = c(5, 3, 0, 1, 1), likelihood = "poisson", shape = 3, rate = 3) fit predict(fit) summary(fit) ## Normal/Normal pure Bayesian model cm("bayes", data = c(5, 3, 0, 1, 1), likelihood = "normal", sd.lik = 2, mean = 2, sd = 1)
Compute probability density function or cumulative distribution function of the payment per payment or payment per loss random variable under any combination of the following coverage modifications: deductible, limit, coinsurance, inflation.
coverage(pdf, cdf, deductible = 0, franchise = FALSE, limit = Inf, coinsurance = 1, inflation = 0, per.loss = FALSE)
coverage(pdf, cdf, deductible = 0, franchise = FALSE, limit = Inf, coinsurance = 1, inflation = 0, per.loss = FALSE)
pdf , cdf
|
function object or character string naming a function to compute, respectively, the probability density function and cumulative distribution function of a probability law. |
deductible |
a unique positive numeric value. |
franchise |
logical; |
limit |
a unique positive numeric value larger than
|
coinsurance |
a unique value between 0 and 1; the proportion of coinsurance. |
inflation |
a unique value between 0 and 1; the rate of inflation. |
per.loss |
logical; |
coverage
returns a function to compute the probability
density function (pdf) or the cumulative distribution function (cdf)
of the distribution of losses under coverage modifications. The pdf
and cdf of unmodified losses are pdf
and cdf
,
respectively.
If pdf
is specified, the pdf is returned; if pdf
is
missing or NULL
, the cdf is returned. Note that cdf
is
needed if there is a deductible or a limit.
An object of mode "function"
with the same arguments as
pdf
or cdf
, except "lower.tail"
,
"log.p"
and "log"
, which are not supported.
Setting arguments of the function returned by coverage
using
formals
may very well not work as expected.
Vincent Goulet [email protected] and Mathieu Pigeon
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
vignette("coverage")
for the exact definitions of the
per payment and per loss random variables under an ordinary or
franchise deductible.
## Default case: pdf of the per payment random variable with ## an ordinary deductible coverage(dgamma, pgamma, deductible = 1) ## Add a limit f <- coverage(dgamma, pgamma, deductible = 1, limit = 7) f <- coverage("dgamma", "pgamma", deductible = 1, limit = 7) # same f(0, shape = 3, rate = 1) f(2, shape = 3, rate = 1) f(6, shape = 3, rate = 1) f(8, shape = 3, rate = 1) curve(dgamma(x, 3, 1), xlim = c(0, 10), ylim = c(0, 0.3)) # original curve(f(x, 3, 1), xlim = c(0.01, 5.99), col = 4, add = TRUE) # modified points(6, f(6, 3, 1), pch = 21, bg = 4) ## Cumulative distribution function F <- coverage(cdf = pgamma, deductible = 1, limit = 7) F(0, shape = 3, rate = 1) F(2, shape = 3, rate = 1) F(6, shape = 3, rate = 1) F(8, shape = 3, rate = 1) curve(pgamma(x, 3, 1), xlim = c(0, 10), ylim = c(0, 1)) # original curve(F(x, 3, 1), xlim = c(0, 5.99), col = 4, add = TRUE) # modified curve(F(x, 3, 1), xlim = c(6, 10), col = 4, add = TRUE) # modified ## With no deductible, all distributions below are identical coverage(dweibull, pweibull, limit = 5) coverage(dweibull, pweibull, per.loss = TRUE, limit = 5) coverage(dweibull, pweibull, franchise = TRUE, limit = 5) coverage(dweibull, pweibull, per.loss = TRUE, franchise = TRUE, limit = 5) ## Coinsurance alone; only case that does not require the cdf coverage(dgamma, coinsurance = 0.8)
## Default case: pdf of the per payment random variable with ## an ordinary deductible coverage(dgamma, pgamma, deductible = 1) ## Add a limit f <- coverage(dgamma, pgamma, deductible = 1, limit = 7) f <- coverage("dgamma", "pgamma", deductible = 1, limit = 7) # same f(0, shape = 3, rate = 1) f(2, shape = 3, rate = 1) f(6, shape = 3, rate = 1) f(8, shape = 3, rate = 1) curve(dgamma(x, 3, 1), xlim = c(0, 10), ylim = c(0, 0.3)) # original curve(f(x, 3, 1), xlim = c(0.01, 5.99), col = 4, add = TRUE) # modified points(6, f(6, 3, 1), pch = 21, bg = 4) ## Cumulative distribution function F <- coverage(cdf = pgamma, deductible = 1, limit = 7) F(0, shape = 3, rate = 1) F(2, shape = 3, rate = 1) F(6, shape = 3, rate = 1) F(8, shape = 3, rate = 1) curve(pgamma(x, 3, 1), xlim = c(0, 10), ylim = c(0, 1)) # original curve(F(x, 3, 1), xlim = c(0, 5.99), col = 4, add = TRUE) # modified curve(F(x, 3, 1), xlim = c(6, 10), col = 4, add = TRUE) # modified ## With no deductible, all distributions below are identical coverage(dweibull, pweibull, limit = 5) coverage(dweibull, pweibull, per.loss = TRUE, limit = 5) coverage(dweibull, pweibull, franchise = TRUE, limit = 5) coverage(dweibull, pweibull, per.loss = TRUE, franchise = TRUE, limit = 5) ## Coinsurance alone; only case that does not require the cdf coverage(dgamma, coinsurance = 0.8)
Conditional Tail Expectation, also called Tail Value-at-Risk.
TVaR
is an alias for CTE
.
CTE(x, ...) ## S3 method for class 'aggregateDist' CTE(x, conf.level = c(0.9, 0.95, 0.99), names = TRUE, ...) TVaR(x, ...)
CTE(x, ...) ## S3 method for class 'aggregateDist' CTE(x, conf.level = c(0.9, 0.95, 0.99), names = TRUE, ...) TVaR(x, ...)
x |
an R object. |
conf.level |
numeric vector of probabilities with values
in |
names |
logical; if true, the result has a |
... |
further arguments passed to or from other methods. |
The Conditional Tail Expectation (or Tail Value-at-Risk) measures the
average of losses above the Value at Risk for some given confidence
level, that is where
is the loss random
variable.
CTE
is a generic function with, currently, only a method for
objects of class "aggregateDist"
.
For the recursive, convolution and simulation methods of
aggregateDist
, the CTE is computed from the definition
using the empirical cdf.
For the normal approximation method, an explicit formula exists:
where is the mean,
the standard
deviation and
the confidence level.
For the Normal Power approximation, the explicit formula given in Castañer et al. (2013) is
where, as above, is the mean,
the standard
deviation,
the confidence level and
is
the skewness.
A numeric vector, named if names
is TRUE
.
Vincent Goulet [email protected] and Tommy Ouellet
Castañer, A. and Claramunt, M.M. and Mármol, M. (2013), Tail value at risk. An analysis with the Normal-Power approximation. In Statistical and Soft Computing Approaches in Insurance Problems, pp. 87-112. Nova Science Publishers, 2013. ISBN 978-1-62618-506-7.
model.freq <- expression(data = rpois(7)) model.sev <- expression(data = rnorm(9, 2)) Fs <- aggregateDist("simulation", model.freq, model.sev, nb.simul = 1000) CTE(Fs)
model.freq <- expression(data = rpois(7)) model.sev <- expression(data = rnorm(9, 2)) Fs <- aggregateDist("simulation", model.freq, model.sev, nb.simul = 1000) CTE(Fs)
Basic dental claims on a policy with a deductible of 50.
dental
dental
A vector containing 10 observations
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (1998), Loss Models, From Data to Decisions, Wiley.
Compute a discrete probability mass function from a continuous cumulative distribution function (cdf) with various methods.
discretise
is an alias for discretize
.
discretize(cdf, from, to, step = 1, method = c("upper", "lower", "rounding", "unbiased"), lev, by = step, xlim = NULL) discretise(cdf, from, to, step = 1, method = c("upper", "lower", "rounding", "unbiased"), lev, by = step, xlim = NULL)
discretize(cdf, from, to, step = 1, method = c("upper", "lower", "rounding", "unbiased"), lev, by = step, xlim = NULL) discretise(cdf, from, to, step = 1, method = c("upper", "lower", "rounding", "unbiased"), lev, by = step, xlim = NULL)
cdf |
an expression written as a function of |
from , to
|
the range over which the function will be discretized. |
step |
numeric; the discretization step (or span, or lag). |
method |
discretization method to use. |
lev |
an expression written as a function of |
by |
an alias for |
xlim |
numeric of length 2; if specified, it serves as default
for |
Usage is similar to curve
.
discretize
returns the probability mass function (pmf) of the
random variable obtained by discretization of the cdf specified in
cdf
.
Let denote the cdf,
the
limited expected value at
,
the step,
the probability mass at
in the discretized distribution and
set
from
and
to
.
Method "upper"
is the forward difference of the cdf :
for .
Method "lower"
is the backward difference of the cdf :
for and
.
Method "rounding"
has the true cdf pass through the
midpoints of the intervals :
for and
. The function assumes the cdf is continuous. Any
adjusment necessary for discrete distributions can be done via
cdf
.
Method "unbiased"
matches the first moment of the discretized
and the true distributions. The probabilities are as follows:
A numeric vector of probabilities suitable for use in
aggregateDist
.
Vincent Goulet [email protected]
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
x <- seq(0, 5, 0.5) op <- par(mfrow = c(1, 1), col = "black") ## Upper and lower discretization fu <- discretize(pgamma(x, 1), method = "upper", from = 0, to = 5, step = 0.5) fl <- discretize(pgamma(x, 1), method = "lower", from = 0, to = 5, step = 0.5) curve(pgamma(x, 1), xlim = c(0, 5)) par(col = "blue") plot(stepfun(head(x, -1), diffinv(fu)), pch = 19, add = TRUE) par(col = "green") plot(stepfun(x, diffinv(fl)), pch = 19, add = TRUE) par(col = "black") ## Rounding (or midpoint) discretization fr <- discretize(pgamma(x, 1), method = "rounding", from = 0, to = 5, step = 0.5) curve(pgamma(x, 1), xlim = c(0, 5)) par(col = "blue") plot(stepfun(head(x, -1), diffinv(fr)), pch = 19, add = TRUE) par(col = "black") ## First moment matching fb <- discretize(pgamma(x, 1), method = "unbiased", lev = levgamma(x, 1), from = 0, to = 5, step = 0.5) curve(pgamma(x, 1), xlim = c(0, 5)) par(col = "blue") plot(stepfun(x, diffinv(fb)), pch = 19, add = TRUE) par(op)
x <- seq(0, 5, 0.5) op <- par(mfrow = c(1, 1), col = "black") ## Upper and lower discretization fu <- discretize(pgamma(x, 1), method = "upper", from = 0, to = 5, step = 0.5) fl <- discretize(pgamma(x, 1), method = "lower", from = 0, to = 5, step = 0.5) curve(pgamma(x, 1), xlim = c(0, 5)) par(col = "blue") plot(stepfun(head(x, -1), diffinv(fu)), pch = 19, add = TRUE) par(col = "green") plot(stepfun(x, diffinv(fl)), pch = 19, add = TRUE) par(col = "black") ## Rounding (or midpoint) discretization fr <- discretize(pgamma(x, 1), method = "rounding", from = 0, to = 5, step = 0.5) curve(pgamma(x, 1), xlim = c(0, 5)) par(col = "blue") plot(stepfun(head(x, -1), diffinv(fr)), pch = 19, add = TRUE) par(col = "black") ## First moment matching fb <- discretize(pgamma(x, 1), method = "unbiased", lev = levgamma(x, 1), from = 0, to = 5, step = 0.5) curve(pgamma(x, 1), xlim = c(0, 5)) par(col = "blue") plot(stepfun(x, diffinv(fb)), pch = 19, add = TRUE) par(op)
Compute the empirical limited expected value for individual or grouped data.
elev(x, ...) ## Default S3 method: elev(x, ...) ## S3 method for class 'grouped.data' elev(x, ...) ## S3 method for class 'elev' print(x, digits = getOption("digits") - 2, ...) ## S3 method for class 'elev' summary(object, ...) ## S3 method for class 'elev' knots(Fn, ...) ## S3 method for class 'elev' plot(x, ..., main = NULL, xlab = "x", ylab = "Empirical LEV")
elev(x, ...) ## Default S3 method: elev(x, ...) ## S3 method for class 'grouped.data' elev(x, ...) ## S3 method for class 'elev' print(x, digits = getOption("digits") - 2, ...) ## S3 method for class 'elev' summary(object, ...) ## S3 method for class 'elev' knots(Fn, ...) ## S3 method for class 'elev' plot(x, ..., main = NULL, xlab = "x", ylab = "Empirical LEV")
x |
a vector or an object of class |
digits |
number of significant digits to use, see
|
Fn , object
|
an R object inheriting from |
main |
main title. |
xlab , ylab
|
labels of x and y axis. |
... |
arguments to be passed to subsequent methods. |
The limited expected value (LEV) at of a random variable
is
. For
individual data
, the
empirical LEV
is thus
Methods of elev
exist for individual data or for grouped data
created with grouped.data
. The formula in this case is
too long to show here. See the reference for details.
For elev
, a function of class "elev"
, inheriting from the
"function"
class.
Vincent Goulet [email protected] and Mathieu Pigeon
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (1998), Loss Models, From Data to Decisions, Wiley.
grouped.data
to create grouped data objects;
stepfun
for related documentation (even though the
empirical LEV is not a step function).
data(gdental) lev <- elev(gdental) lev summary(lev) knots(lev) # the group boundaries lev(knots(lev)) # empirical lev at boundaries lev(c(80, 200, 2000)) # and at other limits plot(lev, type = "o", pch = 16)
data(gdental) lev <- elev(gdental) lev summary(lev) knots(lev) # the group boundaries lev(knots(lev)) # empirical lev at boundaries lev(c(80, 200, 2000)) # and at other limits plot(lev, type = "o", pch = 16)
Raw empirical moments for individual and grouped data.
emm(x, order = 1, ...) ## Default S3 method: emm(x, order = 1, ...) ## S3 method for class 'grouped.data' emm(x, order = 1, ...)
emm(x, order = 1, ...) ## Default S3 method: emm(x, order = 1, ...) ## S3 method for class 'grouped.data' emm(x, order = 1, ...)
x |
a vector or matrix of individual data, or an object of class
|
order |
order of the moment. Must be positive. |
... |
further arguments passed to or from other methods. |
Arguments ...
are passed to colMeans
;
na.rm = TRUE
may be useful for individual data with missing
values.
For individual data, the th empirical moment is
.
For grouped data with group boundaries and group frequencies
, the
th empirical moment is
where .
A named vector or matrix of moments.
Vincent Goulet [email protected] and Mathieu Pigeon
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (1998), Loss Models, From Data to Decisions, Wiley.
mean
and mean.grouped.data
for simpler
access to the first moment.
## Individual data data(dental) emm(dental, order = 1:3) ## Grouped data data(gdental) emm(gdental) x <- grouped.data(cj = gdental[, 1], nj1 = sample(1:100, nrow(gdental)), nj2 = sample(1:100, nrow(gdental))) emm(x) # same as mean(x)
## Individual data data(dental) emm(dental, order = 1:3) ## Grouped data data(gdental) emm(gdental) x <- grouped.data(cj = gdental[, 1], nj1 = sample(1:100, nrow(gdental)), nj2 = sample(1:100, nrow(gdental))) emm(x) # same as mean(x)
Raw moments, limited moments and moment generating function for the
exponential distribution with rate rate
(i.e., mean
1/rate
).
mexp(order, rate = 1) levexp(limit, rate = 1, order = 1) mgfexp(t, rate = 1, log = FALSE)
mexp(order, rate = 1) levexp(limit, rate = 1, order = 1) mgfexp(t, rate = 1, log = FALSE)
order |
order of the moment. |
limit |
limit of the loss variable. |
rate |
vector of rates. |
t |
numeric vector. |
log |
logical; if |
The th raw moment of the random variable
is
, the
th limited moment at some limit
is
and the moment
generating function is
,
.
mexp
gives the th raw moment,
levexp
gives the th moment of the limited loss
variable, and
mgfexp
gives the moment generating function in t
.
Invalid arguments will result in return value NaN
, with a warning.
Vincent Goulet [email protected], Christophe Dutang and Mathieu Pigeon.
Johnson, N. L. and Kotz, S. (1970), Continuous Univariate Distributions, Volume 1, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
mexp(2, 3) - mexp(1, 3)^2 levexp(10, 3, order = 2) mgfexp(1,2)
mexp(2, 3) - mexp(1, 3)^2 levexp(10, 3, order = 2) mgfexp(1,2)
Extract or replace subsets of grouped data objects.
## S3 method for class 'grouped.data' x[i, j] ## S3 replacement method for class 'grouped.data' x[i, j] <- value
## S3 method for class 'grouped.data' x[i, j] ## S3 replacement method for class 'grouped.data' x[i, j] <- value
x |
an object of class |
i , j
|
elements to extract or replace. |
value |
a suitable replacement value. |
Objects of class "grouped.data"
can mostly be indexed like data
frames, with the following restrictions:
For [
, the extracted object must keep a group
boundaries column and at least one group frequencies column to
remain of class "grouped.data"
;
For [<-
, it is not possible to replace group boundaries
and group frequencies simultaneously;
When replacing group boundaries, length(value) ==
length(i) + 1
.
x[, 1]
will return the plain vector of group boundaries.
Replacement of non adjacent group boundaries is not possible for obvious reasons.
Otherwise, extraction and replacement should work just like for data frames.
For [
an object of class "grouped.data"
, a data frame or a
vector.
For [<-
an object of class "grouped.data"
.
Currently [[
, [[<-
, $
and $<-
are not
specifically supported, but should work as usual on group frequency
columns.
Vincent Goulet [email protected]
[.data.frame
for extraction and replacement
methods of data frames, grouped.data
to create grouped
data objects.
data(gdental) (x <- gdental[1]) # select column 1 class(x) # no longer a grouped.data object class(gdental[2]) # same gdental[, 1] # group boundaries gdental[, 2] # group frequencies gdental[1:4,] # a subset gdental[c(1, 3, 5),] # avoid this gdental[1:2, 1] <- c(0, 30, 60) # modified boundaries gdental[, 2] <- 10 # modified frequencies ## Not run: gdental[1, ] <- 2 # not allowed
data(gdental) (x <- gdental[1]) # select column 1 class(x) # no longer a grouped.data object class(gdental[2]) # same gdental[, 1] # group boundaries gdental[, 2] # group frequencies gdental[1:4,] # a subset gdental[c(1, 3, 5),] # avoid this gdental[1:2, 1] <- c(0, 30, 60) # modified boundaries gdental[, 2] <- 10 # modified frequencies ## Not run: gdental[1, ] <- 2 # not allowed
Density function, distribution function, quantile function, random generation,
raw moments and limited moments for the Feller Pareto distribution
with parameters min
, shape1
, shape2
, shape3
and
scale
.
dfpareto(x, min, shape1, shape2, shape3, rate = 1, scale = 1/rate, log = FALSE) pfpareto(q, min, shape1, shape2, shape3, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qfpareto(p, min, shape1, shape2, shape3, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rfpareto(n, min, shape1, shape2, shape3, rate = 1, scale = 1/rate) mfpareto(order, min, shape1, shape2, shape3, rate = 1, scale = 1/rate) levfpareto(limit, min, shape1, shape2, shape3, rate = 1, scale = 1/rate, order = 1)
dfpareto(x, min, shape1, shape2, shape3, rate = 1, scale = 1/rate, log = FALSE) pfpareto(q, min, shape1, shape2, shape3, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qfpareto(p, min, shape1, shape2, shape3, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rfpareto(n, min, shape1, shape2, shape3, rate = 1, scale = 1/rate) mfpareto(order, min, shape1, shape2, shape3, rate = 1, scale = 1/rate) levfpareto(limit, min, shape1, shape2, shape3, rate = 1, scale = 1/rate, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
min |
lower bound of the support of the distribution. |
shape1 , shape2 , shape3 , scale
|
parameters. Must be strictly positive. |
rate |
an alternative way to specify the scale. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The Feller-Pareto distribution with parameters min
,
shape1
,
shape2
,
shape3
and
scale
, has
density:
for ,
,
,
,
and
.
(Here
is the function implemented
by R's
gamma()
and defined in its help.)
The Feller-Pareto is the distribution of the random variable
where has a beta distribution with parameters
and
.
The Feller-Pareto defines a large family of distributions encompassing
the transformed beta family and many variants of the Pareto
distribution. Setting yields the
transformed beta distribution.
The Feller-Pareto distribution also has the following direct special cases:
A Pareto IV distribution when shape3
== 1
;
A Pareto III distribution when shape1
shape3 == 1
;
A Pareto II distribution when shape1
shape2 == 1
;
A Pareto I distribution when shape1
shape2 == 1
and min = scale
.
The th raw moment of the random variable
is
for nonnegative integer values of
.
The th limited moment at some limit
is
for nonnegative integer values of
and
,
not a negative integer.
Note that the range of admissible values for in raw and
limited moments is larger when
.
dfpareto
gives the density,
pfpareto
gives the distribution function,
qfpareto
gives the quantile function,
rfpareto
generates random deviates,
mfpareto
gives the th raw moment, and
levfpareto
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
levfpareto
computes the limited expected value using
betaint
.
For the Feller-Pareto and other Pareto distributions, we use the classification of Arnold (2015) with the parametrization of Klugman et al. (2012).
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected] and Nicholas Langevin
Dutang, C., Goulet, V., Langevin, N. (2022). Feller-Pareto and Related Distributions: Numerical Implementation and Actuarial Applications. Journal of Statistical Software, 103(6), 1–22. doi:10.18637/jss.v103.i06.
Abramowitz, M. and Stegun, I. A. (1972), Handbook of Mathematical Functions, Dover.
Arnold, B. C. (2015), Pareto Distributions, Second Edition, CRC Press.
Kleiber, C. and Kotz, S. (2003), Statistical Size Distributions in Economics and Actuarial Sciences, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dtrbeta
for the transformed beta distribution.
exp(dfpareto(2, 1, 2, 3, 4, 5, log = TRUE)) p <- (1:10)/10 pfpareto(qfpareto(p, 1, 2, 3, 4, 5), 1, 2, 3, 4, 5) ## variance mfpareto(2, 1, 2, 3, 4, 5) - mfpareto(1, 1, 2, 3, 4, 5)^2 ## case with shape1 - order/shape2 > 0 levfpareto(10, 1, 2, 3, 4, scale = 1, order = 2) ## case with shape1 - order/shape2 < 0 levfpareto(20, 10, 0.1, 14, 2, scale = 1.5, order = 2)
exp(dfpareto(2, 1, 2, 3, 4, 5, log = TRUE)) p <- (1:10)/10 pfpareto(qfpareto(p, 1, 2, 3, 4, 5), 1, 2, 3, 4, 5) ## variance mfpareto(2, 1, 2, 3, 4, 5) - mfpareto(1, 1, 2, 3, 4, 5)^2 ## case with shape1 - order/shape2 > 0 levfpareto(10, 1, 2, 3, 4, scale = 1, order = 2) ## case with shape1 - order/shape2 < 0 levfpareto(20, 10, 0.1, 14, 2, scale = 1.5, order = 2)
Raw moments, limited moments and moment generating function for the
Gamma distribution with parameters shape
and scale
.
mgamma(order, shape, rate = 1, scale = 1/rate) levgamma(limit, shape, rate = 1, scale = 1/rate, order = 1) mgfgamma(t, shape, rate = 1, scale = 1/rate, log = FALSE)
mgamma(order, shape, rate = 1, scale = 1/rate) levgamma(limit, shape, rate = 1, scale = 1/rate, order = 1) mgfgamma(t, shape, rate = 1, scale = 1/rate, log = FALSE)
order |
order of the moment. |
limit |
limit of the loss variable. |
rate |
an alternative way to specify the scale. |
shape , scale
|
shape and scale parameters. Must be strictly positive. |
t |
numeric vector. |
log |
logical; if |
The th raw moment of the random variable
is
, the
th limited moment at some limit
is
and the moment
generating function is
,
.
mgamma
gives the th raw moment,
levgamma
gives the th moment of the limited loss
variable, and
mgfgamma
gives the moment generating function in t
.
Invalid arguments will result in return value NaN
, with a warning.
Vincent Goulet [email protected], Christophe Dutang and Mathieu Pigeon
Johnson, N. L. and Kotz, S. (1970), Continuous Univariate Distributions, Volume 1, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
mgamma(2, 3, 4) - mgamma(1, 3, 4)^2 levgamma(10, 3, 4, order = 2) mgfgamma(1,3,2)
mgamma(2, 3, 4) - mgamma(1, 3, 4)^2 levgamma(10, 3, 4, order = 2) mgfgamma(1,3,2)
Grouped dental claims, that is presented in a number of claims per claim amount group form.
gdental
gdental
An object of class "grouped.data"
(inheriting from class
"data.frame"
) consisting of 10 rows and 2 columns. The
environment of the object contains the plain vector of cj
of
group boundaries
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (1998), Loss Models, From Data to Decisions, Wiley.
grouped.data
for a description of grouped data objects.
Density function, distribution function, quantile function, random generation,
raw moments and limited moments for the Generalized Beta distribution
with parameters shape1
, shape2
, shape3
and
scale
.
dgenbeta(x, shape1, shape2, shape3, rate = 1, scale = 1/rate, log = FALSE) pgenbeta(q, shape1, shape2, shape3, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qgenbeta(p, shape1, shape2, shape3, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rgenbeta(n, shape1, shape2, shape3, rate = 1, scale = 1/rate) mgenbeta(order, shape1, shape2, shape3, rate = 1, scale = 1/rate) levgenbeta(limit, shape1, shape2, shape3, rate = 1, scale = 1/rate, order = 1)
dgenbeta(x, shape1, shape2, shape3, rate = 1, scale = 1/rate, log = FALSE) pgenbeta(q, shape1, shape2, shape3, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qgenbeta(p, shape1, shape2, shape3, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rgenbeta(n, shape1, shape2, shape3, rate = 1, scale = 1/rate) mgenbeta(order, shape1, shape2, shape3, rate = 1, scale = 1/rate) levgenbeta(limit, shape1, shape2, shape3, rate = 1, scale = 1/rate, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
shape1 , shape2 , shape3 , scale
|
parameters. Must be strictly positive. |
rate |
an alternative way to specify the scale. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The generalized beta distribution with parameters shape1
,
shape2
,
shape3
and
scale
, has
density:
for ,
,
,
and
. (Here
is the function implemented
by R's
gamma()
and defined in its help.)
The generalized beta is the distribution of the random variable
where has a beta distribution with parameters
and
.
The th raw moment of the random variable
is
and the
th limited moment at some limit
is
,
.
dgenbeta
gives the density,
pgenbeta
gives the distribution function,
qgenbeta
gives the quantile function,
rgenbeta
generates random deviates,
mgenbeta
gives the th raw moment, and
levgenbeta
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
This is not the generalized three-parameter beta distribution defined on page 251 of Johnson et al, 1995.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected]
Johnson, N. L., Kotz, S. and Balakrishnan, N. (1995) Continuous Univariate Distributions, Volume 2, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
exp(dgenbeta(2, 2, 3, 4, 0.2, log = TRUE)) p <- (1:10)/10 pgenbeta(qgenbeta(p, 2, 3, 4, 0.2), 2, 3, 4, 0.2) mgenbeta(2, 1, 2, 3, 0.25) - mgenbeta(1, 1, 2, 3, 0.25) ^ 2 levgenbeta(10, 1, 2, 3, 0.25, order = 2)
exp(dgenbeta(2, 2, 3, 4, 0.2, log = TRUE)) p <- (1:10)/10 pgenbeta(qgenbeta(p, 2, 3, 4, 0.2), 2, 3, 4, 0.2) mgenbeta(2, 1, 2, 3, 0.25) - mgenbeta(1, 1, 2, 3, 0.25) ^ 2 levgenbeta(10, 1, 2, 3, 0.25, order = 2)
Density function, distribution function, quantile function, random generation,
raw moments and limited moments for the Generalized Pareto
distribution with parameters shape1
, shape2
and
scale
.
dgenpareto(x, shape1, shape2, rate = 1, scale = 1/rate, log = FALSE) pgenpareto(q, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qgenpareto(p, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rgenpareto(n, shape1, shape2, rate = 1, scale = 1/rate) mgenpareto(order, shape1, shape2, rate = 1, scale = 1/rate) levgenpareto(limit, shape1, shape2, rate = 1, scale = 1/rate, order = 1)
dgenpareto(x, shape1, shape2, rate = 1, scale = 1/rate, log = FALSE) pgenpareto(q, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qgenpareto(p, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rgenpareto(n, shape1, shape2, rate = 1, scale = 1/rate) mgenpareto(order, shape1, shape2, rate = 1, scale = 1/rate) levgenpareto(limit, shape1, shape2, rate = 1, scale = 1/rate, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
shape1 , shape2 , scale
|
parameters. Must be strictly positive. |
rate |
an alternative way to specify the scale. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The Generalized Pareto distribution with parameters shape1
,
shape2
and
scale
has density:
for ,
,
and
.
(Here
is the function implemented
by R's
gamma()
and defined in its help.)
The Generalized Pareto is the distribution of the random variable
where has a beta distribution with parameters
and
.
The Generalized Pareto distribution has the following special cases:
A Pareto distribution when shape2 ==
1
;
An Inverse Pareto distribution when
shape1 == 1
.
The th raw moment of the random variable
is
,
.
The th limited moment at some limit
is
,
and
not a
negative integer.
dgenpareto
gives the density,
pgenpareto
gives the distribution function,
qgenpareto
gives the quantile function,
rgenpareto
generates random deviates,
mgenpareto
gives the th raw moment, and
levgenpareto
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
levgenpareto
computes the limited expected value using
betaint
.
Distribution also known as the Beta of the Second Kind. See also Kleiber and Kotz (2003) for alternative names and parametrizations.
The Generalized Pareto distribution defined here is different from the one in Embrechts et al. (1997) and in Wikipedia; see also Kleiber and Kotz (2003, section 3.12). One may most likely compute quantities for the latter using functions for the Pareto distribution with the appropriate change of parametrization.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected] and Mathieu Pigeon
Embrechts, P., Klüppelberg, C. and Mikisch, T. (1997), Modelling Extremal Events for Insurance and Finance, Springer.
Kleiber, C. and Kotz, S. (2003), Statistical Size Distributions in Economics and Actuarial Sciences, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
exp(dgenpareto(3, 3, 4, 4, log = TRUE)) p <- (1:10)/10 pgenpareto(qgenpareto(p, 3, 3, 1), 3, 3, 1) qgenpareto(.3, 3, 4, 4, lower.tail = FALSE) ## variance mgenpareto(2, 3, 2, 1) - mgenpareto(1, 3, 2, 1)^2 ## case with shape1 - order > 0 levgenpareto(10, 3, 3, scale = 1, order = 2) ## case with shape1 - order < 0 levgenpareto(10, 1.5, 3, scale = 1, order = 2)
exp(dgenpareto(3, 3, 4, 4, log = TRUE)) p <- (1:10)/10 pgenpareto(qgenpareto(p, 3, 3, 1), 3, 3, 1) qgenpareto(.3, 3, 4, 4, lower.tail = FALSE) ## variance mgenpareto(2, 3, 2, 1) - mgenpareto(1, 3, 2, 1)^2 ## case with shape1 - order > 0 levgenpareto(10, 3, 3, scale = 1, order = 2) ## case with shape1 - order < 0 levgenpareto(10, 1.5, 3, scale = 1, order = 2)
Creation of grouped data objects, from either a provided set of group boundaries and group frequencies, or from individual data using automatic or specified breakpoints.
grouped.data(..., breaks = "Sturges", include.lowest = TRUE, right = TRUE, nclass = NULL, group = FALSE, row.names = NULL, check.rows = FALSE, check.names = TRUE)
grouped.data(..., breaks = "Sturges", include.lowest = TRUE, right = TRUE, nclass = NULL, group = FALSE, row.names = NULL, check.rows = FALSE, check.names = TRUE)
... |
arguments of the form |
breaks |
same as for
In the last three cases the number is a suggestion only; the
breakpoints will be set to |
include.lowest |
logical; if |
right |
logical; indicating if the intervals should be closed on the right (and open on the left) or vice versa. |
nclass |
numeric (integer); equivalent to |
group |
logical; an alternative way to force grouping of individual data. |
row.names , check.rows , check.names
|
arguments identical to those
of |
A grouped data object is a special form of data frame consisting of one column of contiguous group boundaries and one or more columns of frequencies within each group.
The function can create a grouped data object from two types of arguments.
Group boundaries and frequencies. This is the default mode of
operation if the call has at least two elements in ...
.
The first argument will then be taken as the vector of group boundaries. This vector must be exactly one element longer than the other arguments, which will be taken as vectors of group frequencies. All arguments are coerced to data frames.
Individual data. This mode of operation is active if there
is a single argument in ...
, or if either breaks
or nclass
is specified or group
is TRUE
.
Arguments of ...
are first grouped using
hist
. If needed, breakpoints are set using the first
argument.
Missing (NA
) frequencies are replaced by zeros, with a
warning.
Extraction and replacement methods exist for grouped.data
objects, but working on non adjacent groups will most likely yield
useless results.
An object of class
c("grouped.data", "data.frame")
with
an environment containing the vector cj
of group boundaries.
Vincent Goulet [email protected], Mathieu Pigeon and Louis-Philippe Pouliot
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (1998), Loss Models, From Data to Decisions, Wiley.
[.grouped.data
for extraction and replacement methods.
data.frame
for usual data frame creation and
manipulation.
hist
for details on the calculation of breakpoints.
## Most common usage using a predetermined set of group ## boundaries and group frequencies. cj <- c(0, 25, 50, 100, 250, 500, 1000) nj <- c(30, 31, 57, 42, 45, 10) (x <- grouped.data(Group = cj, Frequency = nj)) class(x) x[, 1] # group boundaries x[, 2] # group frequencies ## Multiple frequency columns are supported x <- sample(1:100, 9) y <- sample(1:100, 9) grouped.data(cj = 1:10, nj.1 = x, nj.2 = y) ## Alternative usage with grouping of individual data. grouped.data(x) # automatic breakpoints grouped.data(x, breaks = 7) # forced number of groups grouped.data(x, breaks = c(0,25,75,100)) # specified groups grouped.data(x, y, breaks = c(0,25,75,100)) # multiple data sets ## Not run: ## Providing two or more data sets and automatic breakpoints is ## very error-prone since the range of the first data set has to ## include the ranges of all the other data sets. range(x) range(y) grouped.data(x, y, group = TRUE) ## End(Not run)
## Most common usage using a predetermined set of group ## boundaries and group frequencies. cj <- c(0, 25, 50, 100, 250, 500, 1000) nj <- c(30, 31, 57, 42, 45, 10) (x <- grouped.data(Group = cj, Frequency = nj)) class(x) x[, 1] # group boundaries x[, 2] # group frequencies ## Multiple frequency columns are supported x <- sample(1:100, 9) y <- sample(1:100, 9) grouped.data(cj = 1:10, nj.1 = x, nj.2 = y) ## Alternative usage with grouping of individual data. grouped.data(x) # automatic breakpoints grouped.data(x, breaks = 7) # forced number of groups grouped.data(x, breaks = c(0,25,75,100)) # specified groups grouped.data(x, y, breaks = c(0,25,75,100)) # multiple data sets ## Not run: ## Providing two or more data sets and automatic breakpoints is ## very error-prone since the range of the first data set has to ## include the ranges of all the other data sets. range(x) range(y) grouped.data(x, y, group = TRUE) ## End(Not run)
Density function, distribution function, quantile function, random
generation and raw moments for the Gumbel extreme value distribution
with parameters alpha
and scale
.
dgumbel(x, alpha, scale, log = FALSE) pgumbel(q, alpha, scale, lower.tail = TRUE, log.p = FALSE) qgumbel(p, alpha, scale, lower.tail = TRUE, log.p = FALSE) rgumbel(n, alpha, scale) mgumbel(order, alpha, scale) mgfgumbel(t, alpha, scale, log = FALSE)
dgumbel(x, alpha, scale, log = FALSE) pgumbel(q, alpha, scale, lower.tail = TRUE, log.p = FALSE) qgumbel(p, alpha, scale, lower.tail = TRUE, log.p = FALSE) rgumbel(n, alpha, scale) mgumbel(order, alpha, scale) mgfgumbel(t, alpha, scale, log = FALSE)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
alpha |
location parameter. |
scale |
parameter. Must be strictly positive. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. Only values |
t |
numeric vector. |
The Gumbel distribution with parameters alpha
and
scale
has distribution
function:
for ,
and
.
The mode of the distribution is in , the mean is
, where
is the Euler-Mascheroni constant, and the variance is
.
dgumbel
gives the density,
pgumbel
gives the distribution function,
qgumbel
gives the quantile function,
rgumbel
generates random deviates,
mgumbel
gives the th raw moment,
, and
mgfgamma
gives the moment generating function in t
.
Invalid arguments will result in return value NaN
, with a warning.
Distribution also knonw as the generalized extreme value distribution Type-I.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected]
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dgumbel(c(-5, 0, 10, 20), 0.5, 2) p <- (1:10)/10 pgumbel(qgumbel(p, 2, 3), 2, 3) curve(pgumbel(x, 0.5, 2), from = -5, to = 20, col = "red") curve(pgumbel(x, 1.0, 2), add = TRUE, col = "green") curve(pgumbel(x, 1.5, 3), add = TRUE, col = "blue") curve(pgumbel(x, 3.0, 4), add = TRUE, col = "cyan") a <- 3; s <- 4 mgumbel(1, a, s) # mean a - s * digamma(1) # same mgumbel(2, a, s) - mgumbel(1, a, s)^2 # variance (pi * s)^2/6 # same
dgumbel(c(-5, 0, 10, 20), 0.5, 2) p <- (1:10)/10 pgumbel(qgumbel(p, 2, 3), 2, 3) curve(pgumbel(x, 0.5, 2), from = -5, to = 20, col = "red") curve(pgumbel(x, 1.0, 2), add = TRUE, col = "green") curve(pgumbel(x, 1.5, 3), add = TRUE, col = "blue") curve(pgumbel(x, 3.0, 4), add = TRUE, col = "cyan") a <- 3; s <- 4 mgumbel(1, a, s) # mean a - s * digamma(1) # same mgumbel(2, a, s) - mgumbel(1, a, s)^2 # variance (pi * s)^2/6 # same
Hachemeister (1975) data set giving average claim amounts in private passenger bodily injury insurance in five U.S. states over 12 quarters between July 1970 and June 1973 and the corresponding number of claims.
hachemeister
hachemeister
A matrix with 5 rows and the following 25 columns:
state
the state number;
ratio.1
, ..., ratio.12
the average claim amounts;
weight.1
, ..., weight.12
the corresponding number of claims.
Hachemeister, C. A. (1975), Credibility for regression models with application to trend, Proceedings of the Berkeley Actuarial Research Conference on Credibility, Academic Press.
This method for the generic function hist
is mainly
useful to plot the histogram of grouped data. If plot = FALSE
,
the resulting object of class "histogram"
is returned for
compatibility with hist.default
, but does not contain
much information not already in x
.
## S3 method for class 'grouped.data' hist(x, freq = NULL, probability = !freq, density = NULL, angle = 45, col = NULL, border = NULL, main = paste("Histogram of" , xname), xlim = range(x), ylim = NULL, xlab = xname, ylab, axes = TRUE, plot = TRUE, labels = FALSE, ...)
## S3 method for class 'grouped.data' hist(x, freq = NULL, probability = !freq, density = NULL, angle = 45, col = NULL, border = NULL, main = paste("Histogram of" , xname), xlim = range(x), ylim = NULL, xlab = xname, ylab, axes = TRUE, plot = TRUE, labels = FALSE, ...)
x |
an object of class |
freq |
logical; if |
probability |
an alias for |
density |
the density of shading lines, in lines per inch.
The default value of |
angle |
the slope of shading lines, given as an angle in degrees (counter-clockwise). |
col |
a colour to be used to fill the bars.
The default of |
border |
the color of the border around the bars. The default is to use the standard foreground color. |
main , xlab , ylab
|
these arguments to |
xlim , ylim
|
the range of x and y values with sensible defaults.
Note that |
axes |
logical. If |
plot |
logical. If |
labels |
logical or character. Additionally draw labels on top
of bars, if not |
... |
further graphical parameters passed to
|
An object of class "histogram"
which is a list with components:
breaks |
the |
counts |
|
density |
the relative frequencies within each group
|
intensities |
same as |
mids |
the |
xname |
a character string with the actual |
equidist |
logical, indicating if the distances between
|
The resulting value does not depend on the values of
the arguments freq
(or probability
)
or plot
. This is intentionally different from S.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (1998), Loss Models, From Data to Decisions, Wiley.
hist
and hist.default
for histograms of
individual data and fancy examples.
data(gdental) hist(gdental)
data(gdental) hist(gdental)
Density function, distribution function, quantile function, random
generation, raw moments and limited moments for the Inverse Burr
distribution with parameters shape1
, shape2
and
scale
.
dinvburr(x, shape1, shape2, rate = 1, scale = 1/rate, log = FALSE) pinvburr(q, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qinvburr(p, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rinvburr(n, shape1, shape2, rate = 1, scale = 1/rate) minvburr(order, shape1, shape2, rate = 1, scale = 1/rate) levinvburr(limit, shape1, shape2, rate = 1, scale = 1/rate, order = 1)
dinvburr(x, shape1, shape2, rate = 1, scale = 1/rate, log = FALSE) pinvburr(q, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qinvburr(p, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rinvburr(n, shape1, shape2, rate = 1, scale = 1/rate) minvburr(order, shape1, shape2, rate = 1, scale = 1/rate) levinvburr(limit, shape1, shape2, rate = 1, scale = 1/rate, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
shape1 , shape2 , scale
|
parameters. Must be strictly positive. |
rate |
an alternative way to specify the scale. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The inverse Burr distribution with parameters shape1
,
shape2
and
scale
, has density:
for ,
,
and
.
The inverse Burr is the distribution of the random variable
where has a beta distribution with parameters
and
.
The inverse Burr distribution has the following special cases:
A Loglogistic distribution when shape1
== 1
;
An Inverse Pareto distribution when
shape2 == 1
;
An Inverse Paralogistic distribution
when shape1 == shape2
.
The th raw moment of the random variable
is
,
.
The th limited moment at some limit
is
,
and
not a negative integer.
dinvburr
gives the density,
invburr
gives the distribution function,
qinvburr
gives the quantile function,
rinvburr
generates random deviates,
minvburr
gives the th raw moment, and
levinvburr
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
levinvburr
computes the limited expected value using
betaint
.
Also known as the Dagum distribution. See also Kleiber and Kotz (2003) for alternative names and parametrizations.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected] and Mathieu Pigeon
Kleiber, C. and Kotz, S. (2003), Statistical Size Distributions in Economics and Actuarial Sciences, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
exp(dinvburr(2, 2, 3, 1, log = TRUE)) p <- (1:10)/10 pinvburr(qinvburr(p, 2, 3, 1), 2, 3, 1) ## variance minvburr(2, 2, 3, 1) - minvburr(1, 2, 3, 1) ^ 2 ## case with 1 - order/shape2 > 0 levinvburr(10, 2, 3, 1, order = 2) ## case with 1 - order/shape2 < 0 levinvburr(10, 2, 1.5, 1, order = 2)
exp(dinvburr(2, 2, 3, 1, log = TRUE)) p <- (1:10)/10 pinvburr(qinvburr(p, 2, 3, 1), 2, 3, 1) ## variance minvburr(2, 2, 3, 1) - minvburr(1, 2, 3, 1) ^ 2 ## case with 1 - order/shape2 > 0 levinvburr(10, 2, 3, 1, order = 2) ## case with 1 - order/shape2 < 0 levinvburr(10, 2, 1.5, 1, order = 2)
Density function, distribution function, quantile function, random generation
raw moments and limited moments for the Inverse Exponential
distribution with parameter scale
.
dinvexp(x, rate = 1, scale = 1/rate, log = FALSE) pinvexp(q, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qinvexp(p, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rinvexp(n, rate = 1, scale = 1/rate) minvexp(order, rate = 1, scale = 1/rate) levinvexp(limit, rate = 1, scale = 1/rate, order)
dinvexp(x, rate = 1, scale = 1/rate, log = FALSE) pinvexp(q, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qinvexp(p, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rinvexp(n, rate = 1, scale = 1/rate) minvexp(order, rate = 1, scale = 1/rate) levinvexp(limit, rate = 1, scale = 1/rate, order)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
scale |
parameter. Must be strictly positive. |
rate |
an alternative way to specify the scale. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The inverse exponential distribution with parameter scale
has density:
for and
.
The th raw moment of the random variable
is
,
, and the
th limited moment at
some limit
is
, all
.
dinvexp
gives the density,
pinvexp
gives the distribution function,
qinvexp
gives the quantile function,
rinvexp
generates random deviates,
minvexp
gives the th raw moment, and
levinvexp
calculates the th limited moment.
Invalid arguments will result in return value NaN
, with a warning.
levinvexp
computes the limited expected value using
gammainc
from package expint.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected] and Mathieu Pigeon
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
exp(dinvexp(2, 2, log = TRUE)) p <- (1:10)/10 pinvexp(qinvexp(p, 2), 2) minvexp(0.5, 2)
exp(dinvexp(2, 2, log = TRUE)) p <- (1:10)/10 pinvexp(qinvexp(p, 2), 2) minvexp(0.5, 2)
Density function, distribution function, quantile function, random generation,
raw moments, and limited moments for the Inverse Gamma distribution
with parameters shape
and scale
.
dinvgamma(x, shape, rate = 1, scale = 1/rate, log = FALSE) pinvgamma(q, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qinvgamma(p, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rinvgamma(n, shape, rate = 1, scale = 1/rate) minvgamma(order, shape, rate = 1, scale = 1/rate) levinvgamma(limit, shape, rate = 1, scale = 1/rate, order = 1) mgfinvgamma(t, shape, rate =1, scale = 1/rate, log =FALSE)
dinvgamma(x, shape, rate = 1, scale = 1/rate, log = FALSE) pinvgamma(q, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qinvgamma(p, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rinvgamma(n, shape, rate = 1, scale = 1/rate) minvgamma(order, shape, rate = 1, scale = 1/rate) levinvgamma(limit, shape, rate = 1, scale = 1/rate, order = 1) mgfinvgamma(t, shape, rate =1, scale = 1/rate, log =FALSE)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
shape , scale
|
parameters. Must be strictly positive. |
rate |
an alternative way to specify the scale. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
t |
numeric vector. |
The inverse gamma distribution with parameters shape
and
scale
has density:
for ,
and
.
(Here
is the function implemented
by R's
gamma()
and defined in its help.)
The special case shape == 1
is an
Inverse Exponential distribution.
The th raw moment of the random variable
is
,
, and the
th
limited moment at some limit
is
, all
.
The moment generating function is given by .
dinvgamma
gives the density,
pinvgamma
gives the distribution function,
qinvgamma
gives the quantile function,
rinvgamma
generates random deviates,
minvgamma
gives the th raw moment,
levinvgamma
gives the th moment of the limited loss
variable, and
mgfinvgamma
gives the moment generating function in t
.
Invalid arguments will result in return value NaN
, with a warning.
levinvgamma
computes the limited expected value using
gammainc
from package expint.
Also known as the Vinci distribution. See also Kleiber and Kotz (2003) for alternative names and parametrizations.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected] and Mathieu Pigeon
Kleiber, C. and Kotz, S. (2003), Statistical Size Distributions in Economics and Actuarial Sciences, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
exp(dinvgamma(2, 3, 4, log = TRUE)) p <- (1:10)/10 pinvgamma(qinvgamma(p, 2, 3), 2, 3) minvgamma(-1, 2, 2) ^ 2 levinvgamma(10, 2, 2, order = 1) mgfinvgamma(-1, 3, 2)
exp(dinvgamma(2, 3, 4, log = TRUE)) p <- (1:10)/10 pinvgamma(qinvgamma(p, 2, 3), 2, 3) minvgamma(-1, 2, 2) ^ 2 levinvgamma(10, 2, 2, order = 1) mgfinvgamma(-1, 3, 2)
Density function, distribution function, quantile function, random
generation, raw moments, limited moments and moment generating
function for the Inverse Gaussian distribution with parameters
mean
and shape
.
dinvgauss(x, mean, shape = 1, dispersion = 1/shape, log = FALSE) pinvgauss(q, mean, shape = 1, dispersion = 1/shape, lower.tail = TRUE, log.p = FALSE) qinvgauss(p, mean, shape = 1, dispersion = 1/shape, lower.tail = TRUE, log.p = FALSE, tol = 1e-14, maxit = 100, echo = FALSE, trace = echo) rinvgauss(n, mean, shape = 1, dispersion = 1/shape) minvgauss(order, mean, shape = 1, dispersion = 1/shape) levinvgauss(limit, mean, shape = 1, dispersion = 1/shape, order = 1) mgfinvgauss(t, mean, shape = 1, dispersion = 1/shape, log = FALSE)
dinvgauss(x, mean, shape = 1, dispersion = 1/shape, log = FALSE) pinvgauss(q, mean, shape = 1, dispersion = 1/shape, lower.tail = TRUE, log.p = FALSE) qinvgauss(p, mean, shape = 1, dispersion = 1/shape, lower.tail = TRUE, log.p = FALSE, tol = 1e-14, maxit = 100, echo = FALSE, trace = echo) rinvgauss(n, mean, shape = 1, dispersion = 1/shape) minvgauss(order, mean, shape = 1, dispersion = 1/shape) levinvgauss(limit, mean, shape = 1, dispersion = 1/shape, order = 1) mgfinvgauss(t, mean, shape = 1, dispersion = 1/shape, log = FALSE)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
mean , shape
|
parameters. Must be strictly positive. Infinite values are supported. |
dispersion |
an alternative way to specify the shape. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. Only |
limit |
limit of the loss variable. |
tol |
small positive value. Tolerance to assess convergence in the Newton computation of quantiles. |
maxit |
positive integer; maximum number of recursions in the Newton computation of quantiles. |
echo , trace
|
logical; echo the recursions to screen in the Newton computation of quantiles. |
t |
numeric vector. |
The inverse Gaussian distribution with parameters mean
and
dispersion
has density:
for ,
and
.
The limiting case is an inverse
chi-squared distribution (or inverse gamma with
shape
and
rate
phi
). This distribution has no
finite strictly positive, integer moments.
The limiting case is an infinite spike in
.
If the random variable is IG
, then
is IG
.
The th raw moment of the random variable
is
,
, the limited expected
value at some limit
is
and
the moment generating function is
.
The moment generating function of the inverse guassian is defined for
t <= 1/(2 * mean^2 * phi)
.
dinvgauss
gives the density,
pinvgauss
gives the distribution function,
qinvgauss
gives the quantile function,
rinvgauss
generates random deviates,
minvgauss
gives the th raw moment,
levinvgauss
gives the limited expected value, and
mgfinvgauss
gives the moment generating function in t
.
Invalid arguments will result in return value NaN
, with a warning.
Functions dinvgauss
, pinvgauss
and qinvgauss
are
C implementations of functions of the same name in package
statmod; see Giner and Smyth (2016).
Devroye (1986, chapter 4) provides a nice presentation of the algorithm to generate random variates from an inverse Gaussian distribution.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected]
Giner, G. and Smyth, G. K. (2016), “statmod: Probability Calculations for the Inverse Gaussian Distribution”, R Journal, vol. 8, no 1, p. 339-351. https://journal.r-project.org/archive/2016-1/giner-smyth.pdf
Chhikara, R. S. and Folk, T. L. (1989), The Inverse Gaussian Distribution: Theory, Methodology and Applications, Decker.
Devroye, L. (1986), Non-Uniform Random Variate Generation, Springer-Verlag. http://luc.devroye.org/rnbookindex.html
dinvgamma
for the inverse gamma distribution.
dinvgauss(c(-1, 0, 1, 2, Inf), mean = 1.5, dis = 0.7) dinvgauss(c(-1, 0, 1, 2, Inf), mean = Inf, dis = 0.7) dinvgauss(c(-1, 0, 1, 2, Inf), mean = 1.5, dis = Inf) # spike at zero ## Typical graphical representations of the inverse Gaussian ## distribution. First fixed mean and varying shape; second ## varying mean and fixed shape. col = c("red", "blue", "green", "cyan", "yellow", "black") par = c(0.125, 0.5, 1, 2, 8, 32) curve(dinvgauss(x, 1, par[1]), from = 0, to = 2, col = col[1]) for (i in 2:6) curve(dinvgauss(x, 1, par[i]), add = TRUE, col = col[i]) curve(dinvgauss(x, par[1], 1), from = 0, to = 2, col = col[1]) for (i in 2:6) curve(dinvgauss(x, par[i], 1), add = TRUE, col = col[i]) pinvgauss(qinvgauss((1:10)/10, 1.5, shape = 2), 1.5, 2) minvgauss(1:4, 1.5, 2) levinvgauss(c(0, 0.5, 1, 1.2, 10, Inf), 1.5, 2)
dinvgauss(c(-1, 0, 1, 2, Inf), mean = 1.5, dis = 0.7) dinvgauss(c(-1, 0, 1, 2, Inf), mean = Inf, dis = 0.7) dinvgauss(c(-1, 0, 1, 2, Inf), mean = 1.5, dis = Inf) # spike at zero ## Typical graphical representations of the inverse Gaussian ## distribution. First fixed mean and varying shape; second ## varying mean and fixed shape. col = c("red", "blue", "green", "cyan", "yellow", "black") par = c(0.125, 0.5, 1, 2, 8, 32) curve(dinvgauss(x, 1, par[1]), from = 0, to = 2, col = col[1]) for (i in 2:6) curve(dinvgauss(x, 1, par[i]), add = TRUE, col = col[i]) curve(dinvgauss(x, par[1], 1), from = 0, to = 2, col = col[1]) for (i in 2:6) curve(dinvgauss(x, par[i], 1), add = TRUE, col = col[i]) pinvgauss(qinvgauss((1:10)/10, 1.5, shape = 2), 1.5, 2) minvgauss(1:4, 1.5, 2) levinvgauss(c(0, 0.5, 1, 1.2, 10, Inf), 1.5, 2)
Density function, distribution function, quantile function, random generation,
raw moments and limited moments for the Inverse Paralogistic
distribution with parameters shape
and scale
.
dinvparalogis(x, shape, rate = 1, scale = 1/rate, log = FALSE) pinvparalogis(q, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qinvparalogis(p, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rinvparalogis(n, shape, rate = 1, scale = 1/rate) minvparalogis(order, shape, rate = 1, scale = 1/rate) levinvparalogis(limit, shape, rate = 1, scale = 1/rate, order = 1)
dinvparalogis(x, shape, rate = 1, scale = 1/rate, log = FALSE) pinvparalogis(q, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qinvparalogis(p, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rinvparalogis(n, shape, rate = 1, scale = 1/rate) minvparalogis(order, shape, rate = 1, scale = 1/rate) levinvparalogis(limit, shape, rate = 1, scale = 1/rate, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
shape , scale
|
parameters. Must be strictly positive. |
rate |
an alternative way to specify the scale. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The inverse paralogistic distribution with parameters shape
and
scale
has density:
for ,
and
.
The th raw moment of the random variable
is
,
.
The th limited moment at some limit
is
,
and
not a negative integer.
dinvparalogis
gives the density,
pinvparalogis
gives the distribution function,
qinvparalogis
gives the quantile function,
rinvparalogis
generates random deviates,
minvparalogis
gives the th raw moment, and
levinvparalogis
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
levinvparalogis
computes computes the limited expected value using
betaint
.
See Kleiber and Kotz (2003) for alternative names and parametrizations.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected] and Mathieu Pigeon
Kleiber, C. and Kotz, S. (2003), Statistical Size Distributions in Economics and Actuarial Sciences, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
exp(dinvparalogis(2, 3, 4, log = TRUE)) p <- (1:10)/10 pinvparalogis(qinvparalogis(p, 2, 3), 2, 3) ## first negative moment minvparalogis(-1, 2, 2) ## case with 1 - order/shape > 0 levinvparalogis(10, 2, 2, order = 1) ## case with 1 - order/shape < 0 levinvparalogis(10, 2/3, 2, order = 1)
exp(dinvparalogis(2, 3, 4, log = TRUE)) p <- (1:10)/10 pinvparalogis(qinvparalogis(p, 2, 3), 2, 3) ## first negative moment minvparalogis(-1, 2, 2) ## case with 1 - order/shape > 0 levinvparalogis(10, 2, 2, order = 1) ## case with 1 - order/shape < 0 levinvparalogis(10, 2/3, 2, order = 1)
Density function, distribution function, quantile function, random generation
raw moments and limited moments for the Inverse Pareto distribution
with parameters shape
and scale
.
dinvpareto(x, shape, scale, log = FALSE) pinvpareto(q, shape, scale, lower.tail = TRUE, log.p = FALSE) qinvpareto(p, shape, scale, lower.tail = TRUE, log.p = FALSE) rinvpareto(n, shape, scale) minvpareto(order, shape, scale) levinvpareto(limit, shape, scale, order = 1)
dinvpareto(x, shape, scale, log = FALSE) pinvpareto(q, shape, scale, lower.tail = TRUE, log.p = FALSE) qinvpareto(p, shape, scale, lower.tail = TRUE, log.p = FALSE) rinvpareto(n, shape, scale) minvpareto(order, shape, scale) levinvpareto(limit, shape, scale, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
shape , scale
|
parameters. Must be strictly positive. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The inverse Pareto distribution with parameters shape
and
scale
has density:
for ,
and
.
The th raw moment of the random variable
is
,
.
The th limited moment at some limit
is
,
.
dinvpareto
gives the density,
pinvpareto
gives the distribution function,
qinvpareto
gives the quantile function,
rinvpareto
generates random deviates,
minvpareto
gives the th raw moment, and
levinvpareto
calculates the th limited moment.
Invalid arguments will result in return value NaN
, with a warning.
Evaluation of levinvpareto
is done using numerical integration.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected] and Mathieu Pigeon
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
exp(dinvpareto(2, 3, 4, log = TRUE)) p <- (1:10)/10 pinvpareto(qinvpareto(p, 2, 3), 2, 3) minvpareto(0.5, 1, 2)
exp(dinvpareto(2, 3, 4, log = TRUE)) p <- (1:10)/10 pinvpareto(qinvpareto(p, 2, 3), 2, 3) minvpareto(0.5, 1, 2)
Density function, distribution function, quantile function, random generation,
raw moments, and limited moments for the Inverse Transformed Gamma
distribution with parameters shape1
, shape2
and
scale
.
dinvtrgamma(x, shape1, shape2, rate = 1, scale = 1/rate, log = FALSE) pinvtrgamma(q, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qinvtrgamma(p, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rinvtrgamma(n, shape1, shape2, rate = 1, scale = 1/rate) minvtrgamma(order, shape1, shape2, rate = 1, scale = 1/rate) levinvtrgamma(limit, shape1, shape2, rate = 1, scale = 1/rate, order = 1)
dinvtrgamma(x, shape1, shape2, rate = 1, scale = 1/rate, log = FALSE) pinvtrgamma(q, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qinvtrgamma(p, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rinvtrgamma(n, shape1, shape2, rate = 1, scale = 1/rate) minvtrgamma(order, shape1, shape2, rate = 1, scale = 1/rate) levinvtrgamma(limit, shape1, shape2, rate = 1, scale = 1/rate, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
shape1 , shape2 , scale
|
parameters. Must be strictly positive. |
rate |
an alternative way to specify the scale. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The inverse transformed gamma distribution with parameters
shape1
,
shape2
and
scale
, has density:
for ,
,
and
.
(Here
is the function implemented
by R's
gamma()
and defined in its help.)
The inverse transformed gamma is the distribution of the random
variable
where
has a gamma distribution with shape parameter
and scale parameter
or, equivalently, of the
random variable
with
a gamma distribution with shape parameter
and scale parameter
.
The inverse transformed gamma distribution defines a family of distributions with the following special cases:
An Inverse Gamma distribution when
shape2 == 1
;
An Inverse Weibull distribution when
shape1 == 1
;
An Inverse Exponential distribution when
shape1 == shape2 == 1
;
The th raw moment of the random variable
is
,
, and
the
th limited moment at some limit
is
for all
.
dinvtrgamma
gives the density,
pinvtrgamma
gives the distribution function,
qinvtrgamma
gives the quantile function,
rinvtrgamma
generates random deviates,
minvtrgamma
gives the th raw moment, and
levinvtrgamma
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
levinvtrgamma
computes the limited expected value using
gammainc
from package expint.
Distribution also known as the Inverse Generalized Gamma. See also Kleiber and Kotz (2003) for alternative names and parametrizations.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected] and Mathieu Pigeon
Kleiber, C. and Kotz, S. (2003), Statistical Size Distributions in Economics and Actuarial Sciences, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
exp(dinvtrgamma(2, 3, 4, 5, log = TRUE)) p <- (1:10)/10 pinvtrgamma(qinvtrgamma(p, 2, 3, 4), 2, 3, 4) minvtrgamma(2, 3, 4, 5) levinvtrgamma(200, 3, 4, 5, order = 2)
exp(dinvtrgamma(2, 3, 4, 5, log = TRUE)) p <- (1:10)/10 pinvtrgamma(qinvtrgamma(p, 2, 3, 4), 2, 3, 4) minvtrgamma(2, 3, 4, 5) levinvtrgamma(200, 3, 4, 5, order = 2)
Density function, distribution function, quantile function, random generation,
raw moments and limited moments for the Inverse Weibull distribution
with parameters shape
and scale
.
dinvweibull(x, shape, rate = 1, scale = 1/rate, log = FALSE) pinvweibull(q, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qinvweibull(p, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rinvweibull(n, shape, rate = 1, scale = 1/rate) minvweibull(order, shape, rate = 1, scale = 1/rate) levinvweibull(limit, shape, rate = 1, scale = 1/rate, order = 1)
dinvweibull(x, shape, rate = 1, scale = 1/rate, log = FALSE) pinvweibull(q, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qinvweibull(p, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rinvweibull(n, shape, rate = 1, scale = 1/rate) minvweibull(order, shape, rate = 1, scale = 1/rate) levinvweibull(limit, shape, rate = 1, scale = 1/rate, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
shape , scale
|
parameters. Must be strictly positive. |
rate |
an alternative way to specify the scale. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The inverse Weibull distribution with parameters shape
and
scale
has density:
for ,
and
.
The special case shape == 1
is an
Inverse Exponential distribution.
The th raw moment of the random variable
is
,
, and the
th
limited moment at some limit
is
, all
.
dinvweibull
gives the density,
pinvweibull
gives the distribution function,
qinvweibull
gives the quantile function,
rinvweibull
generates random deviates,
minvweibull
gives the th raw moment, and
levinvweibull
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
levinvweibull
computes the limited expected value using
gammainc
from package expint.
Distribution also knonw as the log-Gompertz. See also Kleiber and Kotz (2003) for alternative names and parametrizations.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected] and Mathieu Pigeon
Kleiber, C. and Kotz, S. (2003), Statistical Size Distributions in Economics and Actuarial Sciences, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
exp(dinvweibull(2, 3, 4, log = TRUE)) p <- (1:10)/10 pinvweibull(qinvweibull(p, 2, 3), 2, 3) mlgompertz(-1, 3, 3) levinvweibull(10, 2, 3, order = 1)
exp(dinvweibull(2, 3, 4, log = TRUE)) p <- (1:10)/10 pinvweibull(qinvweibull(p, 2, 3), 2, 3) mlgompertz(-1, 3, 3) levinvweibull(10, 2, 3, order = 1)
Density function, distribution function, quantile function and random
generation for the Logarithmic (or log-series) distribution with parameter
prob
.
dlogarithmic(x, prob, log = FALSE) plogarithmic(q, prob, lower.tail = TRUE, log.p = FALSE) qlogarithmic(p, prob, lower.tail = TRUE, log.p = FALSE) rlogarithmic(n, prob)
dlogarithmic(x, prob, log = FALSE) plogarithmic(q, prob, lower.tail = TRUE, log.p = FALSE) qlogarithmic(p, prob, lower.tail = TRUE, log.p = FALSE) rlogarithmic(n, prob)
x |
vector of (strictly positive integer) quantiles. |
q |
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
prob |
parameter. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
The logarithmic (or log-series) distribution with parameter
prob
has probability mass function
with and for
,
.
The logarithmic distribution is the limiting case of the
zero-truncated negative binomial distribution with size
parameter equal to . Note that in this context, parameter
prob
generally corresponds to the probability of failure
of the zero-truncated negative binomial.
If an element of x
is not integer, the result of
dlogarithmic
is zero, with a warning.
The quantile is defined as the smallest value such that
, where
is the distribution function.
dlogarithmic
gives the probability mass function,
plogarithmic
gives the distribution function,
qlogarithmic
gives the quantile function, and
rlogarithmic
generates random deviates.
Invalid prob
will result in return value NaN
, with a warning.
The length of the result is determined by n
for
rlogarithmic
, and is the maximum of the lengths of the
numerical arguments for the other functions.
qlogarithmic
is based on qbinom
et al.; it uses the
Cornish–Fisher Expansion to include a skewness correction to a normal
approximation, followed by a search.
rlogarithmic
is an implementation of the LS and LK algorithms
of Kemp (1981) with automatic selection. As suggested by Devroye
(1986), the LS algorithm is used when prob < 0.95
, and the LK
algorithm otherwise.
Vincent Goulet [email protected]
Johnson, N. L., Kemp, A. W. and Kotz, S. (2005), Univariate Discrete Distributions, Third Edition, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
Kemp, A. W. (1981), “Efficient Generation of Logarithmically Distributed Pseudo-Random Variables”, Journal of the Royal Statistical Society, Series C, vol. 30, p. 249-253.
Devroye, L. (1986), Non-Uniform Random Variate Generation, Springer-Verlag. http://luc.devroye.org/rnbookindex.html
dztnbinom
for the zero-truncated negative binomial
distribution.
## Table 1 of Kemp (1981) [also found in Johnson et al. (2005), chapter 7] p <- c(0.1, 0.3, 0.5, 0.7, 0.8, 0.85, 0.9, 0.95, 0.99, 0.995, 0.999, 0.9999) round(rbind(dlogarithmic(1, p), dlogarithmic(2, p), plogarithmic(9, p, lower.tail = FALSE), -p/((1 - p) * log(1 - p))), 2) qlogarithmic(plogarithmic(1:10, 0.9), 0.9) x <- rlogarithmic(1000, 0.8) y <- sort(unique(x)) plot(y, table(x)/length(x), type = "h", lwd = 2, pch = 19, col = "black", xlab = "x", ylab = "p(x)", main = "Empirical vs theoretical probabilities") points(y, dlogarithmic(y, prob = 0.8), pch = 19, col = "red") legend("topright", c("empirical", "theoretical"), lty = c(1, NA), pch = c(NA, 19), col = c("black", "red"))
## Table 1 of Kemp (1981) [also found in Johnson et al. (2005), chapter 7] p <- c(0.1, 0.3, 0.5, 0.7, 0.8, 0.85, 0.9, 0.95, 0.99, 0.995, 0.999, 0.9999) round(rbind(dlogarithmic(1, p), dlogarithmic(2, p), plogarithmic(9, p, lower.tail = FALSE), -p/((1 - p) * log(1 - p))), 2) qlogarithmic(plogarithmic(1:10, 0.9), 0.9) x <- rlogarithmic(1000, 0.8) y <- sort(unique(x)) plot(y, table(x)/length(x), type = "h", lwd = 2, pch = 19, col = "black", xlab = "x", ylab = "p(x)", main = "Empirical vs theoretical probabilities") points(y, dlogarithmic(y, prob = 0.8), pch = 19, col = "red") legend("topright", c("empirical", "theoretical"), lty = c(1, NA), pch = c(NA, 19), col = c("black", "red"))
Density function, distribution function, quantile function, random generation,
raw moments and limited moments for the Loggamma distribution with
parameters shapelog
and ratelog
.
dlgamma(x, shapelog, ratelog, log = FALSE) plgamma(q, shapelog, ratelog, lower.tail = TRUE, log.p = FALSE) qlgamma(p, shapelog, ratelog, lower.tail = TRUE, log.p = FALSE) rlgamma(n, shapelog, ratelog) mlgamma(order, shapelog, ratelog) levlgamma(limit, shapelog, ratelog, order = 1)
dlgamma(x, shapelog, ratelog, log = FALSE) plgamma(q, shapelog, ratelog, lower.tail = TRUE, log.p = FALSE) qlgamma(p, shapelog, ratelog, lower.tail = TRUE, log.p = FALSE) rlgamma(n, shapelog, ratelog) mlgamma(order, shapelog, ratelog) levlgamma(limit, shapelog, ratelog, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
shapelog , ratelog
|
parameters. Must be strictly positive. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The loggamma distribution with parameters shapelog
and
ratelog
has density:
for ,
and
.
(Here
is the function implemented
by R's
gamma()
and defined in its help.)
The loggamma is the distribution of the random variable
, where
has a gamma distribution with
shape parameter
and scale parameter
.
The th raw moment of the random variable
is
and the
th limited moment at some limit
is
,
.
dlgamma
gives the density,
plgamma
gives the distribution function,
qlgamma
gives the quantile function,
rlgamma
generates random deviates,
mlgamma
gives the th raw moment, and
levlgamma
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected] and Mathieu Pigeon
Hogg, R. V. and Klugman, S. A. (1984), Loss Distributions, Wiley.
exp(dlgamma(2, 3, 4, log = TRUE)) p <- (1:10)/10 plgamma(qlgamma(p, 2, 3), 2, 3) mlgamma(2, 3, 4) - mlgamma(1, 3, 4)^2 levlgamma(10, 3, 4, order = 2)
exp(dlgamma(2, 3, 4, log = TRUE)) p <- (1:10)/10 plgamma(qlgamma(p, 2, 3), 2, 3) mlgamma(2, 3, 4) - mlgamma(1, 3, 4)^2 levlgamma(10, 3, 4, order = 2)
Density function, distribution function, quantile function, random generation,
raw moments and limited moments for the Loglogistic distribution with
parameters shape
and scale
.
dllogis(x, shape, rate = 1, scale = 1/rate, log = FALSE) pllogis(q, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qllogis(p, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rllogis(n, shape, rate = 1, scale = 1/rate) mllogis(order, shape, rate = 1, scale = 1/rate) levllogis(limit, shape, rate = 1, scale = 1/rate, order = 1)
dllogis(x, shape, rate = 1, scale = 1/rate, log = FALSE) pllogis(q, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qllogis(p, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rllogis(n, shape, rate = 1, scale = 1/rate) mllogis(order, shape, rate = 1, scale = 1/rate) levllogis(limit, shape, rate = 1, scale = 1/rate, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
shape , scale
|
parameters. Must be strictly positive. |
rate |
an alternative way to specify the scale. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The loglogistic distribution with parameters shape
and
scale
has density:
for ,
and
.
The th raw moment of the random variable
is
,
.
The th limited moment at some limit
is
,
and
not a negative integer.
dllogis
gives the density,
pllogis
gives the distribution function,
qllogis
gives the quantile function,
rllogis
generates random deviates,
mllogis
gives the th raw moment, and
levllogis
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
levllogis
computes the limited expected value using
betaint
.
Also known as the Fisk distribution. See also Kleiber and Kotz (2003) for alternative names and parametrizations.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected] and Mathieu Pigeon
Kleiber, C. and Kotz, S. (2003), Statistical Size Distributions in Economics and Actuarial Sciences, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dpareto3
for an equivalent distribution with a location
parameter.
exp(dllogis(2, 3, 4, log = TRUE)) p <- (1:10)/10 pllogis(qllogis(p, 2, 3), 2, 3) ## mean mllogis(1, 2, 3) ## case with 1 - order/shape > 0 levllogis(10, 2, 3, order = 1) ## case with 1 - order/shape < 0 levllogis(10, 2/3, 3, order = 1)
exp(dllogis(2, 3, 4, log = TRUE)) p <- (1:10)/10 pllogis(qllogis(p, 2, 3), 2, 3) ## mean mllogis(1, 2, 3) ## case with 1 - order/shape > 0 levllogis(10, 2, 3, order = 1) ## case with 1 - order/shape < 0 levllogis(10, 2/3, 3, order = 1)
Raw moments and limited moments for the Lognormal distribution whose
logarithm has mean equal to meanlog
and standard deviation
equal to sdlog
.
mlnorm(order, meanlog = 0, sdlog = 1) levlnorm(limit, meanlog = 0, sdlog = 1, order = 1)
mlnorm(order, meanlog = 0, sdlog = 1) levlnorm(limit, meanlog = 0, sdlog = 1, order = 1)
order |
order of the moment. |
limit |
limit of the loss variable. |
meanlog , sdlog
|
mean and standard deviation of the distribution
on the log scale with default values of |
mlnorm
gives the th raw moment and
levlnorm
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
Vincent Goulet [email protected] and Mathieu Pigeon
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
Lognormal
for details on the lognormal distribution and
functions [dpqr]lnorm
.
mlnorm(2, 3, 4) - mlnorm(1, 3, 4)^2 levlnorm(10, 3, 4, order = 2)
mlnorm(2, 3, 4) - mlnorm(1, 3, 4)^2 levlnorm(10, 3, 4, order = 2)
Minimum distance fitting of univariate distributions, allowing parameters to be held fixed if desired.
mde(x, fun, start, measure = c("CvM", "chi-square", "LAS"), weights = NULL, ...)
mde(x, fun, start, measure = c("CvM", "chi-square", "LAS"), weights = NULL, ...)
x |
a vector or an object of class |
fun |
function returning a cumulative distribution (for
|
start |
a named list giving the parameters to be optimized with initial values |
measure |
either |
weights |
weights; see Details. |
... |
Additional parameters, either for |
The Cramer-von Mises method ("CvM"
) minimizes the squared
difference between the theoretical cdf and the empirical cdf at the
data points (for individual data) or the ogive at the knots (for
grouped data).
The modified chi-square method ("chi-square"
) minimizes the
modified chi-square statistic for grouped data, that is the squared
difference between the expected and observed frequency within each
group.
The layer average severity method ("LAS"
) minimizes the
squared difference between the theoretical and empirical limited
expected value within each group for grouped data.
All sum of squares can be weighted. If arguments weights
is
missing, weights default to 1 for measure = "CvM"
and
measure = "LAS"
; for measure = "chi-square"
, weights
default to , where
is the frequency
in group
.
Optimization is performed using optim
. For
one-dimensional problems the Nelder-Mead method is used and for
multi-dimensional problems the BFGS method, unless arguments named
lower
or upper
are supplied when L-BFGS-B
is used
or method
is supplied explicitly.
An object of class "mde"
, a list with two components:
estimate |
the parameter estimates, and |
distance |
the distance. |
Vincent Goulet [email protected] and Mathieu Pigeon
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (1998), Loss Models, From Data to Decisions, Wiley.
## Individual data example data(dental) mde(dental, pexp, start = list(rate = 1/200), measure = "CvM") ## Example 2.21 of Klugman et al. (1998) data(gdental) mde(gdental, pexp, start = list(rate = 1/200), measure = "CvM") mde(gdental, pexp, start = list(rate = 1/200), measure = "chi-square") mde(gdental, levexp, start = list(rate = 1/200), measure = "LAS") ## Two-parameter distribution example try(mde(gdental, ppareto, start = list(shape = 3, scale = 600), measure = "CvM")) # no convergence ## Working in log scale often solves the problem pparetolog <- function(x, shape, scale) ppareto(x, exp(shape), exp(scale)) ( p <- mde(gdental, pparetolog, start = list(shape = log(3), scale = log(600)), measure = "CvM") ) exp(p$estimate)
## Individual data example data(dental) mde(dental, pexp, start = list(rate = 1/200), measure = "CvM") ## Example 2.21 of Klugman et al. (1998) data(gdental) mde(gdental, pexp, start = list(rate = 1/200), measure = "CvM") mde(gdental, pexp, start = list(rate = 1/200), measure = "chi-square") mde(gdental, levexp, start = list(rate = 1/200), measure = "LAS") ## Two-parameter distribution example try(mde(gdental, ppareto, start = list(shape = 3, scale = 600), measure = "CvM")) # no convergence ## Working in log scale often solves the problem pparetolog <- function(x, shape, scale) ppareto(x, exp(shape), exp(scale)) ( p <- mde(gdental, pparetolog, start = list(shape = log(3), scale = log(600)), measure = "CvM") ) exp(p$estimate)
Mean of grouped data objects.
## S3 method for class 'grouped.data' mean(x, ...)
## S3 method for class 'grouped.data' mean(x, ...)
x |
an object of class |
... |
further arguments passed to or from other methods. |
The mean of grouped data with group boundaries and group frequencies
is
where
is the midpoint of the
th interval, and
.
A named vector of means.
Vincent Goulet [email protected]
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (1998), Loss Models, From Data to Decisions, Wiley.
grouped.data
to create grouped data objects;
emm
to compute higher moments.
data(gdental) mean(gdental)
data(gdental) mean(gdental)
Raw moments and moment generating function for the normal distribution with
mean equal to mean
and standard deviation equal to sd
.
mnorm(order, mean = 0, sd = 1) mgfnorm(t, mean = 0, sd = 1, log = FALSE)
mnorm(order, mean = 0, sd = 1) mgfnorm(t, mean = 0, sd = 1, log = FALSE)
order |
vector of integers; order of the moment. |
mean |
vector of means. |
sd |
vector of standard deviations. |
t |
numeric vector. |
log |
logical; if |
The th raw moment of the random variable
is
and the moment generating function is
.
Only integer moments are supported.
mnorm
gives the th raw moment and
mgfnorm
gives the moment generating function in t
.
Invalid arguments will result in return value NaN
, with a warning.
Vincent Goulet [email protected], Christophe Dutang
Johnson, N. L. and Kotz, S. (1970), Continuous Univariate Distributions, Volume 1, Wiley.
mgfnorm(0:4,1,2) mnorm(3)
mgfnorm(0:4,1,2) mnorm(3)
Compute a smoothed empirical distribution function for grouped data.
ogive(x, ...) ## Default S3 method: ogive(x, y = NULL, breaks = "Sturges", nclass = NULL, ...) ## S3 method for class 'grouped.data' ogive(x, ...) ## S3 method for class 'ogive' print(x, digits = getOption("digits") - 2, ...) ## S3 method for class 'ogive' summary(object, ...) ## S3 method for class 'ogive' knots(Fn, ...) ## S3 method for class 'ogive' plot(x, main = NULL, xlab = "x", ylab = "F(x)", ...)
ogive(x, ...) ## Default S3 method: ogive(x, y = NULL, breaks = "Sturges", nclass = NULL, ...) ## S3 method for class 'grouped.data' ogive(x, ...) ## S3 method for class 'ogive' print(x, digits = getOption("digits") - 2, ...) ## S3 method for class 'ogive' summary(object, ...) ## S3 method for class 'ogive' knots(Fn, ...) ## S3 method for class 'ogive' plot(x, main = NULL, xlab = "x", ylab = "F(x)", ...)
x |
for the generic and all but the default method, an object of
class |
y |
a vector of group frequencies. |
breaks , nclass
|
arguments passed to |
digits |
number of significant digits to use, see
|
Fn , object
|
an R object inheriting from |
main |
main title. |
xlab , ylab
|
labels of x and y axis. |
... |
arguments to be passed to subsequent methods. |
The ogive is a linear interpolation of the empirical cumulative distribution function.
The equation of the ogive is
for and where
are the
group
boundaries and
is the empirical distribution function of
the sample.
For ogive
, a function of class "ogive"
, inheriting from the
"function"
class.
Vincent Goulet [email protected] and Mathieu Pigeon
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (1998), Loss Models, From Data to Decisions, Wiley.
grouped.data
to create grouped data objects;
quantile.grouped.data
for the inverse function;
approxfun
, which is used to compute the ogive;
stepfun
for related documentation (even though the ogive
is not a step function).
## Most common usage: create ogive from grouped data object. Fn <- ogive(gdental) Fn summary(Fn) knots(Fn) # the group boundaries Fn(knots(Fn)) # true values of the empirical cdf Fn(c(80, 200, 2000)) # linear interpolations plot(Fn) # graphical representation ## Alternative 1: create ogive directly from individual data ## without first creating a grouped data object. ogive(dental) # automatic class boundaries ogive(dental, breaks = c(0, 50, 200, 500, 1500, 2000)) ## Alternative 2: create ogive from set of group boundaries and ## group frequencies. cj <- c(0, 25, 50, 100, 250, 500, 1000) nj <- c(30, 31, 57, 42, 45, 10) ogive(cj, nj)
## Most common usage: create ogive from grouped data object. Fn <- ogive(gdental) Fn summary(Fn) knots(Fn) # the group boundaries Fn(knots(Fn)) # true values of the empirical cdf Fn(c(80, 200, 2000)) # linear interpolations plot(Fn) # graphical representation ## Alternative 1: create ogive directly from individual data ## without first creating a grouped data object. ogive(dental) # automatic class boundaries ogive(dental, breaks = c(0, 50, 200, 500, 1500, 2000)) ## Alternative 2: create ogive from set of group boundaries and ## group frequencies. cj <- c(0, 25, 50, 100, 250, 500, 1000) nj <- c(30, 31, 57, 42, 45, 10) ogive(cj, nj)
Density function, distribution function, quantile function, random generation,
raw moments and limited moments for the Paralogistic distribution with
parameters shape
and scale
.
dparalogis(x, shape, rate = 1, scale = 1/rate, log = FALSE) pparalogis(q, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qparalogis(p, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rparalogis(n, shape, rate = 1, scale = 1/rate) mparalogis(order, shape, rate = 1, scale = 1/rate) levparalogis(limit, shape, rate = 1, scale = 1/rate, order = 1)
dparalogis(x, shape, rate = 1, scale = 1/rate, log = FALSE) pparalogis(q, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qparalogis(p, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rparalogis(n, shape, rate = 1, scale = 1/rate) mparalogis(order, shape, rate = 1, scale = 1/rate) levparalogis(limit, shape, rate = 1, scale = 1/rate, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
shape , scale
|
parameters. Must be strictly positive. |
rate |
an alternative way to specify the scale. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The paralogistic distribution with parameters shape
and
scale
has density:
for ,
and
.
The th raw moment of the random variable
is
,
.
The th limited moment at some limit
is
,
and
not a negative integer.
dparalogis
gives the density,
pparalogis
gives the distribution function,
qparalogis
gives the quantile function,
rparalogis
generates random deviates,
mparalogis
gives the th raw moment, and
levparalogis
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
levparalogis
computes the limited expected value using
betaint
.
See Kleiber and Kotz (2003) for alternative names and parametrizations.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected] and Mathieu Pigeon
Kleiber, C. and Kotz, S. (2003), Statistical Size Distributions in Economics and Actuarial Sciences, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
exp(dparalogis(2, 3, 4, log = TRUE)) p <- (1:10)/10 pparalogis(qparalogis(p, 2, 3), 2, 3) ## variance mparalogis(2, 2, 3) - mparalogis(1, 2, 3)^2 ## case with shape - order/shape > 0 levparalogis(10, 2, 3, order = 2) ## case with shape - order/shape < 0 levparalogis(10, 1.25, 3, order = 2)
exp(dparalogis(2, 3, 4, log = TRUE)) p <- (1:10)/10 pparalogis(qparalogis(p, 2, 3), 2, 3) ## variance mparalogis(2, 2, 3) - mparalogis(1, 2, 3)^2 ## case with shape - order/shape > 0 levparalogis(10, 2, 3, order = 2) ## case with shape - order/shape < 0 levparalogis(10, 1.25, 3, order = 2)
Density function, distribution function, quantile function, random generation,
raw moments and limited moments for the Pareto distribution with
parameters shape
and scale
.
dpareto(x, shape, scale, log = FALSE) ppareto(q, shape, scale, lower.tail = TRUE, log.p = FALSE) qpareto(p, shape, scale, lower.tail = TRUE, log.p = FALSE) rpareto(n, shape, scale) mpareto(order, shape, scale) levpareto(limit, shape, scale, order = 1)
dpareto(x, shape, scale, log = FALSE) ppareto(q, shape, scale, lower.tail = TRUE, log.p = FALSE) qpareto(p, shape, scale, lower.tail = TRUE, log.p = FALSE) rpareto(n, shape, scale) mpareto(order, shape, scale) levpareto(limit, shape, scale, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
shape , scale
|
parameters. Must be strictly positive. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The Pareto distribution with parameters shape
and
scale
has density:
for ,
and
.
There are many different definitions of the Pareto distribution in the literature; see Arnold (2015) or Kleiber and Kotz (2003). In the nomenclature of actuar, The “Pareto distribution” does not have a location parameter. The version with a location parameter is the Pareto II.
The th raw moment of the random variable
is
,
.
The th limited moment at some limit
is
,
and
not a
negative integer.
dpareto
gives the density,
ppareto
gives the distribution function,
qpareto
gives the quantile function,
rpareto
generates random deviates,
mpareto
gives the th raw moment, and
levpareto
gives the th moment of the limited loss variable.
Invalid arguments will result in return value NaN
, with a warning.
levpareto
computes the limited expected value using
betaint
.
The version of the Pareto defined for is named
Single Parameter Pareto, or Pareto I, in actuar.
Vincent Goulet [email protected] and Mathieu Pigeon
Kleiber, C. and Kotz, S. (2003), Statistical Size Distributions in Economics and Actuarial Sciences, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dpareto2
for an equivalent distribution with location
parameter.
dpareto1
for the Single Parameter Pareto distribution.
"distributions"
package vignette for details on the
interrelations between the continuous size distributions in
actuar and complete formulas underlying the above functions.
exp(dpareto(2, 3, 4, log = TRUE)) p <- (1:10)/10 ppareto(qpareto(p, 2, 3), 2, 3) ## variance mpareto(2, 4, 1) - mpareto(1, 4, 1)^2 ## case with shape - order > 0 levpareto(10, 3, scale = 1, order = 2) ## case with shape - order < 0 levpareto(10, 1.5, scale = 1, order = 2)
exp(dpareto(2, 3, 4, log = TRUE)) p <- (1:10)/10 ppareto(qpareto(p, 2, 3), 2, 3) ## variance mpareto(2, 4, 1) - mpareto(1, 4, 1)^2 ## case with shape - order > 0 levpareto(10, 3, scale = 1, order = 2) ## case with shape - order < 0 levpareto(10, 1.5, scale = 1, order = 2)
Density function, distribution function, quantile function, random generation,
raw moments and limited moments for the Pareto II distribution with
parameters min
, shape
and scale
.
dpareto2(x, min, shape, rate = 1, scale = 1/rate, log = FALSE) ppareto2(q, min, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qpareto2(p, min, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rpareto2(n, min, shape, rate = 1, scale = 1/rate) mpareto2(order, min, shape, rate = 1, scale = 1/rate) levpareto2(limit, min, shape, rate = 1, scale = 1/rate, order = 1)
dpareto2(x, min, shape, rate = 1, scale = 1/rate, log = FALSE) ppareto2(q, min, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qpareto2(p, min, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rpareto2(n, min, shape, rate = 1, scale = 1/rate) mpareto2(order, min, shape, rate = 1, scale = 1/rate) levpareto2(limit, min, shape, rate = 1, scale = 1/rate, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
min |
lower bound of the support of the distribution. |
shape , scale
|
parameters. Must be strictly positive. |
rate |
an alternative way to specify the scale. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The Pareto II (or “type II”) distribution with parameters
min
,
shape
and
scale
has density:
for ,
,
and
.
The Pareto II is the distribution of the random variable
where has a beta distribution with parameters
and
. It derives from the Feller-Pareto
distribution with
.
Setting
yields the familiar
Pareto distribution.
The Pareto I (or Single parameter Pareto)
distribution is a special case of the Pareto II with min ==
scale
.
The th raw moment of the random variable
is
for nonnegative integer values of
.
The th limited moment at some limit
is
for nonnegative integer values of
and
,
not a negative integer.
dpareto2
gives the density,
ppareto2
gives the distribution function,
qpareto2
gives the quantile function,
rpareto2
generates random deviates,
mpareto2
gives the th raw moment, and
levpareto2
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
levpareto2
computes the limited expected value using
betaint
.
For Pareto distributions, we use the classification of Arnold (2015) with the parametrization of Klugman et al. (2012).
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected]
Arnold, B.C. (2015), Pareto Distributions, Second Edition, CRC Press.
Kleiber, C. and Kotz, S. (2003), Statistical Size Distributions in Economics and Actuarial Sciences, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dpareto
for the Pareto distribution without a location
parameter.
exp(dpareto2(1, min = 10, 3, 4, log = TRUE)) p <- (1:10)/10 ppareto2(qpareto2(p, min = 10, 2, 3), min = 10, 2, 3) ## variance mpareto2(2, min = 10, 4, 1) - mpareto2(1, min = 10, 4, 1)^2 ## case with shape - order > 0 levpareto2(10, min = 10, 3, scale = 1, order = 2) ## case with shape - order < 0 levpareto2(10, min = 10, 1.5, scale = 1, order = 2)
exp(dpareto2(1, min = 10, 3, 4, log = TRUE)) p <- (1:10)/10 ppareto2(qpareto2(p, min = 10, 2, 3), min = 10, 2, 3) ## variance mpareto2(2, min = 10, 4, 1) - mpareto2(1, min = 10, 4, 1)^2 ## case with shape - order > 0 levpareto2(10, min = 10, 3, scale = 1, order = 2) ## case with shape - order < 0 levpareto2(10, min = 10, 1.5, scale = 1, order = 2)
Density function, distribution function, quantile function, random generation,
raw moments and limited moments for the Pareto III distribution with
parameters min
, shape
and scale
.
dpareto3(x, min, shape, rate = 1, scale = 1/rate, log = FALSE) ppareto3(q, min, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qpareto3(p, min, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rpareto3(n, min, shape, rate = 1, scale = 1/rate) mpareto3(order, min, shape, rate = 1, scale = 1/rate) levpareto3(limit, min, shape, rate = 1, scale = 1/rate, order = 1)
dpareto3(x, min, shape, rate = 1, scale = 1/rate, log = FALSE) ppareto3(q, min, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qpareto3(p, min, shape, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rpareto3(n, min, shape, rate = 1, scale = 1/rate) mpareto3(order, min, shape, rate = 1, scale = 1/rate) levpareto3(limit, min, shape, rate = 1, scale = 1/rate, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
min |
lower bound of the support of the distribution. |
shape , scale
|
parameters. Must be strictly positive. |
rate |
an alternative way to specify the scale. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The Pareto III (or “type III”) distribution with parameters
min
,
shape
and
scale
has density:
for ,
,
and
.
The Pareto III is the distribution of the random variable
where has a uniform distribution on
. It derives
from the Feller-Pareto
distribution with
.
Setting
yields the loglogistic
distribution.
The th raw moment of the random variable
is
for nonnegative integer values of
.
The th limited moment at some limit
is
for nonnegative integer values of
and
,
not a negative integer.
dpareto3
gives the density,
ppareto3
gives the distribution function,
qpareto3
gives the quantile function,
rpareto3
generates random deviates,
mpareto3
gives the th raw moment, and
levpareto3
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
levpareto3
computes the limited expected value using
betaint
.
For Pareto distributions, we use the classification of Arnold (2015) with the parametrization of Klugman et al. (2012).
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected]
Arnold, B.C. (2015), Pareto Distributions, Second Edition, CRC Press.
Kleiber, C. and Kotz, S. (2003), Statistical Size Distributions in Economics and Actuarial Sciences, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dllogis
for the loglogistic distribution.
exp(dpareto3(1, min = 10, 3, 4, log = TRUE)) p <- (1:10)/10 ppareto3(qpareto3(p, min = 10, 2, 3), min = 10, 2, 3) ## mean mpareto3(1, min = 10, 2, 3) ## case with 1 - order/shape > 0 levpareto3(20, min = 10, 2, 3, order = 1) ## case with 1 - order/shape < 0 levpareto3(20, min = 10, 2/3, 3, order = 1)
exp(dpareto3(1, min = 10, 3, 4, log = TRUE)) p <- (1:10)/10 ppareto3(qpareto3(p, min = 10, 2, 3), min = 10, 2, 3) ## mean mpareto3(1, min = 10, 2, 3) ## case with 1 - order/shape > 0 levpareto3(20, min = 10, 2, 3, order = 1) ## case with 1 - order/shape < 0 levpareto3(20, min = 10, 2/3, 3, order = 1)
Density function, distribution function, quantile function, random generation,
raw moments and limited moments for the Pareto IV distribution with
parameters min
, shape1
, shape2
and scale
.
dpareto4(x, min, shape1, shape2, rate = 1, scale = 1/rate, log = FALSE) ppareto4(q, min, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qpareto4(p, min, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rpareto4(n, min, shape1, shape2, rate = 1, scale = 1/rate) mpareto4(order, min, shape1, shape2, rate = 1, scale = 1/rate) levpareto4(limit, min, shape1, shape2, rate = 1, scale = 1/rate, order = 1)
dpareto4(x, min, shape1, shape2, rate = 1, scale = 1/rate, log = FALSE) ppareto4(q, min, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qpareto4(p, min, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rpareto4(n, min, shape1, shape2, rate = 1, scale = 1/rate) mpareto4(order, min, shape1, shape2, rate = 1, scale = 1/rate) levpareto4(limit, min, shape1, shape2, rate = 1, scale = 1/rate, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
min |
lower bound of the support of the distribution. |
shape1 , shape2 , scale
|
parameters. Must be strictly positive. |
rate |
an alternative way to specify the scale. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The Pareto IV (or “type IV”) distribution with parameters
min
,
shape1
,
shape2
and
scale
has density:
for ,
,
,
and
.
The Pareto IV is the distribution of the random variable
where has a beta distribution with parameters
and
. It derives from the Feller-Pareto
distribution with
. Setting
yields the Burr distribution.
The Pareto IV distribution also has the following direct special cases:
A Pareto III distribution when shape1
== 1
;
A Pareto II distribution when shape1
== 1
.
The th raw moment of the random variable
is
for nonnegative integer values of
.
The th limited moment at some limit
is
for nonnegative integer values of
and
,
not a negative integer.
dpareto4
gives the density,
ppareto4
gives the distribution function,
qpareto4
gives the quantile function,
rpareto4
generates random deviates,
mpareto4
gives the th raw moment, and
levpareto4
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
levpareto4
computes the limited expected value using
betaint
.
For Pareto distributions, we use the classification of Arnold (2015) with the parametrization of Klugman et al. (2012).
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected]
Arnold, B.C. (2015), Pareto Distributions, Second Edition, CRC Press.
Kleiber, C. and Kotz, S. (2003), Statistical Size Distributions in Economics and Actuarial Sciences, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dburr
for the Burr distribution.
exp(dpareto4(1, min = 10, 2, 3, log = TRUE)) p <- (1:10)/10 ppareto4(qpareto4(p, min = 10, 2, 3, 2), min = 10, 2, 3, 2) ## variance mpareto4(2, min = 10, 2, 3, 1) - mpareto4(1, min = 10, 2, 3, 1) ^ 2 ## case with shape1 - order/shape2 > 0 levpareto4(10, min = 10, 2, 3, 1, order = 2) ## case with shape1 - order/shape2 < 0 levpareto4(10, min = 10, 1.5, 0.5, 1, order = 2)
exp(dpareto4(1, min = 10, 2, 3, log = TRUE)) p <- (1:10)/10 ppareto4(qpareto4(p, min = 10, 2, 3, 2), min = 10, 2, 3, 2) ## variance mpareto4(2, min = 10, 2, 3, 1) - mpareto4(1, min = 10, 2, 3, 1) ^ 2 ## case with shape1 - order/shape2 > 0 levpareto4(10, min = 10, 2, 3, 1, order = 2) ## case with shape1 - order/shape2 < 0 levpareto4(10, min = 10, 1.5, 0.5, 1, order = 2)
Density, distribution function, random generation, raw moments and
moment generating function for the (continuous) Phase-type
distribution with parameters prob
and rates
.
dphtype(x, prob, rates, log = FALSE) pphtype(q, prob, rates, lower.tail = TRUE, log.p = FALSE) rphtype(n, prob, rates) mphtype(order, prob, rates) mgfphtype(t, prob, rates, log = FALSE)
dphtype(x, prob, rates, log = FALSE) pphtype(q, prob, rates, lower.tail = TRUE, log.p = FALSE) rphtype(n, prob, rates) mphtype(order, prob, rates) mgfphtype(t, prob, rates, log = FALSE)
x , q
|
vector of quantiles. |
n |
number of observations. If |
prob |
vector of initial probabilities for each of the transient
states of the underlying Markov chain. The initial probability of
the absorbing state is |
rates |
square matrix of the rates of transition among the states of the underlying Markov chain. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
t |
numeric vector. |
The phase-type distribution with parameters prob
and
rates
has density:
for and
, where
is a column vector with all components equal to one,
is the exit rates vector and
denotes the matrix exponential of
. The
matrix exponential of a matrix
is defined as
the Taylor series
The parameters of the distribution must satisfy
,
,
and
.
The th raw moment of the random variable
is
and the moment generating function is
.
dphasetype
gives the density,
pphasetype
gives the distribution function,
rphasetype
generates random deviates,
mphasetype
gives the th raw moment, and
mgfphasetype
gives the moment generating function in x
.
Invalid arguments will result in return value NaN
, with a warning.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected] and Christophe Dutang
https://en.wikipedia.org/wiki/Phase-type_distribution
Neuts, M. F. (1981), Generating random variates from a distribution of phase type, WSC '81: Proceedings of the 13th conference on Winter simulation, IEEE Press.
## Erlang(3, 2) distribution T <- cbind(c(-2, 0, 0), c(2, -2, 0), c(0, 2, -2)) pi <- c(1,0,0) x <- 0:10 dphtype(x, pi, T) # density dgamma(x, 3, 2) # same pphtype(x, pi, T) # cdf pgamma(x, 3, 2) # same rphtype(10, pi, T) # random values mphtype(1, pi, T) # expected value curve(mgfphtype(x, pi, T), from = -10, to = 1)
## Erlang(3, 2) distribution T <- cbind(c(-2, 0, 0), c(2, -2, 0), c(0, 2, -2)) pi <- c(1,0,0) x <- 0:10 dphtype(x, pi, T) # density dgamma(x, 3, 2) # same pphtype(x, pi, T) # cdf pgamma(x, 3, 2) # same rphtype(10, pi, T) # random values mphtype(1, pi, T) # expected value curve(mgfphtype(x, pi, T), from = -10, to = 1)
Density function, distribution function, quantile function and random
generation for the Poisson-inverse Gaussian discrete distribution with
parameters mean
and shape
.
dpoisinvgauss(x, mean, shape = 1, dispersion = 1/shape, log = FALSE) ppoisinvgauss(q, mean, shape = 1, dispersion = 1/shape, lower.tail = TRUE, log.p = FALSE) qpoisinvgauss(p, mean, shape = 1, dispersion = 1/shape, lower.tail = TRUE, log.p = FALSE) rpoisinvgauss(n, mean, shape = 1, dispersion = 1/shape)
dpoisinvgauss(x, mean, shape = 1, dispersion = 1/shape, log = FALSE) ppoisinvgauss(q, mean, shape = 1, dispersion = 1/shape, lower.tail = TRUE, log.p = FALSE) qpoisinvgauss(p, mean, shape = 1, dispersion = 1/shape, lower.tail = TRUE, log.p = FALSE) rpoisinvgauss(n, mean, shape = 1, dispersion = 1/shape)
x |
vector of (positive integer) quantiles. |
q |
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
mean , shape
|
parameters. Must be strictly positive. Infinite values are supported. |
dispersion |
an alternative way to specify the shape. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
The Poisson-inverse Gaussian distribution is the result of the continuous mixture between a Poisson distribution and an inverse Gaussian, that is, the distribution with probability mass function
where is the density
function of the inverse Gaussian distribution with parameters
mean
and
dispersion
(see
dinvgauss
).
The resulting probability mass function is
for ,
,
and where
is the modified Bessel function of the third
kind implemented by R's
besselK()
and defined in its
help.
The limiting case has well defined
probability mass and distribution functions, but has no finite
strictly positive, integer moments. The pmf in this case reduces to
The limiting case is a degenerate distribution in
.
If an element of x
is not integer, the result of
dpoisinvgauss
is zero, with a warning.
The quantile is defined as the smallest value such that
, where
is the distribution function.
dpoisinvgauss
gives the probability mass function,
ppoisinvgauss
gives the distribution function,
qpoisinvgauss
gives the quantile function, and
rpoisinvgauss
generates random deviates.
Invalid arguments will result in return value NaN
, with a warning.
The length of the result is determined by n
for
rpoisinvgauss
, and is the maximum of the lengths of the
numerical arguments for the other functions.
[dpqr]pig
are aliases for [dpqr]poisinvgauss
.
qpoisinvgauss
is based on qbinom
et al.; it uses the
Cornish–Fisher Expansion to include a skewness correction to a normal
approximation, followed by a search.
Vincent Goulet [email protected]
Holla, M. S. (1966), “On a Poisson-Inverse Gaussian Distribution”, Metrika, vol. 15, p. 377-384.
Johnson, N. L., Kemp, A. W. and Kotz, S. (2005), Univariate Discrete Distributions, Third Edition, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
Shaban, S. A., (1981) “Computation of the poisson-inverse gaussian distribution”, Communications in Statistics - Theory and Methods, vol. 10, no. 14, p. 1389-1399.
dpois
for the Poisson distribution,
dinvgauss
for the inverse Gaussian distribution.
## Tables I and II of Shaban (1981) x <- 0:2 sapply(c(0.4, 0.8, 1), dpoisinvgauss, x = x, mean = 0.1) sapply(c(40, 80, 100, 130), dpoisinvgauss, x = x, mean = 1) qpoisinvgauss(ppoisinvgauss(0:10, 1, dis = 2.5), 1, dis = 2.5) x <- rpoisinvgauss(1000, 1, dis = 2.5) y <- sort(unique(x)) plot(y, table(x)/length(x), type = "h", lwd = 2, pch = 19, col = "black", xlab = "x", ylab = "p(x)", main = "Empirical vs theoretical probabilities") points(y, dpoisinvgauss(y, 1, dis = 2.5), pch = 19, col = "red") legend("topright", c("empirical", "theoretical"), lty = c(1, NA), pch = c(NA, 19), col = c("black", "red"))
## Tables I and II of Shaban (1981) x <- 0:2 sapply(c(0.4, 0.8, 1), dpoisinvgauss, x = x, mean = 0.1) sapply(c(40, 80, 100, 130), dpoisinvgauss, x = x, mean = 1) qpoisinvgauss(ppoisinvgauss(0:10, 1, dis = 2.5), 1, dis = 2.5) x <- rpoisinvgauss(1000, 1, dis = 2.5) y <- sort(unique(x)) plot(y, table(x)/length(x), type = "h", lwd = 2, pch = 19, col = "black", xlab = "x", ylab = "p(x)", main = "Empirical vs theoretical probabilities") points(y, dpoisinvgauss(y, 1, dis = 2.5), pch = 19, col = "red") legend("topright", c("empirical", "theoretical"), lty = c(1, NA), pch = c(NA, 19), col = c("black", "red"))
Quantile and Value-at-Risk methods for objects of class
"aggregateDist"
.
## S3 method for class 'aggregateDist' quantile(x, probs = c(0.25, 0.5, 0.75, 0.9, 0.95, 0.975, 0.99, 0.995), smooth = FALSE, names = TRUE, ...) ## S3 method for class 'aggregateDist' VaR(x, conf.level = c(0.9, 0.95, 0.99), smooth = FALSE, names = TRUE, ...)
## S3 method for class 'aggregateDist' quantile(x, probs = c(0.25, 0.5, 0.75, 0.9, 0.95, 0.975, 0.99, 0.995), smooth = FALSE, names = TRUE, ...) ## S3 method for class 'aggregateDist' VaR(x, conf.level = c(0.9, 0.95, 0.99), smooth = FALSE, names = TRUE, ...)
x |
an object of class |
probs , conf.level
|
numeric vector of probabilities with values
in |
smooth |
logical; when |
names |
logical; if true, the result has a |
... |
further arguments passed to or from other methods. |
The quantiles are taken directly from the cumulative distribution
function defined in x
. Linear interpolation is available for
step functions.
A numeric vector, named if names
is TRUE
.
Vincent Goulet [email protected] and Louis-Philippe Pouliot
model.freq <- expression(data = rpois(3)) model.sev <- expression(data = rlnorm(10, 1.5)) Fs <- aggregateDist("simulation", model.freq, model.sev, nb.simul = 1000) quantile(Fs, probs = c(0.25, 0.5, 0.75)) VaR(Fs)
model.freq <- expression(data = rpois(3)) model.sev <- expression(data = rlnorm(10, 1.5)) Fs <- aggregateDist("simulation", model.freq, model.sev, nb.simul = 1000) quantile(Fs, probs = c(0.25, 0.5, 0.75)) VaR(Fs)
Sample quantiles corresponding to the given probabilities for objects
of class "grouped.data"
.
## S3 method for class 'grouped.data' quantile(x, probs = seq(0, 1, 0.25), names = TRUE, ...) ## S3 method for class 'grouped.data' summary(object, ...)
## S3 method for class 'grouped.data' quantile(x, probs = seq(0, 1, 0.25), names = TRUE, ...) ## S3 method for class 'grouped.data' summary(object, ...)
x , object
|
an object of class |
probs |
numeric vector of probabilities with values
in |
names |
logical; if true, the result has a |
... |
further arguments passed to or from other methods. |
The quantile function is the inverse of the ogive, that is a linear interpolation of the empirical quantile function.
The equation of the quantile function is
for and where
are the
group
boundaries and
is the empirical distribution function of
the sample.
For quantile
, a numeric vector, named if names
is
TRUE
.
For the summary
method, an object of class
c("summaryDefault", "table")
which has specialized
format
and print
methods.
Vincent Goulet [email protected]
ogive
for the smoothed empirical distribution of which
quantile.grouped.data
is an inverse;
mean.grouped.data
and var.grouped.data
to
compute the mean and variance of grouped data.
grouped.data
to create grouped data objects.
data(gdental) quantile(gdental) summary(gdental) Fn <- ogive(gdental) Fn(quantile(gdental)) # inverse function
data(gdental) quantile(gdental) summary(gdental) Fn <- ogive(gdental) Fn(quantile(gdental)) # inverse function
Simulate data for insurance applications allowing hierarchical structures and separate models for the frequency and severity of claims distributions.
rcomphierarc
is an alias for simul
.
rcomphierarc(nodes, model.freq = NULL, model.sev = NULL, weights = NULL) ## S3 method for class 'portfolio' print(x, ...)
rcomphierarc(nodes, model.freq = NULL, model.sev = NULL, weights = NULL) ## S3 method for class 'portfolio' print(x, ...)
nodes |
a vector or a named list giving the number of "nodes" at each level in the hierarchy of the portfolio. The nodes are listed from top (portfolio) to bottom (usually the years of experience). |
model.freq |
a named vector of expressions specifying the
frequency of claims model (see Details); if |
model.sev |
a named vector of expressions specifying the severity
of claims model (see Details); if |
weights |
a vector of weights. |
x |
a |
... |
potential further arguments required by generic. |
The order and the names of the elements in nodes
,
model.freq
and model.sev
must match. At least one of
model.freq
and model.sev
must be non NULL
.
nodes
may be a basic vector, named or not, for non hierarchical
models. The rule above still applies, so model.freq
and
model.sev
should not be named if nodes
is not. However,
for non hierarchical models, rcompound
is faster and has
a simpler interface.
nodes
specifies the hierarchical layout of the portfolio. Each
element of the list is a vector of the number of nodes at a given
level. Vectors are recycled as necessary.
model.freq
and model.sev
specify the simulation models
for claim numbers and claim amounts, respectively. A model is
expressed in a semi-symbolic fashion using an object of mode
expression
. Each element of the object
must be named and should be a complete call to a random number
generation function, with the number of variates omitted. Hierarchical
(or mixtures of) models are achieved by replacing one or more
parameters of a distribution at a given level by any combination of
the names of the levels above. If no mixing is to take place at a
level, the model for this level can be NULL
.
The argument of the random number generation functions for the number
of variates to simulate must be named n
.
Weights will be used wherever the name "weights"
appears in a
model. It is the user's responsibility to ensure that the length of
weights
will match the number of nodes when weights are to be
used. Normally, there should be one weight per node at the lowest
level of the model.
Data is generated in lexicographic order, that is by row in the output matrix.
An object of class
"portfolio"
. A
print
method for this class displays the models used in the
simulation as well as the frequency of claims for each year and entity
in the portfolio.
An object of class "portfolio"
is a list containing the
following components:
data |
a two dimension list where each element is a vector of claim amounts; |
weights |
the vector of weights given in argument reshaped as a
matrix matching element |
classification |
a matrix of integers where each row is a unique
set of subscripts identifying an entity in the portfolio
(e.g. integers |
nodes |
the |
model.freq |
the frequency model as given in argument; |
model.sev |
the severity model as given in argument. |
It is recommended to manipulate objects of class "portfolio"
by
means of the corresponding methods of functions aggregate
,
frequency
and severity
.
Vincent Goulet [email protected], Sébastien Auclair and Louis-Philippe Pouliot
Goulet, V. and Pouliot, L.-P. (2008), Simulation of compound hierarchical models in R, North American Actuarial Journal 12, 401–412.
rcomphierarc.summaries
for the functions to create the
matrices of aggregate claim amounts, frequencies and individual claim
amounts.
rcompound
for a simpler and much faster way to generate
variates from standard, non hierarchical, compound models.
## Two level (contracts and years) portfolio with frequency model ## Nit|Theta_i ~ Poisson(Theta_i), Theta_i ~ Gamma(2, 3) and severity ## model X ~ Lognormal(5, 1) rcomphierarc(nodes = list(contract = 10, year = 5), model.freq = expression(contract = rgamma(2, 3), year = rpois(contract)), model.sev = expression(contract = NULL, year = rlnorm(5, 1))) ## Model with weights and mixtures for both frequency and severity ## models nodes <- list(entity = 8, year = c(5, 4, 4, 5, 3, 5, 4, 5)) mf <- expression(entity = rgamma(2, 3), year = rpois(weights * entity)) ms <- expression(entity = rnorm(5, 1), year = rlnorm(entity, 1)) wit <- sample(2:10, 35, replace = TRUE) pf <- rcomphierarc(nodes, mf, ms, wit) pf # print method weights(pf) # extraction of weights aggregate(pf)[, -1]/weights(pf)[, -1] # ratios ## Four level hierarchical model for frequency only nodes <- list(sector = 3, unit = c(3, 4), employer = c(3, 4, 3, 4, 2, 3, 4), year = 5) mf <- expression(sector = rexp(1), unit = rexp(sector), employer = rgamma(unit, 1), year = rpois(employer)) pf <- rcomphierarc(nodes, mf, NULL) pf # print method aggregate(pf) # aggregate claim amounts frequency(pf) # frequencies severity(pf) # individual claim amounts ## Standard, non hierarchical, compound model with simplified ## syntax (function rcompound() is much faster for such cases) rcomphierarc(10, model.freq = expression(rpois(2)), model.sev = expression(rgamma(2, 3)))
## Two level (contracts and years) portfolio with frequency model ## Nit|Theta_i ~ Poisson(Theta_i), Theta_i ~ Gamma(2, 3) and severity ## model X ~ Lognormal(5, 1) rcomphierarc(nodes = list(contract = 10, year = 5), model.freq = expression(contract = rgamma(2, 3), year = rpois(contract)), model.sev = expression(contract = NULL, year = rlnorm(5, 1))) ## Model with weights and mixtures for both frequency and severity ## models nodes <- list(entity = 8, year = c(5, 4, 4, 5, 3, 5, 4, 5)) mf <- expression(entity = rgamma(2, 3), year = rpois(weights * entity)) ms <- expression(entity = rnorm(5, 1), year = rlnorm(entity, 1)) wit <- sample(2:10, 35, replace = TRUE) pf <- rcomphierarc(nodes, mf, ms, wit) pf # print method weights(pf) # extraction of weights aggregate(pf)[, -1]/weights(pf)[, -1] # ratios ## Four level hierarchical model for frequency only nodes <- list(sector = 3, unit = c(3, 4), employer = c(3, 4, 3, 4, 2, 3, 4), year = 5) mf <- expression(sector = rexp(1), unit = rexp(sector), employer = rgamma(unit, 1), year = rpois(employer)) pf <- rcomphierarc(nodes, mf, NULL) pf # print method aggregate(pf) # aggregate claim amounts frequency(pf) # frequencies severity(pf) # individual claim amounts ## Standard, non hierarchical, compound model with simplified ## syntax (function rcompound() is much faster for such cases) rcomphierarc(10, model.freq = expression(rpois(2)), model.sev = expression(rgamma(2, 3)))
Methods for class "portfolio"
objects.
aggregate
splits portfolio data into subsets and computes
summary statistics for each.
frequency
computes the frequency of claims for subsets of
portfolio data.
severity
extracts the individual claim amounts.
weights
extracts the matrix of weights.
## S3 method for class 'portfolio' aggregate(x, by = names(x$nodes), FUN = sum, classification = TRUE, prefix = NULL, ...) ## S3 method for class 'portfolio' frequency(x, by = names(x$nodes), classification = TRUE, prefix = NULL, ...) ## S3 method for class 'portfolio' severity(x, by = head(names(x$node), -1), splitcol = NULL, classification = TRUE, prefix = NULL, ...) ## S3 method for class 'portfolio' weights(object, classification = TRUE, prefix = NULL, ...)
## S3 method for class 'portfolio' aggregate(x, by = names(x$nodes), FUN = sum, classification = TRUE, prefix = NULL, ...) ## S3 method for class 'portfolio' frequency(x, by = names(x$nodes), classification = TRUE, prefix = NULL, ...) ## S3 method for class 'portfolio' severity(x, by = head(names(x$node), -1), splitcol = NULL, classification = TRUE, prefix = NULL, ...) ## S3 method for class 'portfolio' weights(object, classification = TRUE, prefix = NULL, ...)
x , object
|
an object of class |
by |
character vector of grouping elements using the level names
of the portfolio in |
FUN |
the function to be applied to data subsets. |
classification |
boolean; if |
prefix |
characters to prefix column names with; if |
splitcol |
columns of the data matrix to extract separately; usual matrix indexing methods are supported. |
... |
optional arguments to |
By default, aggregate.portfolio
computes the aggregate claim amounts
for the grouping specified in by
. Any other statistic based on
the individual claim amounts can be used through argument FUN
.
frequency.portfolio
is equivalent to using aggregate.portfolio
with argument FUN
equal to if (identical(x, NA)) NA else
length(x)
.
severity.portfolio
extracts individual claim amounts of a portfolio
by groupings using the default method of severity
.
Argument splitcol
allows to get the individual claim amounts of
specific columns separately.
weights.portfolio
extracts the weight matrix of a portfolio.
A matrix or vector depending on the groupings specified in by
.
For the aggregate
and frequency
methods: if at least one
level other than the last one is used for grouping, the result is a
matrix obtained by binding the appropriate node identifiers extracted
from x$classification
if classification = TRUE
, and the
summaries per grouping. If the last level is used for grouping, the
column names of x$data
are retained; if the last level is not
used for grouping, the column name is replaced by the deparsed name of
FUN
. If only the last level is used (column summaries), a named
vector is returned.
For the severity
method: a list of two elements:
main |
|
split |
same as above, but for the columns specified in
|
For the weights
method: the weight matrix of the portfolio with
node identifiers if classification = TRUE
.
Vincent Goulet [email protected], Louis-Philippe Pouliot.
nodes <- list(sector = 3, unit = c(3, 4), employer = c(3, 4, 3, 4, 2, 3, 4), year = 5) model.freq <- expression(sector = rexp(1), unit = rexp(sector), employer = rgamma(unit, 1), year = rpois(employer)) model.sev <- expression(sector = rnorm(6, 0.1), unit = rnorm(sector, 1), employer = rnorm(unit, 1), year = rlnorm(employer, 1)) pf <- rcomphierarc(nodes, model.freq, model.sev) aggregate(pf) # aggregate claim amount by employer and year aggregate(pf, classification = FALSE) # same, without node identifiers aggregate(pf, by = "sector") # by sector aggregate(pf, by = "y") # by year aggregate(pf, by = c("s", "u"), mean) # average claim amount frequency(pf) # number of claims frequency(pf, prefix = "freq.") # more explicit column names severity(pf) # claim amounts by row severity(pf, by = "year") # by column severity(pf, by = c("s", "u")) # by unit severity(pf, splitcol = "year.5") # last year separate severity(pf, splitcol = 5) # same severity(pf, splitcol = c(FALSE, FALSE, FALSE, FALSE, TRUE)) # same weights(pf) ## For portfolios with weights, the following computes loss ratios. ## Not run: aggregate(pf, classif = FALSE) / weights(pf, classif = FALSE)
nodes <- list(sector = 3, unit = c(3, 4), employer = c(3, 4, 3, 4, 2, 3, 4), year = 5) model.freq <- expression(sector = rexp(1), unit = rexp(sector), employer = rgamma(unit, 1), year = rpois(employer)) model.sev <- expression(sector = rnorm(6, 0.1), unit = rnorm(sector, 1), employer = rnorm(unit, 1), year = rlnorm(employer, 1)) pf <- rcomphierarc(nodes, model.freq, model.sev) aggregate(pf) # aggregate claim amount by employer and year aggregate(pf, classification = FALSE) # same, without node identifiers aggregate(pf, by = "sector") # by sector aggregate(pf, by = "y") # by year aggregate(pf, by = c("s", "u"), mean) # average claim amount frequency(pf) # number of claims frequency(pf, prefix = "freq.") # more explicit column names severity(pf) # claim amounts by row severity(pf, by = "year") # by column severity(pf, by = c("s", "u")) # by unit severity(pf, splitcol = "year.5") # last year separate severity(pf, splitcol = 5) # same severity(pf, splitcol = c(FALSE, FALSE, FALSE, FALSE, TRUE)) # same weights(pf) ## For portfolios with weights, the following computes loss ratios. ## Not run: aggregate(pf, classif = FALSE) / weights(pf, classif = FALSE)
rcompound
generates random variates from a compound model.
rcomppois
is a simplified version for a common case.
rcompound(n, model.freq, model.sev, SIMPLIFY = TRUE) rcomppois(n, lambda, model.sev, SIMPLIFY = TRUE)
rcompound(n, model.freq, model.sev, SIMPLIFY = TRUE) rcomppois(n, lambda, model.sev, SIMPLIFY = TRUE)
n |
number of observations. If |
model.freq , model.sev
|
expressions specifying the frequency and severity simulation models with the number of variates omitted; see Details. |
lambda |
Poisson parameter. |
SIMPLIFY |
boolean; if |
rcompound
generates variates from a random variable of the form
where is the frequency random variable and
are the severity random variables. The latter are mutually
independent, identically distributed and independent from
.
model.freq
and model.sev
specify the simulation models
for the frequency and the severity random variables, respectively. A
model is a complete call to a random number generation function, with
the number of variates omitted. This is similar to
rcomphierarc
, but the calls need not be wrapped into
expression
. Either argument may also be the name of an
object containing an expression, in which case the object will be
evaluated in the parent frame to retrieve the expression.
The argument of the random number generation functions for the number
of variates to simulate must be named n
.
rcomppois
generates variates from the common Compound Poisson
model, that is when random variable is Poisson distributed
with mean
lambda
.
When SIMPLIFY = TRUE
, a vector of aggregate amounts .
When SIMPLIFY = FALSE
, a list of three elements:
aggregate |
vector of aggregate amounts |
frequency |
vector of frequencies |
severity |
vector of severities |
Vincent Goulet [email protected]
rcomphierarc
to simulate from compound hierarchical models.
## Compound Poisson model with gamma severity. rcompound(10, rpois(2), rgamma(2, 3)) rcomppois(10, 2, rgamma(2, 3)) # same ## Frequencies and individual claim amounts along with aggregate ## values. rcomppois(10, 2, rgamma(2, 3), SIMPLIFY = FALSE) ## Wrapping the simulation models into expression() is allowed, but ## not needed. rcompound(10, expression(rpois(2)), expression(rgamma(2, 3))) ## Not run: ## Speed comparison between rcompound() and rcomphierarc(). ## [Also note the simpler syntax for rcompound().] system.time(rcompound(1e6, rpois(2), rgamma(2, 3))) system.time(rcomphierarc(1e6, expression(rpois(2)), expression(rgamma(2, 3)))) ## End(Not run) ## The severity can itself be a compound model. It makes sense ## in such a case to use a zero-truncated frequency distribution ## for the second level model. rcomppois(10, 2, rcompound(rztnbinom(1.5, 0.7), rlnorm(1.2, 1)))
## Compound Poisson model with gamma severity. rcompound(10, rpois(2), rgamma(2, 3)) rcomppois(10, 2, rgamma(2, 3)) # same ## Frequencies and individual claim amounts along with aggregate ## values. rcomppois(10, 2, rgamma(2, 3), SIMPLIFY = FALSE) ## Wrapping the simulation models into expression() is allowed, but ## not needed. rcompound(10, expression(rpois(2)), expression(rgamma(2, 3))) ## Not run: ## Speed comparison between rcompound() and rcomphierarc(). ## [Also note the simpler syntax for rcompound().] system.time(rcompound(1e6, rpois(2), rgamma(2, 3))) system.time(rcomphierarc(1e6, expression(rpois(2)), expression(rgamma(2, 3)))) ## End(Not run) ## The severity can itself be a compound model. It makes sense ## in such a case to use a zero-truncated frequency distribution ## for the second level model. rcomppois(10, 2, rcompound(rztnbinom(1.5, 0.7), rlnorm(1.2, 1)))
Generate random variates from a discrete mixture of distributions.
rmixture(n, probs, models, shuffle = TRUE)
rmixture(n, probs, models, shuffle = TRUE)
n |
number of random variates to generate. If |
probs |
numeric non-negative vector specifying the probability
for each model; is internally normalized to sum 1. Infinite
and missing values are not allowed. Values are recycled as necessary
to match the length of |
models |
vector of expressions specifying the simulation models
with the number of variates omitted; see Details. Models are
recycled as necessary to match the length of |
shuffle |
logical; should the random variates from the distributions be shuffled? |
rmixture
generates variates from a discrete mixture, that is
the random variable with a probability density function of the form
where are densities and
.
The values in probs
will be internally normalized to be
used as probabilities .
The specification of simulation models uses the syntax of
rcomphierarc
. Models are expressed in a
semi-symbolic fashion using an object of mode
expression
where each element is a complete call
to a random number generation function, with the number of variates
omitted.
The argument of the random number generation functions for the number
of variates to simulate must be named n
.
If shuffle
is FALSE
, the output vector contains all the
random variates from the first model, then all the random variates
from the second model, and so on. If the order of the variates is
irrelevant, this cuts the time to generate the variates roughly in
half.
A vector of random variates from the mixture with density .
Building the expressions in models
from the arguments of
another function is delicate. The expressions must be such that
evaluation is possible in the frame of rmixture
or its parent.
See the examples.
Vincent Goulet [email protected]
rcompound
to simulate from compound models.
rcomphierarc
to simulate from compound hierarchical models.
## Mixture of two exponentials (with means 1/3 and 1/7) with equal ## probabilities. rmixture(10, 0.5, expression(rexp(3), rexp(7))) rmixture(10, 42, expression(rexp(3), rexp(7))) # same ## Mixture of two lognormals with different probabilities. rmixture(10, probs = c(0.55, 0.45), models = expression(rlnorm(3.6, 0.6), rlnorm(4.6, 0.3))) ## Building the model expressions in the following example ## works as 'rate' is defined in the parent frame of ## 'rmixture'. probs <- c(2, 5) g <- function(n, p, rate) rmixture(n, p, expression(rexp(rate[1]), rexp(rate[2]))) g(10, probs, c(3, 7)) ## The following example does not work: 'rate' does not exist ## in the evaluation frame of 'rmixture'. f <- function(n, p, model) rmixture(n, p, model) h <- function(n, p, rate) f(n, p, expression(rexp(rate[1]), rexp(rate[2]))) ## Not run: h(10, probs, c(3, 7)) ## Fix: substitute the values in the model expressions. h <- function(n, p, rate) { models <- eval(substitute(expression(rexp(a[1]), rexp(a[2])), list(a = rate))) f(n, p, models) } h(10, probs, c(3, 7))
## Mixture of two exponentials (with means 1/3 and 1/7) with equal ## probabilities. rmixture(10, 0.5, expression(rexp(3), rexp(7))) rmixture(10, 42, expression(rexp(3), rexp(7))) # same ## Mixture of two lognormals with different probabilities. rmixture(10, probs = c(0.55, 0.45), models = expression(rlnorm(3.6, 0.6), rlnorm(4.6, 0.3))) ## Building the model expressions in the following example ## works as 'rate' is defined in the parent frame of ## 'rmixture'. probs <- c(2, 5) g <- function(n, p, rate) rmixture(n, p, expression(rexp(rate[1]), rexp(rate[2]))) g(10, probs, c(3, 7)) ## The following example does not work: 'rate' does not exist ## in the evaluation frame of 'rmixture'. f <- function(n, p, model) rmixture(n, p, model) h <- function(n, p, rate) f(n, p, expression(rexp(rate[1]), rexp(rate[2]))) ## Not run: h(10, probs, c(3, 7)) ## Fix: substitute the values in the model expressions. h <- function(n, p, rate) { models <- eval(substitute(expression(rexp(a[1]), rexp(a[2])), list(a = rate))) f(n, p, models) } h(10, probs, c(3, 7))
Calulation of infinite time probability of ruin in the models of Cramér-Lundberg and Sparre Andersen, that is with exponential or phase-type (including mixtures of exponentials, Erlang and mixture of Erlang) claims interarrival time.
ruin(claims = c("exponential", "Erlang", "phase-type"), par.claims, wait = c("exponential", "Erlang", "phase-type"), par.wait, premium.rate = 1, tol = sqrt(.Machine$double.eps), maxit = 200L, echo = FALSE) ## S3 method for class 'ruin' plot(x, from = NULL, to = NULL, add = FALSE, xlab = "u", ylab = expression(psi(u)), main = "Probability of Ruin", xlim = NULL, ...)
ruin(claims = c("exponential", "Erlang", "phase-type"), par.claims, wait = c("exponential", "Erlang", "phase-type"), par.wait, premium.rate = 1, tol = sqrt(.Machine$double.eps), maxit = 200L, echo = FALSE) ## S3 method for class 'ruin' plot(x, from = NULL, to = NULL, add = FALSE, xlab = "u", ylab = expression(psi(u)), main = "Probability of Ruin", xlim = NULL, ...)
claims |
character; the type of claim severity distribution. |
wait |
character; the type of claim interarrival (wait) time distribution. |
par.claims , par.wait
|
named list containing the parameters of the distribution; see Details. |
premium.rate |
numeric vector of length 1; the premium rate. |
tol , maxit , echo
|
respectively the tolerance level of the
stopping criteria, the maximum number of iterations and whether or
not to echo the procedure when the transition rates matrix is
determined iteratively. Ignored if |
x |
an object of class |
from , to
|
the range over which the function will be plotted. |
add |
logical; if |
xlim |
numeric of length 2; if specified, it serves as default
for |
xlab , ylab
|
label of the x and y axes, respectively. |
main |
main title. |
... |
further graphical parameters accepted by
|
The names of the parameters in par.claims
and par.wait
must the same as in dexp
,
dgamma
or dphtype
, as appropriate.
A model will be a mixture of exponential or Erlang distributions (but
not phase-type) when the parameters are vectors of length
and the parameter list contains a vector
weights
of the
coefficients of the mixture.
Parameters are recycled when needed. Their names can be abbreviated.
Combinations of exponentials as defined in Dufresne and Gerber (1988) are not supported.
Ruin probabilities are evaluated using pphtype
except
when both distributions are exponential, in which case an explicit
formula is used.
When wait != "exponential"
(Sparre Andersen model), the
transition rate matrix of the distribution of
the probability of ruin is determined iteratively using a fixed
point-like algorithm. The stopping criteria used is
where and
are
two successive values of the matrix.
A function of class "ruin"
inheriting from the
"function"
class to compute the probability of ruin given
initial surplus levels. The function has arguments:
u |
numeric vector of initial surplus levels; |
survival |
logical; if |
lower.tail |
an alias for |
Vincent Goulet [email protected], and Christophe Dutang
Asmussen, S. and Rolski, T. (1991), Computational methods in risk theory: A matrix algorithmic approach, Insurance: Mathematics and Economics 10, 259–274.
Dufresne, F. and Gerber, H. U. (1988), Three methods to calculate the probability of ruin, Astin Bulletin 19, 71–90.
Gerber, H. U. (1979), An Introduction to Mathematical Risk Theory, Huebner Foundation.
## Case with an explicit formula: exponential claims and exponential ## interarrival times. psi <- ruin(claims = "e", par.claims = list(rate = 5), wait = "e", par.wait = list(rate = 3)) psi psi(0:10) plot(psi, from = 0, to = 10) ## Mixture of two exponentials for claims, exponential interarrival ## times (Gerber 1979) psi <- ruin(claims = "e", par.claims = list(rate = c(3, 7), w = 0.5), wait = "e", par.wait = list(rate = 3), pre = 1) u <- 0:10 psi(u) (24 * exp(-u) + exp(-6 * u))/35 # same ## Phase-type claims, exponential interarrival times (Asmussen and ## Rolski 1991) p <- c(0.5614, 0.4386) r <- matrix(c(-8.64, 0.101, 1.997, -1.095), 2, 2) lambda <- 1/(1.1 * mphtype(1, p, r)) psi <- ruin(claims = "p", par.claims = list(prob = p, rates = r), wait = "e", par.wait = list(rate = lambda)) psi plot(psi, xlim = c(0, 50)) ## Phase-type claims, mixture of two exponentials for interarrival times ## (Asmussen and Rolski 1991) a <- (0.4/5 + 0.6) * lambda ruin(claims = "p", par.claims = list(prob = p, rates = r), wait = "e", par.wait = list(rate = c(5 * a, a), weights = c(0.4, 0.6)), maxit = 225L)
## Case with an explicit formula: exponential claims and exponential ## interarrival times. psi <- ruin(claims = "e", par.claims = list(rate = 5), wait = "e", par.wait = list(rate = 3)) psi psi(0:10) plot(psi, from = 0, to = 10) ## Mixture of two exponentials for claims, exponential interarrival ## times (Gerber 1979) psi <- ruin(claims = "e", par.claims = list(rate = c(3, 7), w = 0.5), wait = "e", par.wait = list(rate = 3), pre = 1) u <- 0:10 psi(u) (24 * exp(-u) + exp(-6 * u))/35 # same ## Phase-type claims, exponential interarrival times (Asmussen and ## Rolski 1991) p <- c(0.5614, 0.4386) r <- matrix(c(-8.64, 0.101, 1.997, -1.095), 2, 2) lambda <- 1/(1.1 * mphtype(1, p, r)) psi <- ruin(claims = "p", par.claims = list(prob = p, rates = r), wait = "e", par.wait = list(rate = lambda)) psi plot(psi, xlim = c(0, 50)) ## Phase-type claims, mixture of two exponentials for interarrival times ## (Asmussen and Rolski 1991) a <- (0.4/5 + 0.6) * lambda ruin(claims = "p", par.claims = list(prob = p, rates = r), wait = "e", par.wait = list(rate = c(5 * a, a), weights = c(0.4, 0.6)), maxit = 225L)
severity
is a generic function created to manipulate individual
claim amounts. The function invokes particular methods which
depend on the class
of the first argument.
severity(x, ...) ## Default S3 method: severity(x, bycol = FALSE, drop = TRUE, ...)
severity(x, ...) ## Default S3 method: severity(x, bycol = FALSE, drop = TRUE, ...)
x |
an R object. |
bycol |
logical; whether to “unroll” horizontally
( |
... |
further arguments to be passed to or from other methods. |
drop |
logical; if |
Currently, the default method is equivalent to
unroll
. This is liable to change since the link between
the name and the use of the function is rather weak.
A vector or matrix.
Vincent Goulet [email protected] and Louis-Philippe Pouliot
severity.portfolio
for the original motivation of these
functions.
x <- list(c(1:3), c(1:8), c(1:4), c(1:3)) (mat <- matrix(x, 2, 2)) severity(mat) severity(mat, bycol = TRUE)
x <- list(c(1:3), c(1:8), c(1:4), c(1:3)) (mat <- matrix(x, 2, 2)) severity(mat) severity(mat, bycol = TRUE)
Density function, distribution function, quantile function, random generation,
raw moments, and limited moments for the Single-parameter Pareto
distribution with parameter shape
.
dpareto1(x, shape, min, log = FALSE) ppareto1(q, shape, min, lower.tail = TRUE, log.p = FALSE) qpareto1(p, shape, min, lower.tail = TRUE, log.p = FALSE) rpareto1(n, shape, min) mpareto1(order, shape, min) levpareto1(limit, shape, min, order = 1)
dpareto1(x, shape, min, log = FALSE) ppareto1(q, shape, min, lower.tail = TRUE, log.p = FALSE) qpareto1(p, shape, min, lower.tail = TRUE, log.p = FALSE) rpareto1(n, shape, min) mpareto1(order, shape, min) levpareto1(limit, shape, min, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
shape |
parameter. Must be strictly positive. |
min |
lower bound of the support of the distribution. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The single-parameter Pareto, or Pareto I, distribution with parameter
shape
has density:
for ,
and
.
Although there appears to be two parameters, only shape
is a true
parameter. The value of min
must be set in
advance.
The th raw moment of the random variable
is
,
and the
th
limited moment at some limit
is
,
.
dpareto1
gives the density,
ppareto1
gives the distribution function,
qpareto1
gives the quantile function,
rpareto1
generates random deviates,
mpareto1
gives the th raw moment, and
levpareto1
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
For Pareto distributions, we use the classification of Arnold (2015) with the parametrization of Klugman et al. (2012).
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected] and Mathieu Pigeon
Arnold, B.C. (2015), Pareto Distributions, Second Edition, CRC Press.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dpareto
for the two-parameter Pareto distribution.
exp(dpareto1(5, 3, 4, log = TRUE)) p <- (1:10)/10 ppareto1(qpareto1(p, 2, 3), 2, 3) mpareto1(2, 3, 4) - mpareto(1, 3, 4) ^ 2 levpareto(10, 3, 4, order = 2)
exp(dpareto1(5, 3, 4, log = TRUE)) p <- (1:10)/10 ppareto1(qpareto1(p, 2, 3), 2, 3) mpareto1(2, 3, 4) - mpareto(1, 3, 4) ^ 2 levpareto(10, 3, 4, order = 2)
Density function, distribution function, quantile function, random generation,
raw moments and limited moments for the Transformed Beta distribution
with parameters shape1
, shape2
, shape3
and
scale
.
dtrbeta(x, shape1, shape2, shape3, rate = 1, scale = 1/rate, log = FALSE) ptrbeta(q, shape1, shape2, shape3, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qtrbeta(p, shape1, shape2, shape3, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rtrbeta(n, shape1, shape2, shape3, rate = 1, scale = 1/rate) mtrbeta(order, shape1, shape2, shape3, rate = 1, scale = 1/rate) levtrbeta(limit, shape1, shape2, shape3, rate = 1, scale = 1/rate, order = 1)
dtrbeta(x, shape1, shape2, shape3, rate = 1, scale = 1/rate, log = FALSE) ptrbeta(q, shape1, shape2, shape3, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qtrbeta(p, shape1, shape2, shape3, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rtrbeta(n, shape1, shape2, shape3, rate = 1, scale = 1/rate) mtrbeta(order, shape1, shape2, shape3, rate = 1, scale = 1/rate) levtrbeta(limit, shape1, shape2, shape3, rate = 1, scale = 1/rate, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
shape1 , shape2 , shape3 , scale
|
parameters. Must be strictly positive. |
rate |
an alternative way to specify the scale. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The transformed beta distribution with parameters shape1
,
shape2
,
shape3
and
scale
, has
density:
for ,
,
,
and
.
(Here
is the function implemented
by R's
gamma()
and defined in its help.)
The transformed beta is the distribution of the random variable
where has a beta distribution with parameters
and
.
The transformed beta distribution defines a family of distributions with the following special cases:
A Burr distribution when shape3 == 1
;
A loglogistic distribution when shape1
== shape3 == 1
;
A paralogistic distribution when
shape3 == 1
and shape2 == shape1
;
A generalized Pareto distribution when
shape2 == 1
;
A Pareto distribution when shape2 ==
shape3 == 1
;
An inverse Burr distribution when
shape1 == 1
;
An inverse Pareto distribution when
shape2 == shape1 == 1
;
An inverse paralogistic distribution
when shape1 == 1
and shape3 == shape2
.
The th raw moment of the random variable
is
,
.
The th limited moment at some limit
is
,
and
not a negative integer.
dtrbeta
gives the density,
ptrbeta
gives the distribution function,
qtrbeta
gives the quantile function,
rtrbeta
generates random deviates,
mtrbeta
gives the th raw moment, and
levtrbeta
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
levtrbeta
computes the limited expected value using
betaint
.
Distribution also known as the Generalized Beta of the Second Kind and Pearson Type VI. See also Kleiber and Kotz (2003) for alternative names and parametrizations.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected] and Mathieu Pigeon
Kleiber, C. and Kotz, S. (2003), Statistical Size Distributions in Economics and Actuarial Sciences, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dfpareto
for an equivalent distribution with a location
parameter.
exp(dtrbeta(2, 2, 3, 4, 5, log = TRUE)) p <- (1:10)/10 ptrbeta(qtrbeta(p, 2, 3, 4, 5), 2, 3, 4, 5) qpearson6(0.3, 2, 3, 4, 5, lower.tail = FALSE) ## variance mtrbeta(2, 2, 3, 4, 5) - mtrbeta(1, 2, 3, 4, 5)^2 ## case with shape1 - order/shape2 > 0 levtrbeta(10, 2, 3, 4, scale = 1, order = 2) ## case with shape1 - order/shape2 < 0 levtrbeta(10, 1/3, 0.75, 4, scale = 0.5, order = 2)
exp(dtrbeta(2, 2, 3, 4, 5, log = TRUE)) p <- (1:10)/10 ptrbeta(qtrbeta(p, 2, 3, 4, 5), 2, 3, 4, 5) qpearson6(0.3, 2, 3, 4, 5, lower.tail = FALSE) ## variance mtrbeta(2, 2, 3, 4, 5) - mtrbeta(1, 2, 3, 4, 5)^2 ## case with shape1 - order/shape2 > 0 levtrbeta(10, 2, 3, 4, scale = 1, order = 2) ## case with shape1 - order/shape2 < 0 levtrbeta(10, 1/3, 0.75, 4, scale = 0.5, order = 2)
Density function, distribution function, quantile function, random generation,
raw moments and limited moments for the Transformed Gamma distribution
with parameters shape1
, shape2
and scale
.
dtrgamma(x, shape1, shape2, rate = 1, scale = 1/rate, log = FALSE) ptrgamma(q, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qtrgamma(p, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rtrgamma(n, shape1, shape2, rate = 1, scale = 1/rate) mtrgamma(order, shape1, shape2, rate = 1, scale = 1/rate) levtrgamma(limit, shape1, shape2, rate = 1, scale = 1/rate, order = 1)
dtrgamma(x, shape1, shape2, rate = 1, scale = 1/rate, log = FALSE) ptrgamma(q, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) qtrgamma(p, shape1, shape2, rate = 1, scale = 1/rate, lower.tail = TRUE, log.p = FALSE) rtrgamma(n, shape1, shape2, rate = 1, scale = 1/rate) mtrgamma(order, shape1, shape2, rate = 1, scale = 1/rate) levtrgamma(limit, shape1, shape2, rate = 1, scale = 1/rate, order = 1)
x , q
|
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
shape1 , shape2 , scale
|
parameters. Must be strictly positive. |
rate |
an alternative way to specify the scale. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
order |
order of the moment. |
limit |
limit of the loss variable. |
The transformed gamma distribution with parameters shape1
,
shape2
and
scale
has density:
for ,
,
and
.
(Here
is the function implemented
by R's
gamma()
and defined in its help.)
The transformed gamma is the distribution of the random variable
where
has a gamma distribution with shape parameter
and scale parameter
or, equivalently, of the
random variable
with
a gamma distribution with shape parameter
and scale parameter
.
The transformed gamma probability distribution defines a family of distributions with the following special cases:
A Gamma distribution when shape2 == 1
;
A Weibull distribution when shape1 ==
1
;
An Exponential distribution when shape2 ==
shape1 == 1
.
The th raw moment of the random variable
is
and the
th limited moment at some limit
is
,
.
dtrgamma
gives the density,
ptrgamma
gives the distribution function,
qtrgamma
gives the quantile function,
rtrgamma
generates random deviates,
mtrgamma
gives the th raw moment, and
levtrgamma
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
Distribution also known as the Generalized Gamma. See also Kleiber and Kotz (2003) for alternative names and parametrizations.
The "distributions"
package vignette provides the
interrelations between the continuous size distributions in
actuar and the complete formulas underlying the above functions.
Vincent Goulet [email protected] and Mathieu Pigeon
Kleiber, C. and Kotz, S. (2003), Statistical Size Distributions in Economics and Actuarial Sciences, Wiley.
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
exp(dtrgamma(2, 3, 4, 5, log = TRUE)) p <- (1:10)/10 ptrgamma(qtrgamma(p, 2, 3, 4), 2, 3, 4) mtrgamma(2, 3, 4, 5) - mtrgamma(1, 3, 4, 5) ^ 2 levtrgamma(10, 3, 4, 5, order = 2)
exp(dtrgamma(2, 3, 4, 5, log = TRUE)) p <- (1:10)/10 ptrgamma(qtrgamma(p, 2, 3, 4), 2, 3, 4) mtrgamma(2, 3, 4, 5) - mtrgamma(1, 3, 4, 5) ^ 2 levtrgamma(10, 3, 4, 5, order = 2)
Raw moments, limited moments and moment generating function for the
Uniform distribution from min
to max
.
munif(order, min = 0, max = 1) levunif(limit, min = 0, max =1, order = 1) mgfunif(t, min = 0, max = 1, log = FALSE)
munif(order, min = 0, max = 1) levunif(limit, min = 0, max =1, order = 1) mgfunif(t, min = 0, max = 1, log = FALSE)
order |
order of the moment. |
min , max
|
lower and upper limits of the distribution. Must be finite. |
limit |
limit of the random variable. |
t |
numeric vector. |
log |
logical; if |
The th raw moment of the random variable
is
, the
th limited moment at some limit
is
and the moment
generating function is
.
munif
gives the th raw moment,
levunif
gives the th moment of the limited random
variable, and
mgfunif
gives the moment generating function in t
.
Invalid arguments will result in return value NaN
, with a warning.
Vincent Goulet [email protected], Christophe Dutang
https://en.wikipedia.org/wiki/Uniform_distribution_%28continuous%29
munif(-1) munif(1:5) levunif(3, order=1:5) levunif(3, 2, 4) mgfunif(1, 1, 2)
munif(-1) munif(1:5) levunif(3, order=1:5) levunif(3, 2, 4) mgfunif(1, 1, 2)
Displays all values of a matrix of vectors by “unrolling” the object vertically or horizontally.
unroll(x, bycol = FALSE, drop = TRUE)
unroll(x, bycol = FALSE, drop = TRUE)
x |
a list of vectors with a |
bycol |
logical; whether to unroll horizontally
( |
drop |
logical; if |
unroll
returns a matrix where elements of x
are concatenated (“unrolled”) by row (bycol = FALSE
) or
by column (bycol = TRUE
). NA
is used to make
rows/columns of equal length.
Vectors and one dimensional arrays are coerced to row matrices.
A vector or matrix.
Vincent Goulet [email protected] and Louis-Philippe Pouliot
This function was originally written for use in
severity.portfolio
.
x <- list(c(1:3), c(1:8), c(1:4), c(1:3)) (mat <- matrix(x, 2, 2)) unroll(mat) unroll(mat, bycol = TRUE) unroll(mat[1, ]) unroll(mat[1, ], drop = FALSE)
x <- list(c(1:3), c(1:8), c(1:4), c(1:3)) (mat <- matrix(x, 2, 2)) unroll(mat) unroll(mat, bycol = TRUE) unroll(mat[1, ]) unroll(mat[1, ], drop = FALSE)
Generic functions for the variance and standard deviation, and methods for individual and grouped data.
The default methods for individual data are the functions from the stats package.
var(x, ...) ## Default S3 method: var(x, y = NULL, na.rm = FALSE, use, ...) ## S3 method for class 'grouped.data' var(x, ...) sd(x, ...) ## Default S3 method: sd(x, na.rm = FALSE, ...) ## S3 method for class 'grouped.data' sd(x, ...)
var(x, ...) ## Default S3 method: var(x, y = NULL, na.rm = FALSE, use, ...) ## S3 method for class 'grouped.data' var(x, ...) sd(x, ...) ## Default S3 method: sd(x, na.rm = FALSE, ...) ## S3 method for class 'grouped.data' sd(x, ...)
x |
a vector or matrix of individual data, or an object of class
|
y |
see |
na.rm |
see |
use |
see |
... |
further arguments passed to or from other methods. |
This page documents variance and standard deviation computations for
grouped data. For individual data, see var
and
sd
from the stats package.
For grouped data with group boundaries and group frequencies
,
var
computes the sample variance
where
is the midpoint of the
th interval,
is the sample mean (or sample first moment) of the data,
and
.
The sample sample standard deviation is the square root of the sample
variance.
The sample variance for grouped data differs from the variance
computed from the empirical raw moments with emm
in two
aspects. First, it takes into account the degrees of freedom. Second,
it applies Sheppard's correction factor to compensate for the
overestimation of the true variation in the data. For groups of equal
width , Sheppard's correction factor is equal to
.
A named vector of variances or standard deviations.
Vincent Goulet [email protected]. Variance and standard deviation methods for grouped data contributed by Walter Garcia-Fontes [email protected].
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (1998), Loss Models, From Data to Decisions, Wiley.
Heumann, C., Schomaker, M., Shalabh (2016), Introduction to Statistics and Data Analysis, Springer.
grouped.data
to create grouped data objects;
mean.grouped.data
for the mean and emm
for
higher moments.
data(gdental) var(gdental) sd(gdental) ## Illustration of Sheppard's correction factor cj <- c(0, 2, 4, 6, 8) nj <- c(1, 5, 3, 2) gd <- grouped.data(Group = cj, Frequency = nj) (sum(nj) - 1)/sum(nj) * var(gd) (emm(gd, 2) - emm(gd)^2) - 4/12
data(gdental) var(gdental) sd(gdental) ## Illustration of Sheppard's correction factor cj <- c(0, 2, 4, 6, 8) nj <- c(1, 5, 3, 2) gd <- grouped.data(Group = cj, Frequency = nj) (sum(nj) - 1)/sum(nj) * var(gd) (emm(gd, 2) - emm(gd)^2) - 4/12
Value at Risk.
VaR(x, ...)
VaR(x, ...)
x |
an R object. |
... |
further arguments passed to or from other methods. |
This is a generic function with, currently, only a method for objects
of class "aggregateDist"
.
An object of class numeric
.
Vincent Goulet [email protected] and Tommy Ouellet
VaR.aggregateDist
, aggregateDist
Raw moments and limited moments for the Weibull distribution with
parameters shape
and scale
.
mweibull(order, shape, scale = 1) levweibull(limit, shape, scale = 1, order = 1)
mweibull(order, shape, scale = 1) levweibull(limit, shape, scale = 1, order = 1)
order |
order of the moment. |
limit |
limit of the loss variable. |
shape , scale
|
shape and scale parameters, the latter defaulting to 1. |
The th raw moment of the random variable
is
and the
th limited moment at some limit
is
,
.
mweibull
gives the th raw moment and
levweibull
gives the th moment of the limited loss
variable.
Invalid arguments will result in return value NaN
, with a warning.
Vincent Goulet [email protected] and Mathieu Pigeon
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
Weibull
for details on the Weibull distribution and
functions [dpqr]weibull
.
mweibull(2, 3, 4) - mweibull(1, 3, 4)^2 levweibull(10, 3, 4, order = 2)
mweibull(2, 3, 4) - mweibull(1, 3, 4)^2 levweibull(10, 3, 4, order = 2)
Density function, distribution function, quantile function and random
generation for the Zero-Modified Binomial distribution with
parameters size
and prob
, and probability at zero
p0
.
dzmbinom(x, size, prob, p0, log = FALSE) pzmbinom(q, size, prob, p0, lower.tail = TRUE, log.p = FALSE) qzmbinom(p, size, prob, p0, lower.tail = TRUE, log.p = FALSE) rzmbinom(n, size, prob, p0)
dzmbinom(x, size, prob, p0, log = FALSE) pzmbinom(q, size, prob, p0, lower.tail = TRUE, log.p = FALSE) qzmbinom(p, size, prob, p0, lower.tail = TRUE, log.p = FALSE) rzmbinom(n, size, prob, p0)
x |
vector of (strictly positive integer) quantiles. |
q |
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
size |
number of trials (strictly positive integer). |
prob |
probability of success on each trial. |
p0 |
probability mass at zero. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
The zero-modified binomial distribution with size
,
prob
and
p0
is a discrete
mixture between a degenerate distribution at zero and a (standard)
binomial. The probability mass function is
and
for ,
and
, where
is the probability mass
function of the binomial.
The cumulative distribution function is
The mean is and the variance is
,
where
and
are the mean and variance of
the zero-truncated binomial.
In the terminology of Klugman et al. (2012), the zero-modified
binomial is a member of the class of
distributions with
and
.
The special case p0 == 0
is the zero-truncated binomial.
If an element of x
is not integer, the result of
dzmbinom
is zero, with a warning.
The quantile is defined as the smallest value such that
, where
is the distribution function.
dzmbinom
gives the probability mass function,
pzmbinom
gives the distribution function,
qzmbinom
gives the quantile function, and
rzmbinom
generates random deviates.
Invalid size
, prob
or p0
will result in return
value NaN
, with a warning.
The length of the result is determined by n
for
rzmbinom
, and is the maximum of the lengths of the
numerical arguments for the other functions.
Functions {d,p,q}zmbinom
use {d,p,q}binom
for all
but the trivial input values and .
Vincent Goulet [email protected]
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dbinom
for the binomial distribution.
dztbinom
for the zero-truncated binomial distribution.
dzmbinom(1:5, size = 5, prob = 0.4, p0 = 0.2) (1-0.2) * dbinom(1:5, 5, 0.4)/pbinom(0, 5, 0.4, lower = FALSE) # same ## simple relation between survival functions pzmbinom(0:5, 5, 0.4, p0 = 0.2, lower = FALSE) (1-0.2) * pbinom(0:5, 5, 0.4, lower = FALSE) / pbinom(0, 5, 0.4, lower = FALSE) # same qzmbinom(pzmbinom(1:10, 10, 0.6, p0 = 0.1), 10, 0.6, p0 = 0.1) n <- 8; p <- 0.3; p0 <- 0.025 x <- 0:n title <- paste("ZM Binomial(", n, ", ", p, ", p0 = ", p0, ") and Binomial(", n, ", ", p,") PDF", sep = "") plot(x, dzmbinom(x, n, p, p0), type = "h", lwd = 2, ylab = "p(x)", main = title) points(x, dbinom(x, n, p), pch = 19, col = "red") legend("topright", c("ZT binomial probabilities", "Binomial probabilities"), col = c("black", "red"), lty = c(1, 0), lwd = 2, pch = c(NA, 19))
dzmbinom(1:5, size = 5, prob = 0.4, p0 = 0.2) (1-0.2) * dbinom(1:5, 5, 0.4)/pbinom(0, 5, 0.4, lower = FALSE) # same ## simple relation between survival functions pzmbinom(0:5, 5, 0.4, p0 = 0.2, lower = FALSE) (1-0.2) * pbinom(0:5, 5, 0.4, lower = FALSE) / pbinom(0, 5, 0.4, lower = FALSE) # same qzmbinom(pzmbinom(1:10, 10, 0.6, p0 = 0.1), 10, 0.6, p0 = 0.1) n <- 8; p <- 0.3; p0 <- 0.025 x <- 0:n title <- paste("ZM Binomial(", n, ", ", p, ", p0 = ", p0, ") and Binomial(", n, ", ", p,") PDF", sep = "") plot(x, dzmbinom(x, n, p, p0), type = "h", lwd = 2, ylab = "p(x)", main = title) points(x, dbinom(x, n, p), pch = 19, col = "red") legend("topright", c("ZT binomial probabilities", "Binomial probabilities"), col = c("black", "red"), lty = c(1, 0), lwd = 2, pch = c(NA, 19))
Density function, distribution function, quantile function and random
generation for the Zero-Modified Geometric distribution with
parameter prob
and arbitrary probability at zero p0
.
dzmgeom(x, prob, p0, log = FALSE) pzmgeom(q, prob, p0, lower.tail = TRUE, log.p = FALSE) qzmgeom(p, prob, p0, lower.tail = TRUE, log.p = FALSE) rzmgeom(n, prob, p0)
dzmgeom(x, prob, p0, log = FALSE) pzmgeom(q, prob, p0, lower.tail = TRUE, log.p = FALSE) qzmgeom(p, prob, p0, lower.tail = TRUE, log.p = FALSE) rzmgeom(n, prob, p0)
x |
vector of (strictly positive integer) quantiles. |
q |
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
prob |
parameter. |
p0 |
probability mass at zero. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
The zero-modified geometric distribution with prob
and
p0
is a discrete mixture between a
degenerate distribution at zero and a (standard) geometric. The
probability mass function is
and
for ,
and
, where
is the probability mass
function of the geometric.
The cumulative distribution function is
The mean is and the variance is
,
where
and
are the mean and variance of
the zero-truncated geometric.
In the terminology of Klugman et al. (2012), the zero-modified
geometric is a member of the class of
distributions with
and
.
The special case p0 == 0
is the zero-truncated geometric.
If an element of x
is not integer, the result of
dzmgeom
is zero, with a warning.
The quantile is defined as the smallest value such that
, where
is the distribution function.
dzmgeom
gives the (log) probability mass function,
pzmgeom
gives the (log) distribution function,
qzmgeom
gives the quantile function, and
rzmgeom
generates random deviates.
Invalid prob
or p0
will result in return value
NaN
, with a warning.
The length of the result is determined by n
for
rzmgeom
, and is the maximum of the lengths of the
numerical arguments for the other functions.
Functions {d,p,q}zmgeom
use {d,p,q}geom
for all but
the trivial input values and .
Vincent Goulet [email protected]
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dgeom
for the geometric distribution.
dztgeom
for the zero-truncated geometric distribution.
dzmnbinom
for the zero-modified negative binomial, of
which the zero-modified geometric is a special case.
p <- 1/(1 + 0.5) dzmgeom(1:5, prob = p, p0 = 0.6) (1-0.6) * dgeom(1:5, p)/pgeom(0, p, lower = FALSE) # same ## simple relation between survival functions pzmgeom(0:5, p, p0 = 0.2, lower = FALSE) (1-0.2) * pgeom(0:5, p, lower = FALSE)/pgeom(0, p, lower = FALSE) # same qzmgeom(pzmgeom(0:10, 0.3, p0 = 0.6), 0.3, p0 = 0.6)
p <- 1/(1 + 0.5) dzmgeom(1:5, prob = p, p0 = 0.6) (1-0.6) * dgeom(1:5, p)/pgeom(0, p, lower = FALSE) # same ## simple relation between survival functions pzmgeom(0:5, p, p0 = 0.2, lower = FALSE) (1-0.2) * pgeom(0:5, p, lower = FALSE)/pgeom(0, p, lower = FALSE) # same qzmgeom(pzmgeom(0:10, 0.3, p0 = 0.6), 0.3, p0 = 0.6)
Density function, distribution function, quantile function and random
generation for the Zero-Modified Logarithmic (or log-series)
distribution with parameter prob
and arbitrary probability at
zero p0
.
dzmlogarithmic(x, prob, p0, log = FALSE) pzmlogarithmic(q, prob, p0, lower.tail = TRUE, log.p = FALSE) qzmlogarithmic(p, prob, p0, lower.tail = TRUE, log.p = FALSE) rzmlogarithmic(n, prob, p0)
dzmlogarithmic(x, prob, p0, log = FALSE) pzmlogarithmic(q, prob, p0, lower.tail = TRUE, log.p = FALSE) qzmlogarithmic(p, prob, p0, lower.tail = TRUE, log.p = FALSE) rzmlogarithmic(n, prob, p0)
x |
vector of (strictly positive integer) quantiles. |
q |
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
prob |
parameter. |
p0 |
probability mass at zero. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
The zero-modified logarithmic distribution with prob
and
p0
is a discrete mixture between a
degenerate distribution at zero and a (standard) logarithmic. The
probability mass function is
and
for ,
and
, where
is the probability mass
function of the logarithmic.
The cumulative distribution function is
The special case p0 == 0
is the standard logarithmic.
The zero-modified logarithmic distribution is the limiting case of the
zero-modified negative binomial distribution with size
parameter equal to . Note that in this context, parameter
prob
generally corresponds to the probability of failure
of the zero-truncated negative binomial.
If an element of x
is not integer, the result of
dzmlogarithmic
is zero, with a warning.
The quantile is defined as the smallest value such that
, where
is the distribution function.
dzmlogarithmic
gives the probability mass function,
pzmlogarithmic
gives the distribution function,
qzmlogarithmic
gives the quantile function, and
rzmlogarithmic
generates random deviates.
Invalid prob
or p0
will result in return value
NaN
, with a warning.
The length of the result is determined by n
for
rzmlogarithmic
, and is the maximum of the lengths of the
numerical arguments for the other functions.
Functions {d,p,q}zmlogarithmic
use
{d,p,q}logarithmic
for all but the trivial input values and
.
Vincent Goulet [email protected]
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dlogarithmic
for the logarithmic distribution.
dztnbinom
for the zero modified negative binomial
distribution.
p <- 1/(1 + 0.5) dzmlogarithmic(1:5, prob = p, p0 = 0.6) (1-0.6) * dlogarithmic(1:5, p)/plogarithmic(0, p, lower = FALSE) # same ## simple relation between survival functions pzmlogarithmic(0:5, p, p0 = 0.2, lower = FALSE) (1-0.2) * plogarithmic(0:5, p, lower = FALSE)/plogarithmic(0, p, lower = FALSE) # same qzmlogarithmic(pzmlogarithmic(0:10, 0.3, p0 = 0.6), 0.3, p0 = 0.6)
p <- 1/(1 + 0.5) dzmlogarithmic(1:5, prob = p, p0 = 0.6) (1-0.6) * dlogarithmic(1:5, p)/plogarithmic(0, p, lower = FALSE) # same ## simple relation between survival functions pzmlogarithmic(0:5, p, p0 = 0.2, lower = FALSE) (1-0.2) * plogarithmic(0:5, p, lower = FALSE)/plogarithmic(0, p, lower = FALSE) # same qzmlogarithmic(pzmlogarithmic(0:10, 0.3, p0 = 0.6), 0.3, p0 = 0.6)
Density function, distribution function, quantile function and random
generation for the Zero-Modified Negative Binomial distribution with
parameters size
and prob
, and arbitrary probability at
zero p0
.
dzmnbinom(x, size, prob, p0, log = FALSE) pzmnbinom(q, size, prob, p0, lower.tail = TRUE, log.p = FALSE) qzmnbinom(p, size, prob, p0, lower.tail = TRUE, log.p = FALSE) rzmnbinom(n, size, prob, p0)
dzmnbinom(x, size, prob, p0, log = FALSE) pzmnbinom(q, size, prob, p0, lower.tail = TRUE, log.p = FALSE) qzmnbinom(p, size, prob, p0, lower.tail = TRUE, log.p = FALSE) rzmnbinom(n, size, prob, p0)
x |
vector of (strictly positive integer) quantiles. |
q |
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
size |
target for number of successful trials, or dispersion parameter. Must be positive, need not be integer. |
prob |
parameter. |
p0 |
probability mass at zero. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
The zero-modified negative binomial distribution with size
,
prob
and
p0
is a
discrete mixture between a degenerate distribution at zero and a
(standard) negative binomial. The probability mass function is
and
for ,
,
and
, where
is the probability mass
function of the negative binomial.
The cumulative distribution function is
The mean is and the variance is
,
where
and
are the mean and variance of
the zero-truncated negative binomial.
In the terminology of Klugman et al. (2012), the zero-modified
negative binomial is a member of the class of
distributions with
and
.
The special case p0 == 0
is the zero-truncated negative
binomial.
The limiting case size == 0
is the zero-modified logarithmic
distribution with parameters 1 - prob
and p0
.
Unlike the standard negative binomial functions, parametrization
through the mean mu
is not supported to avoid ambiguity as
to whether mu
is the mean of the underlying negative binomial
or the mean of the zero-modified distribution.
If an element of x
is not integer, the result of
dzmnbinom
is zero, with a warning.
The quantile is defined as the smallest value such that
, where
is the distribution function.
dzmnbinom
gives the (log) probability mass function,
pzmnbinom
gives the (log) distribution function,
qzmnbinom
gives the quantile function, and
rzmnbinom
generates random deviates.
Invalid size
, prob
or p0
will result in return
value NaN
, with a warning.
The length of the result is determined by n
for
rzmnbinom
, and is the maximum of the lengths of the
numerical arguments for the other functions.
Functions {d,p,q}zmnbinom
use {d,p,q}nbinom
for all
but the trivial input values and .
Vincent Goulet [email protected]
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dnbinom
for the negative binomial distribution.
dztnbinom
for the zero-truncated negative binomial
distribution.
dzmgeom
for the zero-modified geometric and
dzmlogarithmic
for the zero-modified logarithmic, which
are special cases of the zero-modified negative binomial.
## Example 6.3 of Klugman et al. (2012) p <- 1/(1 + 0.5) dzmnbinom(1:5, size = 2.5, prob = p, p0 = 0.6) (1-0.6) * dnbinom(1:5, 2.5, p)/pnbinom(0, 2.5, p, lower = FALSE) # same ## simple relation between survival functions pzmnbinom(0:5, 2.5, p, p0 = 0.2, lower = FALSE) (1-0.2) * pnbinom(0:5, 2.5, p, lower = FALSE) / pnbinom(0, 2.5, p, lower = FALSE) # same qzmnbinom(pzmnbinom(0:10, 2.5, 0.3, p0 = 0.1), 2.5, 0.3, p0 = 0.1)
## Example 6.3 of Klugman et al. (2012) p <- 1/(1 + 0.5) dzmnbinom(1:5, size = 2.5, prob = p, p0 = 0.6) (1-0.6) * dnbinom(1:5, 2.5, p)/pnbinom(0, 2.5, p, lower = FALSE) # same ## simple relation between survival functions pzmnbinom(0:5, 2.5, p, p0 = 0.2, lower = FALSE) (1-0.2) * pnbinom(0:5, 2.5, p, lower = FALSE) / pnbinom(0, 2.5, p, lower = FALSE) # same qzmnbinom(pzmnbinom(0:10, 2.5, 0.3, p0 = 0.1), 2.5, 0.3, p0 = 0.1)
Density function, distribution function, quantile function, random
generation for the Zero-Modified Poisson distribution with parameter
lambda
and arbitrary probability at zero p0
.
dzmpois(x, lambda, p0, log = FALSE) pzmpois(q, lambda, p0, lower.tail = TRUE, log.p = FALSE) qzmpois(p, lambda, p0, lower.tail = TRUE, log.p = FALSE) rzmpois(n, lambda, p0)
dzmpois(x, lambda, p0, log = FALSE) pzmpois(q, lambda, p0, lower.tail = TRUE, log.p = FALSE) qzmpois(p, lambda, p0, lower.tail = TRUE, log.p = FALSE) rzmpois(n, lambda, p0)
x |
vector of (strictly positive integer) quantiles. |
q |
vector of quantiles. |
p |
vector of probabilities. |
n |
number of values to return. |
lambda |
vector of (non negative) means. |
p0 |
probability mass at zero. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
The zero-modified Poisson distribution is a discrete mixture between a
degenerate distribution at zero and a (standard) Poisson. The
probability mass function is and
for ,
and
, where
is the probability mass
function of the Poisson.
The cumulative distribution function is
The mean is and the variance is
,
where
and
are the mean and variance of
the zero-truncated Poisson.
In the terminology of Klugman et al. (2012), the zero-modified
Poisson is a member of the class of distributions
with
and
.
The special case p0 == 0
is the zero-truncated Poisson.
If an element of x
is not integer, the result of
dzmpois
is zero, with a warning.
The quantile is defined as the smallest value such that
, where
is the distribution function.
dzmpois
gives the (log) probability mass function,
pzmpois
gives the (log) distribution function,
qzmpois
gives the quantile function, and
rzmpois
generates random deviates.
Invalid lambda
or p0
will result in return value
NaN
, with a warning.
The length of the result is determined by n
for
rzmpois
, and is the maximum of the lengths of the
numerical arguments for the other functions.
Functions {d,p,q}zmpois
use {d,p,q}pois
for all
but the trivial input values and .
Vincent Goulet [email protected]
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dpois
for the standard Poisson distribution.
dztpois
for the zero-truncated Poisson distribution.
dzmpois(0:5, lambda = 1, p0 = 0.2) (1-0.2) * dpois(0:5, lambda = 1)/ppois(0, 1, lower = FALSE) # same ## simple relation between survival functions pzmpois(0:5, 1, p0 = 0.2, lower = FALSE) (1-0.2) * ppois(0:5, 1, lower = FALSE) / ppois(0, 1, lower = FALSE) # same qzmpois(pzmpois(0:10, 1, p0 = 0.7), 1, p0 = 0.7)
dzmpois(0:5, lambda = 1, p0 = 0.2) (1-0.2) * dpois(0:5, lambda = 1)/ppois(0, 1, lower = FALSE) # same ## simple relation between survival functions pzmpois(0:5, 1, p0 = 0.2, lower = FALSE) (1-0.2) * ppois(0:5, 1, lower = FALSE) / ppois(0, 1, lower = FALSE) # same qzmpois(pzmpois(0:10, 1, p0 = 0.7), 1, p0 = 0.7)
Density function, distribution function, quantile function and random
generation for the Zero-Truncated Binomial distribution with
parameters size
and prob
.
dztbinom(x, size, prob, log = FALSE) pztbinom(q, size, prob, lower.tail = TRUE, log.p = FALSE) qztbinom(p, size, prob, lower.tail = TRUE, log.p = FALSE) rztbinom(n, size, prob)
dztbinom(x, size, prob, log = FALSE) pztbinom(q, size, prob, lower.tail = TRUE, log.p = FALSE) qztbinom(p, size, prob, lower.tail = TRUE, log.p = FALSE) rztbinom(n, size, prob)
x |
vector of (strictly positive integer) quantiles. |
q |
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
size |
number of trials (strictly positive integer). |
prob |
probability of success on each trial. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
The zero-truncated binomial distribution with size
and
prob
has probability mass function
for and
, and
when
.
The cumulative distribution function is
where is the distribution function of the standard binomial.
The mean is and the variance is
.
In the terminology of Klugman et al. (2012), the zero-truncated
binomial is a member of the class of
distributions with
and
.
If an element of x
is not integer, the result of
dztbinom
is zero, with a warning.
The quantile is defined as the smallest value such that
, where
is the distribution function.
dztbinom
gives the probability mass function,
pztbinom
gives the distribution function,
qztbinom
gives the quantile function, and
rztbinom
generates random deviates.
Invalid size
or prob
will result in return value
NaN
, with a warning.
The length of the result is determined by n
for
rztbinom
, and is the maximum of the lengths of the
numerical arguments for the other functions.
Functions {d,p,q}ztbinom
use {d,p,q}binom
for all
but the trivial input values and .
rztbinom
uses the simple inversion algorithm suggested by
Peter Dalgaard on the r-help mailing list on 1 May 2005
(https://stat.ethz.ch/pipermail/r-help/2005-May/070680.html).
Vincent Goulet [email protected]
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dbinom
for the binomial distribution.
dztbinom(1:5, size = 5, prob = 0.4) dbinom(1:5, 5, 0.4)/pbinom(0, 5, 0.4, lower = FALSE) # same pztbinom(1, 2, prob = 0) # point mass at 1 qztbinom(pztbinom(1:10, 10, 0.6), 10, 0.6) n <- 8; p <- 0.3 x <- 0:n title <- paste("ZT Binomial(", n, ", ", p, ") and Binomial(", n, ", ", p,") PDF", sep = "") plot(x, dztbinom(x, n, p), type = "h", lwd = 2, ylab = "p(x)", main = title) points(x, dbinom(x, n, p), pch = 19, col = "red") legend("topright", c("ZT binomial probabilities", "Binomial probabilities"), col = c("black", "red"), lty = c(1, 0), lwd = 2, pch = c(NA, 19))
dztbinom(1:5, size = 5, prob = 0.4) dbinom(1:5, 5, 0.4)/pbinom(0, 5, 0.4, lower = FALSE) # same pztbinom(1, 2, prob = 0) # point mass at 1 qztbinom(pztbinom(1:10, 10, 0.6), 10, 0.6) n <- 8; p <- 0.3 x <- 0:n title <- paste("ZT Binomial(", n, ", ", p, ") and Binomial(", n, ", ", p,") PDF", sep = "") plot(x, dztbinom(x, n, p), type = "h", lwd = 2, ylab = "p(x)", main = title) points(x, dbinom(x, n, p), pch = 19, col = "red") legend("topright", c("ZT binomial probabilities", "Binomial probabilities"), col = c("black", "red"), lty = c(1, 0), lwd = 2, pch = c(NA, 19))
Density function, distribution function, quantile function and random
generation for the Zero-Truncated Geometric distribution with
parameter prob
.
dztgeom(x, prob, log = FALSE) pztgeom(q, prob, lower.tail = TRUE, log.p = FALSE) qztgeom(p, prob, lower.tail = TRUE, log.p = FALSE) rztgeom(n, prob)
dztgeom(x, prob, log = FALSE) pztgeom(q, prob, lower.tail = TRUE, log.p = FALSE) qztgeom(p, prob, lower.tail = TRUE, log.p = FALSE) rztgeom(n, prob)
x |
vector of (strictly positive integer) quantiles. |
q |
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
prob |
parameter. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
The zero-truncated geometric distribution with prob
has probability mass function
for and
, and
when
.
The cumulative distribution function is
where is the distribution function of the standard geometric.
The mean is and the variance is
.
In the terminology of Klugman et al. (2012), the zero-truncated
geometric is a member of the class of
distributions with
and
.
If an element of x
is not integer, the result of
dztgeom
is zero, with a warning.
The quantile is defined as the smallest value such that
, where
is the distribution function.
dztgeom
gives the (log) probability mass function,
pztgeom
gives the (log) distribution function,
qztgeom
gives the quantile function, and
rztgeom
generates random deviates.
Invalid prob
will result in return value NaN
, with a
warning.
The length of the result is determined by n
for
rztgeom
, and is the maximum of the lengths of the
numerical arguments for the other functions.
Functions {d,p,q}ztgeom
use {d,p,q}geom
for all but
the trivial input values and .
rztgeom
uses the simple inversion algorithm suggested by
Peter Dalgaard on the r-help mailing list on 1 May 2005
(https://stat.ethz.ch/pipermail/r-help/2005-May/070680.html).
Vincent Goulet [email protected]
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dgeom
for the geometric distribution.
dztnbinom
for the zero-truncated negative binomial, of
which the zero-truncated geometric is a special case.
p <- 1/(1 + 0.5) dztgeom(c(1, 2, 3), prob = p) dgeom(c(1, 2, 3), p)/pgeom(0, p, lower = FALSE) # same dgeom(c(1, 2, 3) - 1, p) # same pztgeom(1, prob = 1) # point mass at 1 qztgeom(pztgeom(1:10, 0.3), 0.3)
p <- 1/(1 + 0.5) dztgeom(c(1, 2, 3), prob = p) dgeom(c(1, 2, 3), p)/pgeom(0, p, lower = FALSE) # same dgeom(c(1, 2, 3) - 1, p) # same pztgeom(1, prob = 1) # point mass at 1 qztgeom(pztgeom(1:10, 0.3), 0.3)
Density function, distribution function, quantile function and random
generation for the Zero-Truncated Negative Binomial distribution with
parameters size
and prob
.
dztnbinom(x, size, prob, log = FALSE) pztnbinom(q, size, prob, lower.tail = TRUE, log.p = FALSE) qztnbinom(p, size, prob, lower.tail = TRUE, log.p = FALSE) rztnbinom(n, size, prob)
dztnbinom(x, size, prob, log = FALSE) pztnbinom(q, size, prob, lower.tail = TRUE, log.p = FALSE) qztnbinom(p, size, prob, lower.tail = TRUE, log.p = FALSE) rztnbinom(n, size, prob)
x |
vector of (strictly positive integer) quantiles. |
q |
vector of quantiles. |
p |
vector of probabilities. |
n |
number of observations. If |
size |
target for number of successful trials, or dispersion parameter. Must be positive, need not be integer. |
prob |
parameter. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
The zero-truncated negative binomial distribution with size
and
prob
has probability mass function
for ,
and
, and
when
.
The cumulative distribution function is
where is the distribution function of the standard negative
binomial.
The mean is and the variance is
.
In the terminology of Klugman et al. (2012), the zero-truncated
negative binomial is a member of the class of
distributions with
and
.
The limiting case size == 0
is the
logarithmic distribution with parameter 1 -
prob
.
Unlike the standard negative binomial functions, parametrization
through the mean mu
is not supported to avoid ambiguity as
to whether mu
is the mean of the underlying negative binomial
or the mean of the zero-truncated distribution.
If an element of x
is not integer, the result of
dztnbinom
is zero, with a warning.
The quantile is defined as the smallest value such that
, where
is the distribution function.
dztnbinom
gives the (log) probability mass function,
pztnbinom
gives the (log) distribution function,
qztnbinom
gives the quantile function, and
rztnbinom
generates random deviates.
Invalid size
or prob
will result in return value
NaN
, with a warning.
The length of the result is determined by n
for
rztnbinom
, and is the maximum of the lengths of the
numerical arguments for the other functions.
Functions {d,p,q}ztnbinom
use {d,p,q}nbinom
for all
but the trivial input values and .
rztnbinom
uses the simple inversion algorithm suggested by
Peter Dalgaard on the r-help mailing list on 1 May 2005
(https://stat.ethz.ch/pipermail/r-help/2005-May/070680.html).
Vincent Goulet [email protected]
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dnbinom
for the negative binomial distribution.
dztgeom
for the zero-truncated geometric and
dlogarithmic
for the logarithmic, which are special
cases of the zero-truncated negative binomial.
## Example 6.3 of Klugman et al. (2012) p <- 1/(1 + 0.5) dztnbinom(c(1, 2, 3), size = 2.5, prob = p) dnbinom(c(1, 2, 3), 2.5, p)/pnbinom(0, 2.5, p, lower = FALSE) # same pztnbinom(1, 2, prob = 1) # point mass at 1 dztnbinom(2, size = 1, 0.25) # == dztgeom(2, 0.25) dztnbinom(2, size = 0, 0.25) # == dlogarithmic(2, 0.75) qztnbinom(pztnbinom(1:10, 2.5, 0.3), 2.5, 0.3) x <- rztnbinom(1000, size = 2.5, prob = 0.4) y <- sort(unique(x)) plot(y, table(x)/length(x), type = "h", lwd = 2, pch = 19, col = "black", xlab = "x", ylab = "p(x)", main = "Empirical vs theoretical probabilities") points(y, dztnbinom(y, size = 2.5, prob = 0.4), pch = 19, col = "red") legend("topright", c("empirical", "theoretical"), lty = c(1, NA), lwd = 2, pch = c(NA, 19), col = c("black", "red"))
## Example 6.3 of Klugman et al. (2012) p <- 1/(1 + 0.5) dztnbinom(c(1, 2, 3), size = 2.5, prob = p) dnbinom(c(1, 2, 3), 2.5, p)/pnbinom(0, 2.5, p, lower = FALSE) # same pztnbinom(1, 2, prob = 1) # point mass at 1 dztnbinom(2, size = 1, 0.25) # == dztgeom(2, 0.25) dztnbinom(2, size = 0, 0.25) # == dlogarithmic(2, 0.75) qztnbinom(pztnbinom(1:10, 2.5, 0.3), 2.5, 0.3) x <- rztnbinom(1000, size = 2.5, prob = 0.4) y <- sort(unique(x)) plot(y, table(x)/length(x), type = "h", lwd = 2, pch = 19, col = "black", xlab = "x", ylab = "p(x)", main = "Empirical vs theoretical probabilities") points(y, dztnbinom(y, size = 2.5, prob = 0.4), pch = 19, col = "red") legend("topright", c("empirical", "theoretical"), lty = c(1, NA), lwd = 2, pch = c(NA, 19), col = c("black", "red"))
Density function, distribution function, quantile function, random
generation for the Zero-Truncated Poisson distribution with parameter
lambda
.
dztpois(x, lambda, log = FALSE) pztpois(q, lambda, lower.tail = TRUE, log.p = FALSE) qztpois(p, lambda, lower.tail = TRUE, log.p = FALSE) rztpois(n, lambda)
dztpois(x, lambda, log = FALSE) pztpois(q, lambda, lower.tail = TRUE, log.p = FALSE) qztpois(p, lambda, lower.tail = TRUE, log.p = FALSE) rztpois(n, lambda)
x |
vector of (strictly positive integer) quantiles. |
q |
vector of quantiles. |
p |
vector of probabilities. |
n |
number of values to return. |
lambda |
vector of (non negative) means. |
log , log.p
|
logical; if |
lower.tail |
logical; if |
The zero-truncated Poisson distribution has probability mass function
for , and
when
.
The cumulative distribution function is
where is the distribution function of the standard Poisson.
The mean is and the variance is
.
In the terminology of Klugman et al. (2012), the zero-truncated
Poisson is a member of the class of distributions
with
and
.
If an element of x
is not integer, the result of
dztpois
is zero, with a warning.
The quantile is defined as the smallest value such that
, where
is the distribution function.
dztpois
gives the (log) probability mass function,
pztpois
gives the (log) distribution function,
qztpois
gives the quantile function, and
rztpois
generates random deviates.
Invalid lambda
will result in return value NaN
, with a
warning.
The length of the result is determined by n
for
rztpois
, and is the maximum of the lengths of the
numerical arguments for the other functions.
Functions {d,p,q}ztpois
use {d,p,q}pois
for all
but the trivial input values and .
rztpois
uses the simple inversion algorithm suggested by
Peter Dalgaard on the r-help mailing list on 1 May 2005
(https://stat.ethz.ch/pipermail/r-help/2005-May/070680.html).
Vincent Goulet [email protected]
Klugman, S. A., Panjer, H. H. and Willmot, G. E. (2012), Loss Models, From Data to Decisions, Fourth Edition, Wiley.
dpois
for the standard Poisson distribution.
dztpois(1:5, lambda = 1) dpois(1:5, lambda = 1)/ppois(0, 1, lower = FALSE) # same pztpois(1, lambda = 0) # point mass at 1 qztpois(pztpois(1:10, 1), 1) x <- seq(0, 8) plot(x, dztpois(x, 2), type = "h", lwd = 2, ylab = "p(x)", main = "Zero-Truncated Poisson(2) and Poisson(2) PDF") points(x, dpois(x, 2), pch = 19, col = "red") legend("topright", c("ZT Poisson probabilities", "Poisson probabilities"), col = c("black", "red"), lty = c(1, 0), lwd = 2, pch = c(NA, 19))
dztpois(1:5, lambda = 1) dpois(1:5, lambda = 1)/ppois(0, 1, lower = FALSE) # same pztpois(1, lambda = 0) # point mass at 1 qztpois(pztpois(1:10, 1), 1) x <- seq(0, 8) plot(x, dztpois(x, 2), type = "h", lwd = 2, ylab = "p(x)", main = "Zero-Truncated Poisson(2) and Poisson(2) PDF") points(x, dpois(x, 2), pch = 19, col = "red") legend("topright", c("ZT Poisson probabilities", "Poisson probabilities"), col = c("black", "red"), lty = c(1, 0), lwd = 2, pch = c(NA, 19))