Download cumulative distribution function distribution function cdf Exam C

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts
no text concepts found
Transcript
cumulative distribution function
distribution function
cdf
survival function
Exam C
probability density function
density function
Exam C
probability function
probability mass function
Exam C
hazard rate
force of mortality
failure rate
Exam C
truncated
Exam C
censored
Exam C
k-th raw moment
Exam C
empirical model
Exam C
k-th central moment
Exam C
Exam C
SX (x) = P r(X > x) = 1 − FX (x)
Z
x
F (x) =
f (s)ds = P (X < x)
−∞
f (x) = F 0 (x) = −S 0 (x)
pX (x) = Pr(X = x)
An observation is truncated at d if when it is below
d it is below d it is not recorded but when it is above
d it is recorded at itsobserved value.
The kth raw moment of a random variable is the
expected value of the kth power of the variable, provided it exists. It is denoted by E(X k ) or by µ0k . The
first raw moment is called the mean of the random
variable and is usually denoted by µ.
Z ∞
X
0
k
xk f (x)dx =
xkj p(xj )
µk = E(X ) =
Z
−∞
hX (x) =
fX (x)
SX (x)
An observation is censored at u if, when it is above u
it is recorded as being equal to u but when it is below
u it is recorded at its observed value.
j
∞
E(x) =
S(x)dx
0
The kth central moment of a random variable is
the expected value of the kth power deviation of the
variable from its mean. It is denoted by E[(X − µ)k ]
or by µk . The second central moment is usually called
the variance and denoted σ 2 and the square root, σ,
is called the standard deviation.
Z ∞
µk = E[(X − µ)k ] =
(x − µ)k f (x)dx
−∞
X
=
(xj − µ)k p(xj )
j
The empirical model is a discrete distribution based
on a sample size n which assigns probability 1/n to
each data point.
coefficient of variation
skewness
Exam C
kurtosis
Exam C
symmetric distribution, then it has a skewness
Exam C
limited loss variable
right censored variable
Exam C
limited expected value
Exam C
left censored and shifted variable
And E(left censored and shifted variable)
Exam C
left censored and shifted variable
mean residual life function
complete expectation of life
left truncated and shifted variable
Exam C
mean excess loss function
Exam C
percentile
Exam C
Exam C
µ3
σ3
σ
µ
µ3
=0
σ3
µ4
σ4
½
E[X ∧ u]
Z
k
u
E[(X ∧ u) ] =
0
Z
=
X,
u,
Y =X ∧u=
X<u
X≥u
xk f (x) + uk [1 − F (u)]
u
S(x)dx
0
Y = X − d given that X > d
½
0,
X<d
X − d, X ≥ d
Y = (X − d)+ =
E[(X − d)+ ] = E(X) − E(X ∧ d)
Z ∞
=
S(x)dx
d
The 100pth percentile of a random variable is any
value πp such that F (πp −) ≤ p ≤ F (πp ).
eX (d) = e(d) = E(Y ) = E(X − d|X > d)
E(X − d|X > d) =
E(X) − E(X ∧ d)
1 − F (d)
Sk − E(Sk )
Distribution of lim p
k→∞
V ar(Sk )
median
Exam C
Exam C
moment generating function
E(X n ) =
moment generating function
Exam C
probability generating function
Exam C
pk from probability generating function
Exam C
Mean and Variance
using the probability generating function
Exam C
pgf/mgf of
Pn
j=1
Xj
Exam C
parametric distribution
Exam C
scale distribution
Exam C
Exam C
Normal distribution with mean 0 and variance 1.
(n)
E(X n ) = MX (0)
The 50th percentile, π0.5 is called the median.
MX (t) = E(etX )
(m)
P
(0)
pk = X
m!
The mth derivative of the pgf evalated at 0 devided by
m factorial.
MSk (t) =
k
Y
MXj (t)
j=1
PSk (z) =
k
Y
PXj (z)
X
PX (z) = E(z ) =
∞
X
pk z k
k=0
0
E(X) = PX
(1)
00
0
E[X 2 ] = PX
(1) + PX
(1)
00
0
0
Var(X) = PX
(1) + PX
(1) − [PX
(1)]2
j=1
A parametric distribution is a scale distribution if,
when a random variable from that set of distributions
is multiplied by a positive constant, the resulting random variable is also in that set of distributions.
A parametric distribution is a set of distribution
functions, each member of which is determined by
specifying one or more values called parameters. The
number of parameters is fixed and finite.
scale parameter
parametric distribution family
Exam C
k-point mixture
Exam C
variable-component mixture distribution
Exam C
data-dependant distribution
Exam C
Equilibrium Distribution
Exam C
Survival function and hazard rate of Equilibrium
Distribution
Exam C
Survival function based on equilibrium distribution
Exam C
coherent risk measure
Exam C
Value-at-Risk
Exam C
Exam C
A parametric distribution family is a set of parametric distributions that are related in some meaningful way.
For random variables with nonnegative support, a
scale parameter is a parameter for a scale distribution that meets two conditions. First, when a member of the scale distribution is multiplied by a positive
constant, the scale parameter is multiplied by a positive constant, the scale parameter is multiplied by the
same constant. Second, when a member of the scale
distribution is multiplied by a positive constant, all
other parameters are unchanged.
A variable-component mixture distribution has
a distribution function that can be written as
A random variable Y is a k-point mixture of the
random variables X1 , X2 , · · · , Xk if its cdf is given by
F (x) =
K
X
aj Fj (x),
j=1
K
X
aj = 1, aj > 0
j=1
where all aj > 0 and a1 + a2 + · · · + ak = 1.
Assume X is a continuous distribution f (x) with a survival function S(x) and mean E(X). Then the equilibrium distribution is
fe (x) =
S(x) =
S(x)
,
E(X)
FY (y) = a1 FX1 (y) + a2 FX2 (y) + · · · + ak FXk (y)
A data-dependant distribution is at least as complex as the data or knowledge that produced it, and
the number of “parameters” increases as the number
of data points or amount of knowledge increase.
x≥0
1
E(x) − R0x [ e(t)
]dt
e
e(x)
Z
Se (x) =
fe (t)dt =
x
he (x) =
Let X denote a loss random variable. The Value-atRisk of X at the 100p% level, denoted VaRp (X) or πp ,
is the l00p percentile (or quantile) of the distribution
of X. And for continuous distributions it is the value
of πp satisfying
R∞
∞
x
S(t)dt
,
E(x)
x≥0
fe (x)
S(x)
1
= R∞
=
Se (x)
e(x)
S(t)dt
x
1. Subadditivity: ρ(X + Y ) ≤ ρ(X) + ρ(Y ).
2. Monotonicity: If X ≤ Y for all possible outcomes, then ρ(X) ≤ ρ(Y ).
Pr(X > πp ) = 1 − p
3. Positive homogeneity: For any positive constant
c. ρ(cX) = cρ(X).
VaR is not “coherent” as it does not meet the subaddivity requirement in some cases.
4. Translation invariance: For any positive constant c. ρ(X + c) = ρ(X) + c.
Given: F (x) and f (x) find FY (y) when
Y = θX
Tail-Value-at-Risk
Exam C
Given: F (x) and f (x) find FY (y) when
Y = X 1/τ
Exam C
transformed
Exam C
inverse
Exam C
inverse transformed
Exam C
Exam C
Given: F (x) and f (x) find FY (y) when
Y = eX
Given: F (x) and f (x) find FY (y) when
Y = g(X)
Exam C
Exam C
Raw Moments of a Mixture
Mixture Distribution (FX (x))
Exam C
Exam C
FY (y) = FX
³y ´
θ
³y ´
1
fY (y) = fX
θ
θ
Y = X 1/τ
τ >0
Let X denote a loss random variable. The TailValue-at-Risk of X at the 100p% security level, denoted TVaRp (X), is the expected loss given that the
loss exceeds the l00p percentile (or quantile) of the
distribution of X. And for continuous distributions it
can expressed as
R∞
xf (x)dx
π
TVaRp (X) = E(X|X > πp ) = p
1 − F (πp )
FY (y) = FX (y τ )
fY (y) = τ y τ −1 fX (y τ )
Y = X 1/τ
Y = X 1/τ
τ =< 0 and τ 6= −1
τ = −1
h(y) = g −1 (Y )
FY (y) = FX [h(y)]
fY (y) = |h0 (y)|fX [h(y)]
FY (y) = FX [ln(y)]
1
fY (y) = fX [ln(y)]
y
E(X k ) = E[E(X k | Λ)]
Z
FX (x) =
FX|Λ (x | λ)fΛ (λ)dλ
Important Mixtures:
Y = Poison(Λ)
Λ = Gamma(α, θ)
Variance of a Mixture
Exam C
Important Mixtures:
Y = Exponential(Λ)
Λ = Inv.Gamma(α, θ)
Exam C
Important Mixtures:
Y = Inv.Exponential(Λ)
Λ = Gamma(α, θ)
Exam C
Important Mixtures:
Y = Normal(Λ, σc2 )
Λ = Normal(µ, σd2 )
Exam C
k-component spliced distribution
Exam C
linear exponential family
Exam C
normalizing constant
Exam C
canonical parameter
Exam C
linear exponential family: Mean
Exam C
Exam C
Neg.Bin(r = α, β = θ)
Var(X) = E[V ar(X | Λ)] + V ar[E(X | Λ)]
Inv.Pareto(τ = α, θ)
Pareto(α, θ)
A k-component spliced distribution has a density
function that can be expressed as follows:

a1 f1 (x),
c0 < x < c 1



 a2 f2 (x),
c1 < x < c 2
fX (x) =
..
..

.
.



ak fk (x), ck−1 < x < ck
Normal(µ, σc2 + σd2 )
q(θ)
f (x; θ) =
f (x; θ) =
p(x)er(θ)x
q(θ)
E(X) = µ(θ) =
q 0 (θ)
r0 (θ)q(θ)
p(x)er(θ)x
q(θ)
r(θ)
f (x; θ) =
p(x)er(θ)x
q(θ)
linear exponential family: Variance
(a,b,0) class of distribution
Exam C
(a,b,1) class of distribution
Exam C
Poison (a,b,0) specification
Exam C
Binomial (a,b,0) specification
Exam C
Negative Binomial (a,b,0) specification
Exam C
Exam C
Geometric Relation to Negative Binomial
Geometric (a,b,0) specification
Exam C
memoryless
Exam C
E[(1 + r)X ∧ c]
Exam C
Exam C
Let pk be the pf of a discrete random variable. It is
member of the (a,b,0) class of distributions, provided that there exists constants a and b such that
pk
pk−1
=a+
b
k
Var(X) =
µ0 (θ)
r0 (θ)
k = 1, 2, 3, · · ·
a=0
b=λ
p0 = e−λ
Let pk be the pf of a discrete random variable. It is
member of the (a,b,0) class of distributions, provided that there exists constants a and b such that
pk
pk−1
=a+
b
k
k = 2, 3, 4, · · · p0
=1−
∞
X
pk
k=1
It is called truncated if p0 = 0.
It is called zero-modified if p0 > 0 and is a mixture
of an (a,b,0) class and a distribution where p0 = 1.
(AKA truncated with zeros)
a=
β
1+β
β
1+β
p0 = (1 + β)−r
b = (r − 1)
a=−
q
1−q
b = (m + 1)
q
1−q
p0 = (1 − q)m
Geomtric is Negative Binomial with parameter r = 1.
β
1+β
b=0
a=
p0 = (1 + β)−1
E[(1 + r)X ∧ c] = (1 + r)E [X ∧ c∗ ]
c
c∗ =
1+r
P (X > x + y|X > x) = P (Y > y)
The Geometric and Exponential distributions are
both examples of a memoryless distribution.
per-loss variable
per-payment variable
Exam C
Relationship between per-loss and per-payment
Exam C
Relationship between per-loss and per-payment
Variance
Exam C
franchise deductible
Expectation
Exam C
ordinary deductible
Expectation
Exam C
Exam C
Co-insurance
The loss elimination ratio
Exam C
Co-insurance, deductible and limits variable
Exam C
Exam C
Co-insurance, deductible and limits variable
Expectation
Exam C
½
YP =
undef ined,
X − d,
½
X≤d
X>d
¡
¢ µ
¶2
E [Y L ]2
E(Y L )
−
Var(Y ) =
SX (d)
SX (d)
P
YL =
0,
X≤d
X − d, X > d
If f meets certian conditions [not sure how to define
them, (linear?)] then
f (Y P ) =
f (Y L )
SX (d)
E(Y P ) =
E(Y L )
SX (d)
However,
Var(Y P ) 6=
E(X) − E(X ∧ d)
Var(Y L )
SX (d)
A franchise deductible modifies the ordinary deductible by adding the deductible whenever there is a
positive amount paid. For a franchise deductible the
expected cost per loss is
E(X) − E(X ∧ d) + d[1 − F (d)]
If co-insruance is the only modification, this changes
the loss variable X to the payment variable Y = αX.
E(X) − [E(X) − E(X ∧ d)]
E(X ∧ d)
=
E(X)
E(X)
E(Y L ) = α(1 + r) [E (X ∧ u∗ ) − E (X ∧ d∗ )]
Y
L

 0,
α[(1 + r)X − d], d∗
=

α(u − d),
u∗
X
≤ X
≤ X
< d∗
< u∗
Co-insurance, deductible and limits variable
2nd Raw Moment
individual risk model
Exam C
collective risk model
OR
Compound Model
Exam C
Compound Model
Mean
Exam C
Compound Model
Variance
Exam C
Compound Model
N = Poisson(λ)
Exam C
Exam C
P r(a < S < b) = 0. Then, for a ≤ d ≤ b
E[(S − d)+ ] =
stop-loss insurance
Exam C
Theorem to calculate E[(S − d)+ ] using a discrete
probability function w(P (S = kh) = fk ≥ 0) with
equally spaced (h) nodes.
Exam C
Exam C
Convolution method (X1 + X2 )
Exam C
The individual risk model represents the aggregate
loss as a sum, S = X1 + . . . + Xn , of a fixed number, n, of insurance contracts. The loss amounts for
the n contracts are (X1 , . . . , Xn ), where the Xj s are
assumed to be independent but are not assumed to
be identically distributed. The distribution of the Xj s
usually has a probability mass at zero, corresponding
to the probability of no loss or payment
E[(Y L )2 ] =
= α2 (1 + r)2 {E[(X ∧ u∗ )2 ] − E[(X ∧ d∗ )2 ]
− 2d∗ E(X ∧ u∗ ) + 2d∗ E(X ∧ d∗ )}
S = X1 + . . . + XN with:
E[S] = E[E(S|N )]
= E(N )E(X) = µn µx
1. Conditional on N = n, the random variables
X1 , X2 , . . . , Xn are i.i.d random variables.
2. Conditional on N = n, the common distribution
of the random variables X1 , X2 , . . . , Xn , does not
depend on n.
3. The distribution of N does not depend in any
way on the values of X1 , X2 , . . .
E[S] = µn µx = λµx
V ar[S] =
λ(σx2
+
µ2x )
(N=Poisson)
2
= λE(X )
(N=Poisson)
V ar[S] = V ar[E(S|N )] + E[V ar(S|N )]
= Var(N )E(X)2 + E(N )Var(X)
= σn2 µ2x + µn σx2
b−d
E[(S − d)+ ] =
E[(S − a)+ ]
b−a
d−a
+
E[(S − b)+ ]
b−a
FS (k) =
X
fX1 (j)FX2 (k − j)
Insurance on the aggregate losses, subject to a deductible, is called stop-loss insurance. The expected
cost of this insurance is called the net stop-loss premium and can be computed as E[(S − d)+ ], where d
is the deductible and the notation (.)+ means to use
the value in parenthesis if it is positive but to use zero
otherwise
k = 0, 1, . . .. Then, provided d = jh, with j a nonnegative integer
all j
Z
fS (s) =
fX1 (t)fX2 (s − t)dt
Z
FS (s) =
fX1 (t)FX2 (s − t)dt
E[(S − d)+ ] = h
∞
X
{1 − FS [(m + j)h]}
m=0
E{[S − (j + 1)h]+ } = E[(S − jh)+ ] − h[1 − FS (jh)]
Convolution method (X1 + . . . + Xn )
Bias
Exam C
Exam C
consistent
asymptotically unbiased
Exam C
Exam C
uniformly minimum variance unbiased estimator
UMVUE
mean-squared error (MSE)
Exam C
confidence interval
Exam C
significance level
Exam C
uniformly most powerful
Exam C
p-value
Exam C
Exam C
biasθ̂ = E(θ̂ | θ) − θ
Z
∗n
FX
(x) =
Z
∗n
fX
(x) =
lim P r(|θ̂n − θ| > σ) = 0
n→∞
An estimator, θ̂, is called a uniformly minimum
variance unbiased estimator (UMVUE) if it is
unbiased and for any true value of θ there is no other
unbiased estimator that has a smaller variance.
x
0
∗(n−1)
FX
x
0
∗(n−1)
fX
(x − t)fX (t)dt
(x − t)fX (t)dt
lim E(θ̂n | θ) = θ
n→∞
M SEθ̂ (θ) = E[(θ̂ − θ)2 |θ]
= V ar(θ̂|θ) + [biasθ̂ (θ)]2
The significance levelof a hypothesis test is the probability of making a Type I error given that the null
hypothesis is true. If it can be true in more than one
way, the level of significance is the maximum of such
probabilities. The significance level is usually denoted
by the letter α.
A 100(1 − α)% confidence interval for a parameter
θ is a pair of random values, L and U , computed from
a random sample such that P r(L ≤ θ ≤ U ) ≥ 1 − α
for all θ.
The p-value is the smallest level of significance at
which H0 would be rejected when a specified test procedure is used on a given data set. Once the p-value
has been determined the conclusion at any particular
level α results from computing the p-value to α:
A hypothesis test is uniformly most powerful if no
other test exists that has the same or lower significance
level and for a particular value within the alternative
hypothesis has a smaller probability of making a Type
II error.
1. p-value ≤ α ⇒ reject H0 at level α.
2. p-value > α ⇒ do not reject H0 at level α.
[Probability and Statistics, Devore, 2000]
Log-Transformed Confidence Interval
empirical distribution
Fn (x) =
Sn (x)
Exam C
kernel smoothed distribution
Exam C
data set: variables
Exam C
data summary
Exam C
E[Sn (x)]
Exam C
Exam C
Sample Variance
Var[Sn (x)]
Exam C
Exam C
Empirical estimate of E[(X ∧ u)k ]
Empirical estimate of the variance
Exam C
Exam C
The empirical distribution is obtained by assigning
probability 1/n to each data point.
number of observations ≤ x
n
Sn (x) = 1 − Fn (x)
Fn (x) =
The 100(1 − α)% log-transformed confidence interval for Sn (t) is
³
´
Sn (t)U , Sn (t)(1/U )
where
q


d n (t)]
zα/2 Var[S

U = exp 
Sn (t) ln Sn (t)
A kernel smoothed distribution is obtained by replacing each data point with a continuous random
variable and then assigning probability 1/n to each
such random variable. The random variable used must
be identical except for a location or scale change that
is related to its associated data point.
1. n - insureds
2. di - entry time
3. xi - death time
4. ui - censored time
Y = number greater than x in sample
E[Sn (x)] = S(x) =
Y
n
1. m - death points
2. yj - death point time
3. sj - deaths @ time yj
4. rj - number @ risk @ time yj
n
1 X
(xi − x̄)2
n − 1 i=1


1 X k
xi − uk · [number of xi ’s > u]
n
xi ≤u
Y = number greater than x in sample
Y
n
S(x)[1 − S(x)]
E[Sn (x)] =
n
S(x) =
n
1X
(xi − x̄)2
n i=1
cumulative hazard rate function
Nelson-Åalen estimate
Exam C
variance of the Nelson-Åalen estimate
Exam C
Kaplan-Meier product-limit Estimator
Exam C
Greenwood Approximation
Variance of the Kaplan-Meier product-limit
Estimator
Exam C
log-transformed interval for the Nelson-Åalen
estimate
Exam C
method of moments
Exam C
percentile matching
Exam C
Exam C
likelihood function
log-likelihood function
smoothed empirical estimate
Exam C
Exam C


x < y1
 0,
Pj−1 si
Ĥ(x) =
i=1 ri , yj−1 ≤ x < yj , j = 2, . . . , k

 Pk si , x ≥ y
i
k
i=1 ri
Sn (t) =

1,


³
 Q



H(x) = − ln S(x)
d Ĥ(yj )] =
Var[
0 ≤ t < y1
´
j−1 ri −si
i=1
ri
Qk ³ ri −si ´
i=1
ri
j
X
si
r2
i=1 i
yj−1 ≤ t < yj , j = 2, . . . , k
t ≥ yk
j
= Ĥ(t)U,

where
q

d Ĥ(yj )]
zα/2 Var[

U = exp ±
Ĥ(t)
πgk (θ) = π̂gk , k = 1, 2, . . . , p
X
.
2
d n (yj )] =
Var[S
[Sn (yj )]
i=1
si
ri (ri − si )
µ0k (θ) = µ̂0k , k = 1, 2, . . . , p
F (π̂gk | θ) = gk , k = 1, 2, . . . , p
Here b.c indicates the greatest integer function and
x(1) ≤ x(2) ≤ . . . ≤ x(n) are the order statistics from
the sample.
L(θ) =
n
Y
Pr(Xj ∈ Aj | θ)
j=1
l(θ) = ln L(θ)
π̂g = (1 − h)x(j) + hx(j+1) , where
j = b(n + 1)gc and h = (n + 1)g − j
Information function
Variance of the likelihood estimate (θ)
Exam C
Delta Method (single variable)
Exam C
Delta Method (general)
Exam C
Non-normal confidence interval
Exam C
p-p plot
Exam C
Exam C
Kolmogorov-Smirnov Test
D(x) plot
Exam C
Anderson-Darling test
Exam C
Chi-Square Goodness-of-fit
Exam C
Exam C
−1
·
¸
∂2
I(θ) = −E
ln L(θ)
∂θ2
"µ
¶2 #
∂2
ln L(θ)
=E
∂θ2
Z
∂2
= −n f (x; θ) 2 ln[f (x; θ)]dx
∂θ
Var(θ̂) = [I(θ)]
Let θ̂ n = (θ̂1n , . . . , θ̂kn )T be a multivariate parameter
vector of dimension k based on a sample size of n.
Assume that θ̂ n is asymptotically normal with mean
of θ̂ and covariance matrix Ω/n. Then g(θ1 , . . . , θk ) is
asymptotically normal with
Let θ̂n be a parameter estimated using a sample size
of n. Assume that θ̂n is asymptotically normal with
mean of µ and variance of σ 2 /n. Then g(θ̂n ) is asymptotically normal with
E[g(θ)] = g(θ̂n ))
E[g(θ)] = g(θ̂ n )
Var[g(θ)] = (∂g)T
Var[g(θ))] = [g 0 (θ̂n )]2
Ω
∂g
n
(
(Fn (xj ), F ∗ (xj )), where
j
Fn (xj ) =
n+1
θ : l(θ) ≥ l(θ̂) −
χ2α/2
)
2
where the first term is the loglikelihood value at the
maximum likelihood estimate and the second term is
the 1 − α percentile from the chi-square distribution
with degrees of freedom equal to the number of estimated parameters.
D = max |Fn (x) − F ∗ (x)|
D(x) = Fn (x) − F ∗ (x), where
j
Fn (xj ) =
n
t≤x≤u
Z
k
X
n(p̂j − pnj )2
χ =
p̂j
j=1
2
u
[Fn (x) − F ∗ (x)]2 ∗
f (x)dx
∗
∗
t F (x)[1 − F (x)]
= −nF ∗ (u)
A =n
2
k
X
(Ej − Oj )2
=
Ej
j=1
σ2
n
+n
k
X
[1 − Fn (yj )]2 (ln[1 − F ∗ (yj )] − ln[1 − F ∗ (yj+1 )])
j=0
The critical values for this test comes from the chisquare distribution with degrees of freedom of (k − 1 −
r). Where k is the number of terms in the sum and r
is the number of estimated parameters values.
+n
k
X
j=0
Fn (yj )2 [ln F ∗ (yj+1 ) − ln F ∗ (yj )]
likelihood ratio test
Schwarz Bayesian Criterion
Exam C
Full credibility
(Single Variable Case)
Exam C
Full credibility
(Poisson)
Exam C
Exam C
Full Credibility
Compound Distributions
n
X
Si
Full Credibility
Compound Distributions
# of Si ’s
i=1
Exam C
Full Credibility
Compound Distributions
n
X
Si
µ
i=1 y
Exam C
Full Credibility
Compound Distributions
N = P oisson(λ)
# of Si ’s
Exam C
Exam C
Full Credibility
Compound Distributions
N = P oisson(λ)
n
X
Si
Full Credibility
Compound Distributions
N = P oisson(λ)
n
X
Si
µ
i=1 y
i=1
Exam C
Exam C
Recommends that when ranking models a deduction of
(r/2) ln n should be made from the loglikelihood value,
where r is the number of estimated parameters and n
is the sample size.
L0 = L(θ 0 )
L1 = L(θ 1 )
µ ¶
L1
T = 2 ln
= 2 (ln L1 − ln L0 )
L0
The critical values come from a chi-square distribution
with degrees of freedom equal to the number of free
parameters in L1 less the number of free parameters
in L0 .
n ≥ n0
1
λ
W n ≥ n0
µ ¶2
σ
V ar[W ]
2
= n0
= n0 CVW
µ
(E[W ])2
V ar[W ]
σ2
2
= n0
= n0 E[W ]CVW
W n ≥ n0
µ
E[W ]
n ≥ n0
σn2 µ2y + µn σy2
µn µy
n0
σn2 µ2y + µn σy2
(µn µy )2
"
#
σy2
1
1+ 2
n0
λ
µy
n0
σn2 µ2y + µn σy2
µn µ2y
n0
"
n0
σy2
1+ 2
µy
#
"
n0
σy2
µy +
µy
#
Partial Credibility
Model distribution
Exam C
Joint distribution
Exam C
Prior distribution
Exam C
marginal distribution
Exam C
posterior distribution
Exam C
Exam C
Bayes Premium
E(xn+1 )
predictive distribution
Exam C
Exam C
Bayes Premium
E(xn+1 )
with: E[X|θ] = θ
(eg. X Poisson)
Z
∞
xa e−cx dx
0
Exam C
Exam C
The model distribution (mx|θ ) is the probability
distribution for the data as collected given a particular value for the parameter. Its pdf is denoted
mx|θ = fX|Θ (x|θ).
mx|θ = fX|Θ (x|θ) =
n
Y
fX|Θ (xi |θ)
Q = ZW + (1 − Z)M Where,
Z=
Ãs
!
information available
min
,1
information required for full credibility
i=0
The prior distribution (πθ ) is a probability over
the space of possible parameter values. It is denoted
π(θ) and represents our opinion concerning the relative chances that various values of θ are the true value
of the parameter.
The joint distribution (jx,θ ) has pdf
The posterior distribution (pθ|x ) is the conditional
probability distribution of the parameters given the
observed data. Its pdf is
The marginal distribution (gx ) of x has pdf
Z
Z
gx = fX (x) = jx,θ dθ = fX|Θ (x|θ)π(θ)dθ
jx,θ = fX,Θ (x, θ) = mx|θ π(θ) = fX|Θ (x|θ)π(θ)
jx,θ
gx
fX|Θ (x|θ)π(θ)
=R
fX|Θ (x|θ)π(θ)dθ
pθ|x = πΘ|X (θ|x) =
Z
E(xn+1 |x) =
E(x|θ)pθ|x dθ
The predictive distribution is the conditional probability distribution of a new observation y given the
data x = x1 , . . . , xn . Its pdf
Z
fY |X (y|x) = fY |Θ (y|θ)pθ|x dθ
Z
= fY |Θ (y|θ)πΘ|x (θ|x)dθ
Z
Z
∞
0
Γ(a + 1)
ca+1
a!
= a+1 ‘a’ is an integer
c
xa e−cx dx =
E(xn+1 |x) =
θpθ|x dθ
or the mean of the posterior distribution.
Z
0
∞
Conjugate Prior
Poisson-Gamma
λ = gamma(α, θ)
x = Poisson(λ)
−c
x
e
dx
xk
Exam C
Conjugate Prior
Exponential-Inverse Gamma
λ = gamma−1 (α, θ)
x = exp(λ)
Exam C
Conjugate Prior
Binomial-Beta
q = beta(a, b, 1)
x = bin(m, q)
Exam C
Conjugate Prior
Inverse Exponential-Gamma
λ = gamma(α, θ)
x = exp−1 (λ)
Exam C
Conjugate Prior
Normal-Normal
λ = normal(µ, a2 )
x = normal(λ, σ 2 )
Exam C
Conjugate Prior
Uniform-Pareto
λ = single.pareto(α, θ)
x = uniform(0, λ)
Exam C
hypothetical mean
or collective premium
Exam C
Exam C
expected value of the hypothetical means
EVHM
process variance
Exam C
Exam C
µ
X
gamma α +
xi ,
θ
nθ + 1
¶
Z
∞
0
beta(a +
X
xi , b + km −
X
normal
gamma−1 (α + n, θ +
xi , 1)
µ· P
¸ ·
¸
xi
µ
n
1
+ 2 / 2+ 2 ,
σ2
a
σ
a
n
σ2
1
+
−c
ex
Γ(k − 1)
dx =
k
x
ck−1
(k − 2)!
=
ck−1
¶
1
a2
Ã
·
k>1
k≥2
X
1 X 1
gamma α + n,
+
θ
xi
xi )
¸−1 !
µ(θ) = E(Xij |Θi = θ)
single.pareto(α + n, max(x, θ))
µ = E[µ(θ)]
v(θ) = mij V ar(Xij |Θi = θ)
expected value of the process variance
EVPV
variance of the hypothetical means
VHM
Exam C
Bühlmann’s k
or credibility coefficient
Exam C
Bühlmann credibility factor
Exam C
Bühlmann credibility premium
Exam C
Var(X) = f(EVPV,VHM)
Exam C
Non-Paramtric estimation: “µ”
Exam C
Non-Paramtric estimation: “v”
Exam C
Non-Paramtric estimation: “a”
Method using c =
Exam C
Non-Paramtric estimation: “a”
Loss models technique
Exam C
Exam C
a = V ar[µ(θ)]
Zi =
v = E[v(θ)]
mi
mi + v/a
k=
Var(X) = a + v = EV P V + V HM
r
Zi X̄ + (1 − Zi )µ
n
i
XX
¢2
¡
1
mij Xij − X i
(n
−
1)
i=1 i
i=1 j=1
µ=X
v = Pr
Ã
a=
r
1 X 2
m−
m
m i=1 i
!−1 "
r
X
i=1
#
2
mi (X̄i − X̄) − v(r − 1)
v
a
" r
#−1
mi ´
r − 1 X mi ³
1−
c=
r
m
m
i=1
Ã
" r
!
#
X mi
r
vr
2
a=c
(X i − X) −
r − 1 i=1 m
m
³
´
vr
a = c Var(X i ) −
m
Non-Paramtric estimation: ”a”
If µ is given
only data available is for policy holder i
Non-Paramtric estimation: ”a”
µ is given
Exam C
inverse transformed method
Exam C
bootstrap estimate of the mean squared error
Exam C
Chi-Square Test for
number of claims is the result of a sum of a number
(n) of i.i.d random variables (x)
Exam C
Exam C
Pni
vi =
j=1
a=
2
mij (Xij − X i )
r
X
mi
i=1
ni − 1
vi
ai = (X i − µ)2 −
mi
Data: y = {y1 , . . . , yn }
A statistic: θ from the empirical distribution function.
m
(X i − µ)2 −
−1
x = FX
(rand(0, 1))
(Continuous)
F (xj−1 ) ≤ rand(0, 1) < F (xj )
xij = yrandi (1,n)
i = 1, . . . , m;
j = 1, . . . , n
θ̂i = g(xi )
m
´2
1 X³
M SE(θ̂) =
θ̂i − θ = Var(θ̂) + bias2θ̂
m i=1
k
X
(Ej − Oj )2
χ =
Vj
j=1
2
Ej = nE(x)
Vj = nVar(x)
r
v
m
(Discrete)