Академический Документы
Профессиональный Документы
Культура Документы
f (x) d
The hazard rate function h(x) is defined as h(x) = S(x) = dx (ln S(x)).
Rx
The cumulative hazard rate function H(x) is defined as H(x) =
h(t)dt = ln Sx).
3 E[(X )3 ]
The skewness 1 is defined as 1 = = .
3 E[(X )2 ]3/2
4 E[(X )4 ]
The kurtosis 2 is defined as 2 = = .
4 E[(X )2 ]2
The coefficient of variation is .
Cov[X, Y ]
The correlation coefficient XY is defined as XY = .
X Y
The 100pth percentile p is any number such that F (p ) = p. Another name for p is Value-at-Risk at security
level p and is also denoted by VaRp (X).
The probability generating function PX (t) is defined as PX (t) = E[tX ] = MX (ln t).
1 (n)
The Taylor coefficients of PX (t) are the probabilities of X: pn = P r(X = n) = n! PX (0).
A sample is a set of observations from n independent and identically distributed random variables.
A sample mean is the mean of the sample: sum of the observations divided by number of observations.
The covariance matrix of n random variables X1 , X2 , . . . , Xn is the matrix [aij ]i,j = [Cov[Xi , Xj ]]i,j . The covari-
ance matrix is symmetric and positive semi-definite.
1
0.2 Useful Results from Probability Theory
Bayes Theorem
Compound Mean and Variance Formula: If Xi s are independent and identically distributed, N is independent
of each Xi , and S = X1 + X2 + + XN , then
"N # "N #
X X
PS (z) = PN (PX (z)), E Xi = E[N ]E[X], V ar Xi = E[N ]V ar[X] + V ar[N ]E[X]2 .
i=1 i=1
(The mean and variance formulas follow from differentiating the pgf).
1
Variance of Sample Mean: V ar[X] = V ar[X].
n
Bernoulli Shortcut: If X is Bernoulli with values a or b, with probabilities q or 1q, then V ar[X] = (ab)2 q(1q).
Continuity Correction: If X is discrete and we want to approximate X using a continuous variable Y , then we
make the following continuity correction: If a and b are two consecutive values of X and c (a, b), then
If Y is the payment variable with an ordinary deductible d, then Y = max{0, X d}. This variable is called the
payment per loss variable and is denoted by Y L = (X d)+ . We have FY L (x) = FX (x + d).
X = (X d) + (X d)+ = (X d) + Y L .
FX (x + d) FX (d)
The payment per payment variable Y P is defined as Y P = (Xd)+ |X > d. We have FY P (x) = .
1 FX (d)
2
The mean of Y P is denoted by eX (d) and is called the mean excess loss.
Subadditivity: (X + Y ) = (X) + (Y ).
The constant a is 0 for poisson, negative for binomial and positive for negative binomial.
If pk and pk are two distributions in the (a, b, 1) class (with the same a and b) with pgf P (z) and P (z) respectively,
then (P (z) p0 )(1 p0 ) = (P (z) p0 )(1 p0 ).
If pk and pk are two distributions in the (a, b, 1) class (with the same a and b), then pk (1 p0 ) = pk (1 p0 ).
0.7 Poisson/Gamma
If Loss is poisson with parameter and is gamma with parameters (, ), then the unconditional loss frequency
for an insured is negative binomial with parameters r = and = . (Follows from looking at pgf)
If a per loss deductible d is introduced, then we can find the new expected annual aggregate payments by modiying
the loss frequency distribution to payment frequency, changing loss severity to payment per payment and
3
0.9 Aggregate Claims when Severity is Discrete
If S jumps by h 6= 1, then we can study S 0 = S/h and S 0 has integer jumps. We have E[S] = hE[S 0 ]. Thus assume
S jumps by integers.
Recursive Formula: Let frequency N be in (a, b, 1) class, X denote severity which is discrete and S denote the
aggregate loss. Let fk = P r(X = k), pk = P r(N = k), and gk = P r(S = k). Then g0 = PN (f0 ) and
" k #
1 X j
gk = (p1 (a + b)p0 )fk + a+b fj gkj
1 af0 i=1
k
If d is the aggregate deductible and S is the aggregate losses, then the net stop-loss premium is defined as
E[(S d)+ ] = E[S] E[S d].
To calculate net stop-loss premium, we can first calculate E[S d] using
dde1 d
X X
E[S d] = kgk + dP r(S d) = d (d k)gk .
k=0 k=0
The net stop-loss premium E[(S x)+ ] is a linear function of x between possible values of S: If S assumes no values
between d and u and x (d, u), then
ux xd
E[(S x)+ ] = E[(S d)+ ] + E[(S u)+ ].
ud ud
b .
Bias: biasb() = E[|]
The sample mean and sample variance are unbiased estimators of population mean and variance.
Asymptotically unbiased estimator: b is asymptotically unbiased lim biasbn () = 0 lim E[bn |] = .
n n
Consistent (Weakly consistent): b is consistent (AKA weakly consistent) > 0, lim P r(|bn | < ) = 1.
n
Confidence Interval: the 100p% confidence interval of an estimator b of a parameter (using normal approxima-
b
tion) is the solution of p 1 ((1 + p)/2).
v()
2
n n n
2 1 X 2 1 X 1X
Proof that the sample variance S = (Xi x) = Xi Xj is an unbiased estimator of
n 1 i=1 n 1 i=1 n j=1
the population variance:
"
n
# " n # n
" n # n
X X X X X
2 2 2 2 2
E[(n 1)S ] = E (Xi x) = E (Xi 2xXi + x ) = E[Xi ] 2E x Xi + E[x2 ]
i=1 i=1 i=1 i=1 i=1
= nE[X 2 ] 2E[nx2 ] + nE[x2 ] = nE[X 2 ] nE[x2 ] = nE[X 2 ] n Var[x] + E[x]2
n n
!
1 X 1 X
= nE[X 2 ] n Var[ Xi ] + E[ Xi ]2 = nE[X 2 ] Var[X] nE[X]2 = (n 1)Var[X].
n i=1 n i=1
4
0.11 Variance of Empirical Estimators: Complete individual data
For a sample of size n and any interval I, the variance of the empirical estimator of P r(X I) is given by
Case 1: a and b are end points of groups: In this case we can find variance just the way we do for complete
d r(X I)) = 13 (# in I)(# out I)
individual data. Var(P n
Case 2: If a is an end point of a group but b is not: Let J = (ci , ci+1 ) be the group containing b. Let
bci
K = (a, ci ) and = ci+1 ci . Then using linear approximation we have
Case 4: Neither a nor b is an end point of a group: Complicated to write, but has similar idea.
5
0.15 Kernel-Smoothed Distributions: Uniform Kernel
If Y is the empirical distribution and X is the kernel-smoothed distribution using uniform kernel with bandwidth b, then
b2
Var[X] = Var[Y ] + .
3
6
and the variance of the maximum likelihood estimator of is I()1 .
Information matrix for two variable: If l(1 , 2 ) is the log of the likelihood of the observation, then the
information matrix I(1 , 2 ) is defined as
2l
l l
I(1 , 2 ) = EX = EX
i j i,j i j i,j
Asymptotic covariance matrix of MLE for two variables: Asymptotic covariance matrix of MLE for two
variables is the inverse of its information matrix.
Delta Method for Variance of a function: If 1 and 2 are estimated using MLE and g(1 , 2 ) is some function
of 1 and 2 , then an estimate of Var[g(1 , 2 )] is
g1
Var[g(1 , 2 )] [g1 , g2 ] [Covariance Matrix] = g21 Var[1 ] + 2g1 g2 Cov[1 , 2 ] + g22 Var[2 ]
g2
where all the values are evaluated at the estimated values of 1 and 2 .
Confidence interval of MLE: p The confidence interval for a function g(1 , 2 ) where 1 and 2 are estimated
using MLE, is g(1 , 2 ) zp Var[g(1 , 2 )]. The values are evaluated at the estimate.
j
p p plot: is the graph of the points ( n+1 , F (xj )), j = 1, 2, . . . , n.
Hypothesis test rejection and acceptance levels: Reject below and accept above.
7
0.25 Chi-square Goodness of Fit Test
Chi-square Statistic:
k k k
X (Oj Ej )2 X Oj2 X (Oj Ej )2
Q= = n=
j=1
Ej j=1
Ej j=1
Vj
where Oj = #(observed observations in group j), Ej = #(expected observations in group j), k =#(groups),
n =#(observations), Vj =Expected variance in group j.
Read question carefully to find the degrees of freedom.
If the data is not given in grouped form, then group them so that each group has at least 5 expected observations.
Degrees of freedom = #(groups) - #(estimated parameters) - #(restrictions).
Exposure needed (claims needed) so that average claim size is within 100r% of expected claim size 100p% of the
2
y 2
time is e rp = n0 CV 2 where CV is the coefficient of variation of the severity distribution.
Exposure needed so that average number of claims is within 100r% of expected number of claims 100p% of the time
2
y 2
is e rp = n0 CV 2 where CV is its coefficient of variation of the frequency distribution.
8
0.30 Bayesian Estimation and Credibility: Continuous Prior
X has a parameter and has density () called prior. After making n observations x1 , x2 , . . . , xn of X the distribution
of changes (called posterior) and has density
posterior() = #prior()(likelihood of x1 , x2 , . . . , xn given )
where # is chosen so that the total integral is 1.
Expectation of X is called the Bayesian premium and is found using E[X] = E [EX [X|]].
9
0.36 Buhlman-Straub Credibility
This is the same as Buhlman credibility, except that the exposure per period is not 1.
10