Professional Documents
Culture Documents
Agenda
Research Overview
Model Testing
Real Data
Model Enhancement
Further Development
I. Research Overview
severity
state
Loss
Simulation
Model
run simulations
Stochastic
claim and
reserve data
Choose
the best
reserve
method
Compare
against the
simulated
claim data
reserve distributions
Apply
different
reserve
methods
Pass
Test against
real
experience /
model
assumption
0.14
Mean
Method B
Method A
0.12
0.08
Method A Method B
73.7%
81.2%
90.3%
96.7%
96.6%
99.5%
99.9%
98.9%
99.6%
100.0%
<
Reserve
Method B is
good
enough?
0.1
Claim
83.5%
95.7%
99.0%
99.8%
99.9%
0.06
0.04
0.02
0
2 4 6
8 10 12 14 16 18 20 22 24 26 28 30
Aggregate Claim
Claim
Method A
One out of
hundreds of
examples
Method B
Overview
Test some items suggested but not fully addressed in the
CAS LSMWP summary report Modeling Loss Emergence
and Settlement Processes
Fit real claim data to models.
Build two-state regime-switching feature in the LSM to
add an extra layer of flexibility to describe claim data.
Software: LSM and R. The source code of model testing
and model fitting using R is provided.
Model Testing
Real Claim
Data and
Reserve Data
severity
state
Loss
Simulation
Model
run simulations
Stochastic
claim and
reserve data
Choose
the best
reserve
method
Compare
against the
simulated
claim data
reserve distributions
Apply
different
reserve
methods
Pass
Test against
real
experience /
model
assumption
Correlation
Severity trend
Case reserve adequacy distribution
severity
state
Loss
Simulation
Model
run simulations
Stochastic
claim and
reserve data
Choose
the best
reserve
method
Compare
against the
simulated
claim data
reserve distributions
Apply different
reserve
methods
Pass
Test against
real
experience /
model
assumption
Severity
Trend
Correlation
Model Enhancement
Real Claim
Data and
Reserve Data
severity
state
Loss
Simulation
Model
run simulations
Stochastic
claim and
reserve data
Choose
the best
reserve
method
Compare
against the
simulated
claim data
reserve distributions
Apply different
reserve
methods
Pass
Test against
real
experience /
model
assumption
10
11
DAY ONE
9 AM
Tom, our company plans to use
the loss simulation model to
help our reserving works. Lets do
some tests first to get a better
understanding of the model.
Start from the frequency
model.
12
160
freq.ex
120
140
100
100
50
0
Frequency
150
180
200
200
220
100
120
140
160
dataf1
180
200
100
120
140
160
180
200
13
197.4
p value
0.64
size
117.2
9.5
144.2
0.57
Model Assumption
Size
100
Prob.
0.4
Mean ()
150
Variance
375
ML estimation
117
0.448
144.2
321.5
R code extract
# Goodness of fit test
14
DAY ONE
5 PM
15
Correlation
- t Copula
16
Cn (u ) =
ln(1 +
(e u1 1)(e u 2 1) (e u n 1)
(e 1) n 1
>0
Frequencies simulation
- Two Lines with annual frequency Poisson ( = 96)
- Monthly exposure: 1
- Frequency Trend: 1
- Seasonality: 1
- Accident Year: 2000
- Random Seed: 16807
- Frequency correlation: = 8, n = 2
- # of Simulations: 1000
Test Method
- Scatter plot
- Goodness-of-fit test
(k )
n
i =1
3. p value by parametric
bootstrapping
Enterprise
Risk Management
17
Scatter plot
Simulated Frequencies
0.6
0.0
0.2
0.4
Line2
0.8
1.0
0.0
Goodness-of-fit test
-
0.2
0.4
0.6
0.8
1.0
Line1
R code extract
# parameter estimation
fit.gumbel<-fitCopula(gumbel.cop,x,method="ml")
fit.gumbel<-fitCopula(gumbel.cop,x,method="itau")
# Copula Goodness-of-fit test
gofCopula(gumbel.cop, x, N=100, method = "mpl")
gofCopula(gumbel.cop, x, N=100, method = "itau")
18
: correlation matrix
: normal cumulative distribution function
Claim simulation
- One Line with annual frequency Poisson ( = 120)
- Monthly exposure: 1
- Frequency Trend: 1.05
- Seasonality: 1
- Accident Year: 2000
- Random Seed: 16807
- Payment Lag: Exponential with rate = 0.00274, which implies a
mean of 365 days.
- Size of entire loss: Lognormal with = 11.17 and = 0.83
- Correlation between payment lag and size of loss: normal copula
with correlation = 0.85, dimension 2
- # of Simulations: 10
19
Scatter plot
Simulated claim size vs. report lag
0.8
0.6
0.4
0.2
0.0
0.0
0.2
0.4
V2
x[,2]
0.6
0.8
1.0
1.0
0.0
0.2
0.4
0.6
0.8
1.0
x[,1]
Goodness-of-fit test
-
0.0
0.2
0.4
0.6
0.8
1.0
V1
20
DAY THREE
9 AM
21
Severity Trend
The LSM has two ways to model it
Trend factor (cum)
(Persistency of the force of the trend)
22
Severity Trend
Trend factor Test
data
5000
5e+05
trend
2000
2001
2002
2003
Time
2004
2005
-20000
20000
remainder
1e+05
2e+05
3e+05
3e+05
ts1
-5000 0
4e+05
seasonal
5e+05
Decomposition
6e+05
7e+05
1e+05
2006
2000
2001
2002
2003
2004
2005
R code extract
# set up time series
ts1<-ts(data,start=2000,frequency=12)
plot(ts1)
#decomposition
plot(stl(ts1,s.window="periodic"))
# linear trend fitting
trend = time(ts1)-2000
reg = lm(log(ts1)~trend, na.action=NULL)
2006
time
t value
Pr(>|t |)
Estimate
Std. Error
(Intercept)
11.034162 0.007526
1466.1
<2e -16
trend
0.405552 0.002196
184.7
<2e -16
Residual standard error: 0.03226 on 70 degrees of freedom
Multiple R-squared: 0.998,
Adjusted R-squared: 0.9979
F -statistic: 3.412e+04 on 1 and 70 DF, p -value: < 2.2e -16
Enterprise
exp(0.405552) = 1.50013 vs. model input
1.5 Risk Management
23
Severity Trend
24
Severity Trend
Trend persistency Test
QQ plot of severity
6e+05
Seve.ex
4e+05
4e-06
0e+00
2e+05
2e-06
0e+00
yhist
6e-06
8e+05
1e+06
8e-06
0e+00
1e+05
2e+05
3e+05
4e+05
5e+05
6e+05
0e+00
1e+05
2e+05
3e+05
4e+05
5e+05
6e+05
xhist
meanlog
11.32
0.052
sdlog
0.80
0.037
R code extract
# Kolmogorov-Smirnov Tests
ks.test(a,"plnorm", meanlog=11.32,
sdlog=0.8)
# Anderson-Darling Test
library(nortest) ## package loading
ad.test(datas1.norm)
25
DAY FOUR
9 AM
26
/2
1.2856
Enterprise Risk Management
27
1.4
1.5
1.3
1.2
Seve.ex
0.4
0.6
0.8
1.0
1.2
1.4
1.6
1.8
meanlog
0.08
0.014
sdlog
0.32
0.010
28
29
DAY FIVE
5 PM
30
Real Data
Marine claim data for distribution fitting, trend analysis, and correlation
analysis
- two product lines: Property and Liability
- data period: 2006 2010
- accident date, payment date, and final payment amount
1 2
5
6
4
2006
2007
2008
2009
-4
remainder trend
-2
10
seasonal data
15
15
Historical Frequency
ts1
2010
2006
Time
2007
2008
2009
2010
time
31
Real Data
Fit the frequency (continued)
- Linear regression for trend analysis
Log(Monthly Frequency) = Intercept + trend * (time 2006) + error term
Coefficients:
Estimate Std. Error t value
Pr(>|t |)
(Intercept)
1.93060 0.15164 12.732
<2e- 16
trend
-0.14570
0.05919 -2.462
0.0172
Residual standard error: 0.5649 on 52 degrees of freedom.
Multiple R-squared: 0.1044, Adjusted R-squared: 0.08715.
F -statistic: 6.06 on 1 and 52 DF, p -value: 0.01718.
0.0
0.5
1.0
1.5
2.0
2.5
Trend Fitting
log(ts1)
2006
2007
2008
Time
2009
2010
32
Real Data
Fit the frequency (continued)
- Detrend the frequency and fit to the lognormal distribution
meanlog
sdlog
Estimation
9.5539259
Standard Deviation 0.4260991
3.1311762
0.3012976
10
15
20
freq.ex
10
15
detrend$Freq
20
33
Real Data
Fit the Severity
Correlation calibration
Empirical Correlation
1.0
0.0
0.2
0.4
0.6
0.8
1.0
x[,1]
0.6
Line2
0.6
0.4
0.2
0.0
x[,2]
0.8
1.0
0.0
0.2
0.4
0.6
0.8
1.0
Line1
What is missing?
Historical reserve data which
are essential for case reserve
adequacy modeling.
Enterprise Risk Management
34
35
The model
- Two distinct distributions represent different states
- Transition rules from one state to another
P 11: state 1 persistency, the probability that the state will be 1 next
1 + 2 = 1
36
The Simulation
- Steps
1.
2.
3.
4.
5.
- Test Parameters
1 =
1 P22
1 0.7
=
= 0.375
2 P11 P22 2 0.5 0.7
2 =
1 P11
1 0.7
=
= 0.625
2 P11 P22 2 0.5 0.7
0.529508789768443
2
0.0441845036111772
2
0.994539848994464
1
0.21886122901924
1
0.0928565948270261
1
0.797880138037726
2
0.129500501556322
2
0.24027365935035
2
0.797712686471641
1
0.0569291599094868 Enterprise
1
Risk
RN>0.5
RN<0.7
RN>0.7
RN<0.5
RN<0.5
RN>0.5
RN<0.7
RN<0.7
RN>0.7
Management
RN<0.5
37
P
P
0
.
1
0
.
9
22
21
( 1 2 ) = (10.53% 89.47% )
P
P
0
.
1
0
.
9
22
21
( 1 2 ) = (11.11% 88.89%)
38
Set 2
0.0
0.2
0.4
0.6
0.8
0.8
Line2
1.0
0.0
0.0
0.0
0.2
0.2
0.2
0.4
0.4
0.4
Line2
0.6
0.6
0.6
0.8
0.8
1.0
1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0
0.2
0.4
0.6
0.8
1.0
Line1
rcopula(normal.cop, 1000)[,1]
Line1
Set 4
Set 1: State 1 for line 1 and state
1 for line 1
Set 2: State 1 for line 1 and state
2 for line 2
Set 3: State 2 for line 1 and state
1 for line 1
Set 4: State 2 for line 2 and state
2 for line 2
0.6
0.6
0.8
0.8
1.0
1.0
Set 3
0.0
0.0
0.2
0.2
0.4
0.4
Line2
Line2
rcopula(normal.cop, 1000)[,2]
0.0
0.2
0.4
0.6
Line1
0.8
1.0
0.0
0.2
0.4
0.6
Line1
0.8
1.0
39
Interface
Input
40
Interface
Output
- Additional column in claim and transaction output files to record the state
- Showing state and random number while simulating
41
Congratulations!
42
V. Further Development
43
Further Development
Case reserve adequacy test shows that the assumption is
not consistent with simulation data.
This may be caused by the linear interpolation method
used to derive 40% time point case reserve.
It is suggested revising the way in which valuation date is
determined in the LSM. In addition to the simulated
valuation dates based on the waiting-period distribution
assumption as in the LSM, some deterministic time points
can be added as valuation dates.
In the LSM, 0%, 40%, 70%, and 90% time-points, case
reserve adequacy distribution can be input into the model.
Therefore, 0%, 40%, 70% and 90% time points may be
added as deterministic valuation dates.
Enterprise Risk Management
44
Thank you!
45