Mahler.Credibility: Difference between revisions

From BattleActs Wiki
Jump to navigation Jump to search
No edit summary
 
(24 intermediate revisions by 2 users not shown)
Line 3: Line 3:
'''Synopsis''': Tested on the paper but not the appendices.
'''Synopsis''': Tested on the paper but not the appendices.


The paper deals with optimally combining different years of historical data when experience rating. The usual application of credibility, namely ''Data'' * ''Credibility'' + ''Prior Estimate'' * ''Complement of Credibility'' is used. The prior estimate of credibility is normally the average experience for a class of insureds but could also be the relative loss potential of the insured to the class average.
The paper deals with optimally combining different years of historical data when experience rating. The usual application of credibility, namely ''Data'' * ''Credibility'' + ''Prior Estimate'' * (1 - ''Credibility'') is used. The prior estimate of credibility is normally the average experience for a class of insureds but could also be the relative loss potential of the insured to the class average.


The goal is to understand how changes in the risk classification parameters over time results in older years of data adding less credibility to experience rating than expected.
The goal is to understand how changes in the risk classification parameters over time result in older years of data adding less credibility to experience rating than expected.


==Study Tips==
==Study Tips==


To follow...
This reading has historically produced somewhat predictable exam questions which should be straightforward points once you've mastered the basics of the article. However, with the introduction of integrative questions it's less likely you'll get a standalone question on this article as the material can be applied to a wide range of topics.


'''Estimated study time''': x mins, or y hrs, or n<sub>1</sub>-n<sub>2</sub> days, or 1 week,... ''(not including subsequent review time)''
'''Estimated study time''': 8 Hours ''(not including subsequent review time)''


==BattleTable==
==BattleTable==
Line 17: Line 17:
Based on past exams, the '''main things''' you need to know ''(in rough order of importance)'' are:
Based on past exams, the '''main things''' you need to know ''(in rough order of importance)'' are:


* fact A...
* How to apply and interpret the results of a <span style="color:green;">'''Chi-squared test'''</span>.
* fact B...
* Be able to describe <span style="color:blue;">'''alternate tests'''</span> suggested by Mahler.
 
* Understand and apply the '''mean squared error''' statistic.
: {| class='wikitable' style='width: 1000px;'


{|class="wikitable"
|-
|-
! style='width: 175px;' | reference !! style='width: 175px;' | part (a) !! style='width: 175px;' | part (b) !! style='width: 175px;' | part (c) !! part (d)
| Questions from the Fall 2019 exam are held out for practice purposes. (They are included in the CAS practice exam.)  
|}


|-
: {| class='wikitable' style='width: 1000px;'
|| [https://www.battleactsmain.ca/pdf/Exam_(2018_1-Spring)/(2018_1-Spring)_(01).pdf <span style='font-size: 12px; background-color: yellow; border: solid; border-width: 1px; border-radius: 5px; padding: 2px 5px 2px 5px; margin: 5px;'>E</span>] <span style='color: red;'>'''(2018.Spring #1)'''</span>
||
||
||
||


|-
|-
|| [https://www.battleactsmain.ca/pdf/Exam_(2018_1-Spring)/(2018_1-Spring)_(01).pdf <span style='font-size: 12px; background-color: yellow; border: solid; border-width: 1px; border-radius: 5px; padding: 2px 5px 2px 5px; margin: 5px;'>E</span>] <span style='color: red;'>'''(2018.Spring #1)'''</span>
! style='width: 175px;' | reference !! style='width: 175px;' | part (a) !! style='width: 175px;' | part (b) !! style='width: 200px;' | part (c) !! part (d)
||
||
||
||


|-
|- style="border-bottom: 2px solid;"
|| [https://www.battleactsmain.ca/pdf/Exam_(2018_1-Spring)/(2018_1-Spring)_(01).pdf <span style='font-size: 12px; background-color: yellow; border: solid; border-width: 1px; border-radius: 5px; padding: 2px 5px 2px 5px; margin: 5px;'>E</span>] <span style='color: red;'>'''(2018.Spring #1)'''</span>
|| [https://www.battleacts8.ca/8/pdf/Exam_(2018_2-Fall)/(2018_2-Fall)_(01).pdf <span style='font-size: 12px; background-color: yellow; border: solid; border-width: 1px; border-radius: 5px; padding: 2px 5px 2px 5px; margin: 5px;'>E</span>] <span style='color: red;'>'''(2018.Fall #1)'''</span>
||  
|| <span style="color:green;">'''Chi-squared Test'''</span> <br> - calculate
||  
|| '''Mean Square Error''' <br> - use in calculation
||
| style="background-color: lightgrey;" | ''Reinsurance Profit''<br> - <span style="color:brown;">'''outdated'''</span>
||
| style="background-color: lightgrey;" | ''Lognormal Probabilities'' <br> - ''[[Bahnemann.Chapter5]]''


|-
|- style="border-bottom: 2px solid;"
|| [https://www.battleactsmain.ca/pdf/Exam_(2018_1-Spring)/(2018_1-Spring)_(01).pdf <span style='font-size: 12px; background-color: yellow; border: solid; border-width: 1px; border-radius: 5px; padding: 2px 5px 2px 5px; margin: 5px;'>E</span>] <span style='color: red;'>'''(2018.Spring #1)'''</span>
|| [https://www.battleacts8.ca/8/pdf/Exam_(2015_2-Fall)/(2015_2-Fall)_(04).pdf <span style='font-size: 12px; background-color: yellow; border: solid; border-width: 1px; border-radius: 5px; padding: 2px 5px 2px 5px; margin: 5px;'>E</span>] <span style='color: red;'>'''(2015.Fall #4)'''</span>
||  
|| <span style="color:green;">'''Chi-squared Test'''</span> <br> - apply
||  
|| <span style="color:blue;">'''Shifting Frequency'''</span> <br> - alternate tests
||
| style="background-color: lightgrey;" |
||
| style="background-color: lightgrey;" |


|-
|- style="border-bottom: 2px solid;"
|| [https://www.battleactsmain.ca/pdf/Exam_(2018_1-Spring)/(2018_1-Spring)_(01).pdf <span style='font-size: 12px; background-color: yellow; border: solid; border-width: 1px; border-radius: 5px; padding: 2px 5px 2px 5px; margin: 5px;'>E</span>] <span style='color: red;'>'''(2018.Spring #1)'''</span>
|| [https://www.battleacts8.ca/8/pdf/Exam_(2013_2-Fall)/(2013_2-Fall)_(03).pdf <span style='font-size: 12px; background-color: yellow; border: solid; border-width: 1px; border-radius: 5px; padding: 2px 5px 2px 5px; margin: 5px;'>E</span>] <span style='color: red;'>'''(2013.Fall #3)'''</span>
||  
| style="background-color: lightgrey;" | ''Holdout Sample'' <br> - ''[[Couret.Venter]]''
||  
| style="background-color: lightgrey;" | ''Holdout Sample'' <br> - ''[[Couret.Venter]]''
||
| style="background-color: lightgrey;" | ''Sum of Squared Errors Test'' <br> - ''[[Couret.Venter]]''
||
|| '''Frequency Trend''' <br> - describe test


|-
|- style="border-bottom: 2px solid;"
|| [https://www.battleactsmain.ca/pdf/Exam_(2018_1-Spring)/(2018_1-Spring)_(01).pdf <span style='font-size: 12px; background-color: yellow; border: solid; border-width: 1px; border-radius: 5px; padding: 2px 5px 2px 5px; margin: 5px;'>E</span>] <span style='color: red;'>'''(2018.Spring #1)'''</span>
|| [https://www.battleacts8.ca/8/pdf/Exam_(2012_2-Fall)/(2012_2-Fall)_(03).pdf <span style='font-size: 12px; background-color: yellow; border: solid; border-width: 1px; border-radius: 5px; padding: 2px 5px 2px 5px; margin: 5px;'>E</span>] <span style='color: red;'>'''(2012.Fall #3)'''</span>
||  
|| <span style="color:green;">'''Chi-squared Test'''</span> <br> - calculate
||  
|| <span style="color:blue;">'''Shifting Frequency'''</span> <br> - alternate tests
||
| style="background-color: lightgrey;" |
||
| style="background-color: lightgrey;" |


|-
|}
|| [https://www.battleactsmain.ca/pdf/Exam_(2018_1-Spring)/(2018_1-Spring)_(01).pdf <span style='font-size: 12px; background-color: yellow; border: solid; border-width: 1px; border-radius: 5px; padding: 2px 5px 2px 5px; margin: 5px;'>E</span>] <span style='color: red;'>'''(2018.Spring #1)'''</span>
||
||
||
||


<!-- ******** BattleBar Code ******** -->
{|style="border: solid; color:lightgrey; border-radius:10px; border-width:2px; align:center;"
|-
|-
|| [https://www.battleactsmain.ca/pdf/Exam_(2018_1-Spring)/(2018_1-Spring)_(01).pdf <span style='font-size: 12px; background-color: yellow; border: solid; border-width: 1px; border-radius: 5px; padding: 2px 5px 2px 5px; margin: 5px;'>E</span>] <span style='color: red;'>'''(2018.Spring #1)'''</span>
<!-- ******** Full BattleQuiz ******** -->
||
|style="padding:2px"|[https://battleacts8.ca/8/FC.php?selectString=**&filter=both&sortOrder=natural&colorFlag=allFlag&colorStatus=allStatus&priority=importance-high&subsetFlag=miniQuiz&prefix=Mahler&suffix=Credibility&section=all&subSection=all&examRep=all&examYear=all&examTerm=all&quizNum=all<span style="font-size: 20px; background-color: lightgreen; border: solid; border-width: 1px; border-radius: 10px; padding: 2px 10px 2px 10px; margin: 10px;">'''Full BattleQuiz]'''</span>
||
<!--                                   -->
||
<!-- ******** Excel BattleQuiz ******** -->
||
<!--
 
|style="padding:2px"|[https://battleacts8.ca/8/FC.php?selectString=**&filter=both&sortOrder=natural&colorFlag=allFlag&colorStatus=allStatus&priority=importance-high&subsetFlag=miniQuiz&prefix=Mahler&suffix=Credibility&section=all&subSection=all&examRep=all&examYear=all&examTerm=all&quizNum=all<span style="color: red; font-size: 20px; background-color: lightgreen; border: solid; border-width: 1px; border-radius: 10px; border-color: darkblue; padding: 2px 10px 2px 10px; margin: 0px;">'''''Excel BattleQuiz]'''''</span>
-->
<!--                                    -->
<!-- ******** Excel PowerPack Files ******** -->
|style="padding:2px"|[[BattleActs_PowerPack#ppMahler| <span style="color: white; font-size: 12px; background-color: indigo; border: solid; border-width: 2px; border-radius: 10px; border-color: indigo; padding: 1px 3px 1px 3px; margin: 0px;">'''''Excel Files '''''</span>]]
<!--                                    -->
<!-- ******** Forum ******** -->
|style="padding:2px"|[https://battleacts8.ca/8/forum/index.php?p=/categories/mahler-credibility<span style="font-size: 12px; background-color: lightgrey; border: solid; border-width: 1px; border-radius: 10px; padding: 2px 10px 2px 10px; margin: 0px;">'''Forum'''</span>]
<!--                                    -->
<!-- ******** Formula Sheet ******** -->
<!--
|style="padding:2px"|[https://battleacts8.ca/8/forum/categories/mahler-credibility<span style="font-size: 12px; color: darkblue; background-color: lightblue; border: solid; border-width: 1px; border-radius: 10px; padding: 2px 10px 2px 10px; margin: 0px;">'''Formula Sheet'''</span>]
-->
|}
|}
<span style="color: red;">'''You must be <u>logged in</u> or this will not work.'''</span>


==In Plain English!==
==In Plain English!==
Line 90: Line 92:
Next, Mahler asks if the differences in the win/loss record for a fixed team over time can be explained by random fluctuations from the same underlying distribution. He concludes the observed results cannot be explained this way, so the parameters of the distribution which describes the number of losses for a team in a year are changing over time.  
Next, Mahler asks if the differences in the win/loss record for a fixed team over time can be explained by random fluctuations from the same underlying distribution. He concludes the observed results cannot be explained this way, so the parameters of the distribution which describes the number of losses for a team in a year are changing over time.  


One method for testing if parameters shift over time is the standard chi-squared test. Mahler groups the data into 5-year periods (other lengths could also be used) by team and uses the chi-squared test to measure if they could have the same underlying mean over the entire dataset. If a team didn't change its losing percentage over time, then its losing percentage should be normally distributed around its average. The chi-squared statistic is defined as <math>\displaystyle\sum_{i=1}^n\frac{\left(\mbox{Actual Loss}_i - \mbox{Expected Loss}_i\right)^2}{\mbox{Expected Loss}_i}</math>. For a fixed team with ''n'' games, we have <math>n-1</math> degrees of freedom if we have to estimate the expected loss.
One method for testing if parameters shift over time is the standard <span style="color:red;>'''Chi-squared test'''</span>. Mahler groups the data into 5-year periods (other lengths could also be used) by team and uses the Chi-squared test to measure if they could have the same underlying mean over the entire dataset. If a team didn't change its losing percentage over time, then its losing percentage should be normally distributed around its average.
 
The Chi-squared statistic is defined as <math>\displaystyle\sum_{i=1}^n\frac{\left(\mbox{Actual Loss}_i - \mbox{Expected Loss}_i\right)^2}{\mbox{Expected Loss}_i}</math>.  
 
For a fixed team with ''n'' games, we have <math>n-1</math> degrees of freedom if we have to estimate the expected loss.


Another method is to compute the average correlation between risks.  
Another method is to compute the average correlation between risks.  
Line 100: Line 106:
For reference, the <span style = "color:red;">'''correlation coefficient'''</span> is defined as <math>\frac{Cov(X,Y)}{\sigma_X\cdot\sigma_Y} =\frac{E[\left(x-\bar{x}\right)\left(y-\bar{y}\right)]}{\sigma_X\cdot\sigma_Y}</math>.
For reference, the <span style = "color:red;">'''correlation coefficient'''</span> is defined as <math>\frac{Cov(X,Y)}{\sigma_X\cdot\sigma_Y} =\frac{E[\left(x-\bar{x}\right)\left(y-\bar{y}\right)]}{\sigma_X\cdot\sigma_Y}</math>.


Since we are given a lot of years of baseball data, for small ''t'' values the set of pairs ''t'' years apart is relatively large. As ''t'' increases, the volume of data decreases. Mahler notes that a non-zero observed correlation is not necessarily statistically significant as the 95% confidence interval about 0 is approximately <math>\pm0.1</math> and this increases as the number of data points decreases.
Since we are given a lot of years of baseball data, for small ''t'' values the set of pairs which are ''t'' years apart is relatively large. As ''t'' increases, the volume of data decreases. Mahler notes a non-zero observed correlation is not necessarily statistically significant as the 95% confidence interval about 0 is approximately <math>\pm0.1</math> and this increases as the number of data points decreases.


Looking at the results shown in the paper, we observe the correlation method yields correlations which are significantly greater for years that are closer together (small ''t'' values) than those further apart. In fact, after approximately 10 years there is no distinguishable correlation between years. This leads Mahler to conclude recent years can be used to predict the future.
Looking at the results shown in the paper, we observe the correlation method yields correlations which are significantly greater for years that are closer together (small ''t'' values) than those further apart. In fact, after approximately 10 years there is no distinguishable correlation between years.  
This leads Mahler to conclude <u>recent years can be used to predict the future</u>.


Let's see how the CAS might test this in an integrative question. Read through the following solution to 2018 Q1(a) and at a later date make sure you can fill in all the details from memory. '''Insert 2018.Q1a PDF'''
Let's see how the CAS might test this in an integrative question. Read through the following solution to 2018 Q1(a) and at a later date make sure you can fill in all the details from memory.


<u>'''Statement of the Problem'''</u>
: <span class="newwin">[https://www.battleacts8.ca/8/pdf/Q1a_2018.pdf <span style="color: white; font-size: 12px; background-color: green; border: solid; border-width: 2px; border-radius: 10px; border-color: green; padding: 1px 3px 1px 3px; margin: 0px;">'''''2018 Question 1 Part a'''''</span>]</span>


Let ''X'' be the quantity we want to estimate. Let ''Y<sub>i</sub>'' be various known estimates (the estimators) and set ''X'' to a weighted linear combination of the estimators, i.e. <math>X=\sum_i Z_i\cdot Y_i</math>. The goal is to find the optimal set of weights {''Z<sub>i</sub>''} that produces the best estimate of the future losing percentage.
[https://battleacts8.ca/8/FC.php?selectString=**&filter=both&sortOrder=natural&colorFlag=allFlag&colorStatus=allStatus&priority=importance-high&subsetFlag=miniQuiz&prefix=Mahler&suffix=Credibility&section=all&subSection=all&examRep=all&examYear=all&examTerm=all&quizNum=1<span style="font-size: 20px; background-color: aqua; border: solid; border-width: 1px; border-radius: 10px; padding: 2px 10px 2px 10px; margin: 10px;">'''mini BattleQuiz 1]'''</span> <span style="color: red;">'''You must be <u>logged in</u> or this will not work.'''</span>
 
===Statement of the Problem===
Let ''X'' be the quantity we want to estimate. Let ''Y<sub>i</sub>'' be various known estimates (the estimators) and let ''X'' be a weighted linear combination of the estimators, i.e. <math>X=\sum_i Z_i\cdot Y_i</math>. The goal is to find the optimal set of weights {''Z<sub>i</sub>''} which produces the best estimate of the future losing percentage.


===Four Simple Solutions===
===Four Simple Solutions===
Mahler eventually covers six solutions to the problem. We'll follow in his footsteps and begin with four easy options.
Mahler eventually covers six solutions to the problem. We'll follow in his footsteps and begin with four easy methods.


# Assume every risk is average, so set the predicted mean equal to 50%. This assumes all games have a winner/ties are negligible. It also ignores all historical data, so gives 0% credibility to the data.
# Assume every risk is average, so set the predicted mean equal to 50%. This assumes all games have a winner/ties are negligible. It also ignores all historical data, so gives 0% credibility to the data.
# Assume the previous year repeats itself. This gives 100% credibility to the previous year.
# Assume the previous year repeats itself. This gives 100% credibility to the previous year.
# Credibility weight the previous year with the grand mean.  
# Credibility weight the previous year with the grand mean.  
#*The grand mean is some external estimate that is independent of the data. In this case it is 50% because we assume equal likelihood of a win or loss.  
#*The <span style="color:red;">'''grand mean'''</span> is some '''<u>external estimate</u>''' that is independent of the data. In this case it is 50% because we assume equal likelihood of a win or loss.  
#* Cases 1 and 2 are special cases of this, corresponding to 0% and 100% credibility respectively.
#* Cases 1 and 2 are special cases of this, corresponding to 0% and 100% credibility respectively.
# Give equal weight to the most recent ''N'' years of data.  
# Give equal weight to the most recent ''N'' years of data.  
#* This can be further extended by calculating the credibility, ''Z'' of the ''N'' years of data and then giving each prior year weight <math>\frac{Z}{N}</math> and weight <math>1-Z</math> to the grand mean.
#* This can be further extended by calculating the credibility, ''Z'' of the ''N'' years of data and then giving each prior year weight <math>\frac{Z}{N}</math> and weight <math>1-Z</math> to the grand mean.


There are various choices for determining the credibility used in the fourth option. B&uuml;hlmann, Bayesian or classical limited fluctuation credibility methods could be used.
There are various choices for determining the credibility used in the fourth option. B&uuml;hlmann, Bayesian or classical limited fluctuation credibility methods could be used.


To see how the CAS can test this material in an integrative question, read through the following PDF solution to 2018 Q1(b). '''Insert 2018.Q1b PDF'''
To see how the CAS can test this material in an integrative question, read through the following PDF solution to 2018 Q1(b).  
At a later date, you should return to this exam and see if you can fill in all the details from memory.
: <span class="newwin">[https://www.battleacts8.ca/8/pdf/Q1b_2018.pdf <span style="color: white; font-size: 12px; background-color: green; border: solid; border-width: 2px; border-radius: 10px; border-color: green; padding: 1px 3px 1px 3px; margin: 0px;">'''''2018 Question 1 Part b'''''</span>]</span>
At a later date, you should return to this exam question and see if you can fill in all the details from memory.


===Three Criteria for Deciding Between Solutions===
===Three Criteria for Deciding Between Solutions===
Line 135: Line 146:
#* Rephrased - we're looking for credibilities which minimize <math>\mbox{Pr}\left(\frac{|X_{est,team}-X_{actual,team}|}{X_{est,team}}>k\%\right)</math>, where ''k''% is some predetermined threshold.
#* Rephrased - we're looking for credibilities which minimize <math>\mbox{Pr}\left(\frac{|X_{est,team}-X_{actual,team}|}{X_{est,team}}>k\%\right)</math>, where ''k''% is some predetermined threshold.
# Meyers/Dorweiler Credibility:  
# Meyers/Dorweiler Credibility:  
#* Calculate the correlation using the Kendall Tau statistic. The first quantity is the ratio of actual losing percentage to predicted losing percentage. The second quantity is the ratio of the predicted losing percentage to the overall average losing percentage.
#* Calculate the correlation using the Kendall Tau statistic (see Appendix B in the source for the definition). The first quantity is the ratio of actual losing percentage to predicted losing percentage. The second quantity is the ratio of the predicted losing percentage to the overall average losing percentage.


Section 8 of Mahler's paper applies each of these methods in turn to the baseball data. It's best given a quick skim-read. One key takeaway is found in Section 8.3: As the number of previous years, ''N'', used as estimators increases, the credibility of the entire ''N'' year period decreases when using B&uuml;hlmann or Bayesian techniques. This is counterintuitive at first because actuaries typically expect credibility to increase as the volume of data is increased. However, Mahler points out the result is actually what we '''should expect''' given the parameters are shifting significantly over time. This effect is also seen using classical credibility but isn't seen when Meyers/Dorweiler credibility is used.
Section 8 of Mahler's paper applies each of these methods in turn to the baseball data. It's best given a quick skim-read. One key takeaway is found in Section 8.3: As the number of previous years, ''N'', used as estimators increases, the credibility of the entire ''N'' year period decreases when using B&uuml;hlmann or Bayesian techniques. This is counterintuitive at first because actuaries typically expect credibility to increase as the volume of data is increased. However, Mahler points out the result is actually what we '''should expect''' given the parameters are shifting significantly over time. This effect is also seen using classical credibility but isn't seen when Meyers/Dorweiler credibility is used.
Line 144: Line 155:


===More General Solutions===
===More General Solutions===
We now cover the last two method for assigning credibility weights.
We now cover the last two methods for assigning credibility weights.


<u>Method 5</u>: <span style="color:red;">'''Single exponential smoothing'''</span>
<u>Method 5</u>: <span style="color:red;">'''Single exponential smoothing'''</span>


Here we assign geometric decreasing weights. The latest year gets weight ''Z'', the next year weight <math>Z\cdot(1-Z)</math>, and continue in this fashion until the second to oldest year gets weight <math>Z\cdot(1-Z)\ldots\cdot(1-Z)^{n-2}</math> while the oldest year has weight <math>(1-Z)^{n-1}</math>.
Here we assign geometric decreasing weights to ''n'' years of data. The latest year gets weight ''Z'', the next year weight <math>Z\cdot(1-Z)</math>, and continue in this fashion until the second to oldest year gets weight <math>Z\cdot(1-Z)^{n-2}</math> while the oldest year has weight <math>(1-Z)^{n-1}</math>.
Note that the weights sum to 1.
Note the weights sum to 1 (''Alice: "Can you prove this using geometric series?"'').


Method 5 is an example of a mathematical filter. The approach of giving a set of data credibility ''Z'' and the complement weight ''1-Z'' is an application of this where <math>n=1</math>.
Method 5 is an example of a mathematical filter. The approach of giving a set of data credibility ''Z'' and the complement weight ''1 - Z'' is an application of this where <math>n=2</math>. You can also apply exponential smoothing when you have ''k'' years of data and an independently derived complement. In this situation the latest year gets weight ''Z'', the next <math>Z\cdot(1-Z)</math>, continuing until the ''k''<sup>th</sup> year gets weight <math>Z\cdot (1-Z)^{k-1}</math> and the complement gets weight <math>(1-Z)^k</math>.


<u>Method 6</u>:
<u>Method 6</u>:
Line 166: Line 177:


===Equations for Least Squares Credibility===
===Equations for Least Squares Credibility===
Reading the conclusion of the Mahler paper makes this section feel like it should be the most testable. However, to date it appears there haven't been any exam questions on it...
Reading the conclusion of the Mahler paper makes this section feel like it should be testable. However, it didn't show up in the past exams released by the CAS.


The variance of a dataset may be split into two parts. The variance between risks which Mahler denotes by <math>\tau^2</math>, and the variance within risks which Mahler writes as <math>\delta^2+\zeta^2</math>. Here, <math>\delta^2</math> is the process variance and <math>\zeta^2</math> is the parameter variance (variance due to changes in the distribution parameters over time). This sub-division of the within risk isn't necessary for computing credibility here.
The variance of a dataset may be split into two parts. The variance between risks which Mahler denotes by <math>\tau^2</math>, and the variance within risks which Mahler writes as <math>\delta^2+\zeta^2</math>. Here, <math>\delta^2</math> is the process variance and <math>\zeta^2</math> is the parameter variance (variance due to changes in the distribution parameters over time). This sub-division of the within risk isn't necessary for computing credibility here.


Mahler applies method 6, that is, let year ''X<sub>i</sub>'' be given weight ''Z<sub>i</sub>'' and assign the balance to the grand mean ''M'' for the entire dataset. We get the following equation which relates the expected square error between the observation and prediction: <math>\displaystyle\sum_{i=1}^N\sum_{j=1}^N Z_iZ_j(\tau^2+C(|i-j|)) - 2\sum_{i=1}^N Z_i(\tau^2+C(N+\Delta-i)) +\tau^2 +C(0)</math>. Where <math>\tau^2</math> is the between variance, ''C(k)'' is the covariance for data for the same risk, ''k'' years apart (within covariance), ''C(0)'' is the within variance, and &Delta; is the length of time between the latest year of data used and the year being estimated, and ''N'' is the number of years of data being used.
Mahler applies method 6, that is, let year ''X<sub>i</sub>'' be given weight ''Z<sub>i</sub>'' and assign the balance to the grand mean ''M'' for the entire dataset. We get the following equation which relates the expected square error between the observation and prediction: <math>\displaystyle\sum_{i=1}^N\sum_{j=1}^N Z_iZ_j(\tau^2+C(|i-j|)) - 2\sum_{i=1}^N Z_i(\tau^2+C(N+\Delta-i)) +\tau^2 +C(0)</math>. Where <math>\tau^2</math> is the between variance, ''C(k)'' is the covariance for data for the same risk, ''k'' years apart (within covariance), ''C(0)'' is the within variance, &Delta; is the length of time between the latest year of data used and the year being estimated, and ''N'' is the number of years of data being used.


This is viewed as a second degree polynomial for the credibility weights ''Z<sub>i</sub>''. Most likely you would be given this equation on the exam if it was needed.
This is viewed as a second degree polynomial for the credibility weights ''Z<sub>i</sub>''. Most likely you would be given this equation on the exam if it was needed.


==Pop Quiz Answers==
Differentiating the above equation leads to ''N'' equations in ''N'' unknowns so it can be solved exactly. In particular, this equation can be solved for the case when the total assigned credibility ''Z<sub>i</sub>'' is distributed uniformly across the ''N'' years. Note that this method can produce credibilities which do not necessarily decline monotonically, and can even be negative. Negative weights allow more emphasis to be given to other years with an associated reduction in expected squared error.
 
The credibilities may also be restricted so they sum to 1 and hence place no weight on the grand mean. The resulting system of equations may be solved using Lagrange multipliers. In the paper, Mahler's results show the most recent year receives the greatest credibility. However, a monotonic decline is not guaranteed and nor are the weights necessarily all greater than zero.
 
At the end of Section 11, Mahler notes his results are reasonable for estimating the least squares credibility in baseball. However, the transfer of these results to another area is dependent on the covariance structure of the proposed dataset. Details are discussed in the appendix which is out of scope for the exam.
 
The three types of credibility considered in Mahler's paper can be viewed as two categories. B&uuml;hlmann/Bayesian credibility and limited fluctuation credibility attempt to eliminate large errors. B&uuml;hlmann/Bayesian credibility achieves this because large errors will have a disproportionate impact on the sum of square errors. Limited fluctuation credibility does this by minimizing the errors greater than a selected threshold.
 
However, Meyers/Dorweiler credibility may be viewed as concerned with the pattern of the errors. Large errors are not a problem as long as there is no pattern relating them to the experience rating modifications. Mahler concludes it's important to understand what any criterion tests and that there are hazards to solely relying on any one method.
 
===Mahler's Ratemaking Example===
At the end of the paper, Mahler gives a brief overview of an application to ratemaking. The idea is to combine the five most recent annual loss ratios to produce a rate level indication, i.e. <math>N=5</math>. We're told to assume there are three years between the latest year of data and the future average date of loss, i.e. <math>\Delta=3</math>. By assuming a covariance structure, placing no weight on the grand mean (because in ratemaking, what is our best guess for the grand mean?), and requiring the credibilities to sum to 1, there are ''N'' equations in ''N'' unknowns to solve. The equations in Mahler's example are:
*<math>Z_1C(0)+Z_2C(1)+Z_3C(2)+Z_4C(3)+Z_5C(4)=C(7)+\frac{\lambda}{2} </math>
*<math>Z_1C(1)+Z_2C(0)+Z_3C(1)+Z_4C(2)+Z_5C(3)=C(6)+\frac{\lambda}{2} </math>
*<math>Z_1C(2)+Z_2C(1)+Z_3C(0)+Z_4C(1)+Z_5C(2)=C(5)+\frac{\lambda}{2} </math>
*<math>Z_1C(3)+Z_2C(2)+Z_3C(1)+Z_4C(0)+Z_5C(1)=C(4)+\frac{\lambda}{2} </math>
*<math>Z_1C(4)+Z_2C(3)+Z_3C(2)+Z_4C(1)+Z_5C(0)=C(3)+\frac{\lambda}{2} </math>
 
This system can be solved using standard linear algebra techniques once the ''C(k)'' are known. The ''C(k)'' describe the covariance structure between times that are ''k'' years apart. Here's the data used in Mahler's example:
{| class="wikitable"
|-
|Separation in Years||Covariance (0.00001)
|-
|0||130
|-
|1||60
|-
|2||55
|-
|3||50
|-
|4||45
|-
|5||40
|-
|6||35
|-
|7||30
|-
|}
 
We should read this as <math>C(1)=60\cdot 0.00001 =0.0006</math> for example.
 
The example Mahler gives is a bit large to expect on the exam. If you solve the equations carefully you should get (in order) 11.6%, 13.4%, 17.3%, 23.8%, and 33.9% with the more recent data getting the larger weights.
 
If this is asked on the exam, it's more likely to be for two or three years. Go ahead and try solving the following problem:
: <span class="newwin">[https://www.battleacts8.ca/8/pdf/Mahler_Rating.pdf <span style="color: white; font-size: 12px; background-color: green; border: solid; border-width: 2px; border-radius: 10px; border-color: green; padding: 1px 3px 1px 3px; margin: 0px;">'''''An Example of Mahler in Ratemaking'''''</span>]</span>
 
Mahler concludes his work is directly applicable to ratemaking situations where the risk is being compared to a broader average, such as when experience rating individual risks, or calculating class or territory relativities. While in his baseball example the grand mean was always 50% (someone wins and someone loses almost always), this isn't necessarily true for insurance applications.
 
Mahler notes his estimation methods are always balanced. That is, when the dataset is not subdivided (by class or territory for instance), the estimate produces the average for the entire dataset.
 
Finally, (and perhaps most importantly) Mahler concludes when the distributional parameters are changing over time, less weight should be given to older years of data and it is more important to minimize the delay in receiving the data.
 
[https://battleacts8.ca/8/FC.php?selectString=**&filter=both&sortOrder=natural&colorFlag=allFlag&colorStatus=allStatus&priority=importance-high&subsetFlag=miniQuiz&prefix=Mahler&suffix=Credibility&section=all&subSection=all&examRep=all&examYear=all&examTerm=all&quizNum=2<span style="font-size: 20px; background-color: aqua; border: solid; border-width: 1px; border-radius: 10px; padding: 2px 10px 2px 10px; margin: 10px;">'''mini BattleQuiz 2]'''</span> <span style="color: red;">'''You must be <u>logged in</u> or this will not work.'''</span>
 
<!-- ******** BattleBar Code ******** -->
{|style="border: solid; color:lightgrey; border-radius:10px; border-width:2px; align:center;"
|-
<!-- ******** Full BattleQuiz ******** -->
|style="padding:2px"|[https://battleacts8.ca/8/FC.php?selectString=**&filter=both&sortOrder=natural&colorFlag=allFlag&colorStatus=allStatus&priority=importance-high&subsetFlag=miniQuiz&prefix=Mahler&suffix=Credibility&section=all&subSection=all&examRep=all&examYear=all&examTerm=all&quizNum=all<span style="font-size: 20px; background-color: lightgreen; border: solid; border-width: 1px; border-radius: 10px; padding: 2px 10px 2px 10px; margin: 10px;">'''Full BattleQuiz]'''</span>
<!--                                    -->
<!-- ******** Excel BattleQuiz ******** -->
<!--
|style="padding:2px"|[https://battleacts8.ca/8/FC.php?selectString=**&filter=both&sortOrder=natural&colorFlag=allFlag&colorStatus=allStatus&priority=importance-high&subsetFlag=miniQuiz&prefix=Mahler&suffix=Credibility&section=all&subSection=all&examRep=all&examYear=all&examTerm=all&quizNum=all<span style="color: red; font-size: 20px; background-color: lightgreen; border: solid; border-width: 1px; border-radius: 10px; border-color: darkblue; padding: 2px 10px 2px 10px; margin: 0px;">'''''Excel BattleQuiz]'''''</span>
-->
<!--                                    -->
<!-- ******** Excel PowerPack Files ******** -->
|style="padding:2px"|[[BattleActs_PowerPack#ppMahler| <span style="color: white; font-size: 12px; background-color: indigo; border: solid; border-width: 2px; border-radius: 10px; border-color: indigo; padding: 1px 3px 1px 3px; margin: 0px;">'''''Excel Files '''''</span>]]
<!--                                    -->
<!-- ******** Forum ******** -->
|style="padding:2px"|[https://battleacts8.ca/8/forum/index.php?p=/categories/mahler-credibility<span style="font-size: 12px; background-color: lightgrey; border: solid; border-width: 1px; border-radius: 10px; padding: 2px 10px 2px 10px; margin: 0px;">'''Forum'''</span>]
<!--                                    -->
<!-- ******** Formula Sheet ******** -->
<!--
|style="padding:2px"|[https://battleacts8.ca/8/forum/categories/mahler-credibility<span style="font-size: 12px; color: darkblue; background-color: lightblue; border: solid; border-width: 1px; border-radius: 10px; padding: 2px 10px 2px 10px; margin: 0px;">'''Formula Sheet'''</span>]
-->
|}
<span style="color: red;">'''You must be <u>logged in</u> or this will not work.'''</span>

Latest revision as of 10:48, 20 May 2024

Reading: Mahler, H. C., "An Example of Credibility and Shifting Risk Parameters"

Synopsis: Tested on the paper but not the appendices.

The paper deals with optimally combining different years of historical data when experience rating. The usual application of credibility, namely Data * Credibility + Prior Estimate * (1 - Credibility) is used. The prior estimate of credibility is normally the average experience for a class of insureds but could also be the relative loss potential of the insured to the class average.

The goal is to understand how changes in the risk classification parameters over time result in older years of data adding less credibility to experience rating than expected.

Study Tips

This reading has historically produced somewhat predictable exam questions which should be straightforward points once you've mastered the basics of the article. However, with the introduction of integrative questions it's less likely you'll get a standalone question on this article as the material can be applied to a wide range of topics.

Estimated study time: 8 Hours (not including subsequent review time)

BattleTable

Based on past exams, the main things you need to know (in rough order of importance) are:

  • How to apply and interpret the results of a Chi-squared test.
  • Be able to describe alternate tests suggested by Mahler.
  • Understand and apply the mean squared error statistic.
Questions from the Fall 2019 exam are held out for practice purposes. (They are included in the CAS practice exam.)
reference part (a) part (b) part (c) part (d)
E (2018.Fall #1) Chi-squared Test
- calculate
Mean Square Error
- use in calculation
Reinsurance Profit
- outdated
Lognormal Probabilities
- Bahnemann.Chapter5
E (2015.Fall #4) Chi-squared Test
- apply
Shifting Frequency
- alternate tests
E (2013.Fall #3) Holdout Sample
- Couret.Venter
Holdout Sample
- Couret.Venter
Sum of Squared Errors Test
- Couret.Venter
Frequency Trend
- describe test
E (2012.Fall #3) Chi-squared Test
- calculate
Shifting Frequency
- alternate tests
Full BattleQuiz Excel Files Forum

You must be logged in or this will not work.

In Plain English!

The efficiency of an experience rating plan is the reduction in expected squared error due to the use of the plan. The lower the expected squared error, the higher the efficiency.

Mahler uses data on baseball games to illustrate his points. He uses elementary statistics (binomial distribution, normal approximation) to conclude baseball teams do have significant differences between them over the years. Since they have differences between them, this means experience rating should predict future performance with some accuracy.

Next, Mahler asks if the differences in the win/loss record for a fixed team over time can be explained by random fluctuations from the same underlying distribution. He concludes the observed results cannot be explained this way, so the parameters of the distribution which describes the number of losses for a team in a year are changing over time.

One method for testing if parameters shift over time is the standard Chi-squared test. Mahler groups the data into 5-year periods (other lengths could also be used) by team and uses the Chi-squared test to measure if they could have the same underlying mean over the entire dataset. If a team didn't change its losing percentage over time, then its losing percentage should be normally distributed around its average.

The Chi-squared statistic is defined as [math]\displaystyle\sum_{i=1}^n\frac{\left(\mbox{Actual Loss}_i - \mbox{Expected Loss}_i\right)^2}{\mbox{Expected Loss}_i}[/math].

For a fixed team with n games, we have [math]n-1[/math] degrees of freedom if we have to estimate the expected loss.

Another method is to compute the average correlation between risks.

  1. Fix the distance, t, between years.
  2. Consider the set of all pairs of losing percentages which are t years apart for a fixed baseball team. Calculate the correlation coefficient for this dataset.
  3. Repeat step 2 for each team and the average the resulting set of correlation coefficients for the fixed t value.
  4. Repeat steps 2 and 3 for new t values.

For reference, the correlation coefficient is defined as [math]\frac{Cov(X,Y)}{\sigma_X\cdot\sigma_Y} =\frac{E[\left(x-\bar{x}\right)\left(y-\bar{y}\right)]}{\sigma_X\cdot\sigma_Y}[/math].

Since we are given a lot of years of baseball data, for small t values the set of pairs which are t years apart is relatively large. As t increases, the volume of data decreases. Mahler notes a non-zero observed correlation is not necessarily statistically significant as the 95% confidence interval about 0 is approximately [math]\pm0.1[/math] and this increases as the number of data points decreases.

Looking at the results shown in the paper, we observe the correlation method yields correlations which are significantly greater for years that are closer together (small t values) than those further apart. In fact, after approximately 10 years there is no distinguishable correlation between years. This leads Mahler to conclude recent years can be used to predict the future.

Let's see how the CAS might test this in an integrative question. Read through the following solution to 2018 Q1(a) and at a later date make sure you can fill in all the details from memory.

2018 Question 1 Part a

mini BattleQuiz 1 You must be logged in or this will not work.

Statement of the Problem

Let X be the quantity we want to estimate. Let Yi be various known estimates (the estimators) and let X be a weighted linear combination of the estimators, i.e. [math]X=\sum_i Z_i\cdot Y_i[/math]. The goal is to find the optimal set of weights {Zi} which produces the best estimate of the future losing percentage.

Four Simple Solutions

Mahler eventually covers six solutions to the problem. We'll follow in his footsteps and begin with four easy methods.

  1. Assume every risk is average, so set the predicted mean equal to 50%. This assumes all games have a winner/ties are negligible. It also ignores all historical data, so gives 0% credibility to the data.
  2. Assume the previous year repeats itself. This gives 100% credibility to the previous year.
  3. Credibility weight the previous year with the grand mean.
    • The grand mean is some external estimate that is independent of the data. In this case it is 50% because we assume equal likelihood of a win or loss.
    • Cases 1 and 2 are special cases of this, corresponding to 0% and 100% credibility respectively.
  4. Give equal weight to the most recent N years of data.
    • This can be further extended by calculating the credibility, Z of the N years of data and then giving each prior year weight [math]\frac{Z}{N}[/math] and weight [math]1-Z[/math] to the grand mean.

There are various choices for determining the credibility used in the fourth option. Bühlmann, Bayesian or classical limited fluctuation credibility methods could be used.

To see how the CAS can test this material in an integrative question, read through the following PDF solution to 2018 Q1(b).

2018 Question 1 Part b

At a later date, you should return to this exam question and see if you can fill in all the details from memory.

Three Criteria for Deciding Between Solutions

  1. Bühlmann and Bayesian credibility methods which minimize the mean squared error.
    • The sum of squared errors (SSE) is defined as [math]SSE=\sum_{\mbox{team}}\left(\sum_{\mbox{years}}\left(X_{est,team}-X_{actual,team}\right)^2\right)[/math]
    • The mean squared error is defined as [math]\frac{SSE}{\# \mbox{ teams}\cdot\#\mbox{ years}}[/math].
  2. Small chance of large error (classical credibility)
    • When the test is met, there is a probability P of a maximum departure from the expected mean of no more than k percent.
    • We can't directly observe the loss potential as it varies over time. Thus, averaging over time produces an incorrect result.
    • Rephrased - we're looking for credibilities which minimize [math]\mbox{Pr}\left(\frac{|X_{est,team}-X_{actual,team}|}{X_{est,team}}\gt k\%\right)[/math], where k% is some predetermined threshold.
  3. Meyers/Dorweiler Credibility:
    • Calculate the correlation using the Kendall Tau statistic (see Appendix B in the source for the definition). The first quantity is the ratio of actual losing percentage to predicted losing percentage. The second quantity is the ratio of the predicted losing percentage to the overall average losing percentage.

Section 8 of Mahler's paper applies each of these methods in turn to the baseball data. It's best given a quick skim-read. One key takeaway is found in Section 8.3: As the number of previous years, N, used as estimators increases, the credibility of the entire N year period decreases when using Bühlmann or Bayesian techniques. This is counterintuitive at first because actuaries typically expect credibility to increase as the volume of data is increased. However, Mahler points out the result is actually what we should expect given the parameters are shifting significantly over time. This effect is also seen using classical credibility but isn't seen when Meyers/Dorweiler credibility is used.

A question considered by Mahler in Section 8.5 is what constitutes a significant reduction in the mean squared error when using Bühlmann or Bayesian techniques? In the appendices (Alice:"Remember you're not tested on the appendices..."), Mahler derives a theoretical limit for the best reduction in mean squared error that can be achieved by credibility weighting two estimates. The theoretical limit derived relates to the minimum of the squared errors resulting from placing either 100% weight or 0% weight on the data. When N years of data is used under the assumption that the distributional parameters do not shift over time, the best possible reduction in squared error is [math]\frac{1}{2\cdot(N+1)}\%[/math], so for [math]N=1[/math], a 25% reduction in squared error is possible, or as stated in the paper, the credibility weighted estimate has a mean square error at least 75% of that of the mean square error resulting from giving either 100% or 0% weight to the data. Mahler uses this to conclude his measured reduction of the squared error to 83% for the baseball data set is significant.

Since there is less benefit to including older data (due to shifts in the distributional parameters), delays in receiving the data result in a reduction in the accuracy of experience rating. As expected, the mean squared error increases rapidly at first and then tapers off (increases at a decreasing rate) as the data delay widens. Consequently, the credibility of the data decreases at a decreasing rate as the time delay between the latest available data and the prediction time increases.

More General Solutions

We now cover the last two methods for assigning credibility weights.

Method 5: Single exponential smoothing

Here we assign geometric decreasing weights to n years of data. The latest year gets weight Z, the next year weight [math]Z\cdot(1-Z)[/math], and continue in this fashion until the second to oldest year gets weight [math]Z\cdot(1-Z)^{n-2}[/math] while the oldest year has weight [math](1-Z)^{n-1}[/math]. Note the weights sum to 1 (Alice: "Can you prove this using geometric series?").

Method 5 is an example of a mathematical filter. The approach of giving a set of data credibility Z and the complement weight 1 - Z is an application of this where [math]n=2[/math]. You can also apply exponential smoothing when you have k years of data and an independently derived complement. In this situation the latest year gets weight Z, the next [math]Z\cdot(1-Z)[/math], continuing until the kth year gets weight [math]Z\cdot (1-Z)^{k-1}[/math] and the complement gets weight [math](1-Z)^k[/math].

Method 6:

Generalizing method 5, we assign weights [math]Z_1,\ldots,Z_i[/math] to the first i years of data and then the balance, [math]1-\sum Z_i[/math] to the prior estimate, M.

Mahler recommends choosing M to be the grand mean, that is the mean of the entire dataset.

The more years of data used, the harder it is to determine an optimal set of weights [math]Z_1,\ldots,Z_n[/math] for this method.

Section 10 of the paper applies the credibility techniques to the 5th and 6th methods. Only the latest year of data and the previous estimate are considered for method 5. The credibilities aren't particularly sensitive for Bühlmann/Bayesian or limited fluctuation credibility. However, Meyer/Dorweiler credibility suggests optimal results are obtained using low Z values.

The more general weights used in the 6th method are assessed numerically for least squared credibility only due to the complexity of the calculations. Mahler looked at using two or three years of data and observes most of the weight is placed on the most recent year. It's recommended you skim the results in the paper.

Equations for Least Squares Credibility

Reading the conclusion of the Mahler paper makes this section feel like it should be testable. However, it didn't show up in the past exams released by the CAS.

The variance of a dataset may be split into two parts. The variance between risks which Mahler denotes by [math]\tau^2[/math], and the variance within risks which Mahler writes as [math]\delta^2+\zeta^2[/math]. Here, [math]\delta^2[/math] is the process variance and [math]\zeta^2[/math] is the parameter variance (variance due to changes in the distribution parameters over time). This sub-division of the within risk isn't necessary for computing credibility here.

Mahler applies method 6, that is, let year Xi be given weight Zi and assign the balance to the grand mean M for the entire dataset. We get the following equation which relates the expected square error between the observation and prediction: [math]\displaystyle\sum_{i=1}^N\sum_{j=1}^N Z_iZ_j(\tau^2+C(|i-j|)) - 2\sum_{i=1}^N Z_i(\tau^2+C(N+\Delta-i)) +\tau^2 +C(0)[/math]. Where [math]\tau^2[/math] is the between variance, C(k) is the covariance for data for the same risk, k years apart (within covariance), C(0) is the within variance, Δ is the length of time between the latest year of data used and the year being estimated, and N is the number of years of data being used.

This is viewed as a second degree polynomial for the credibility weights Zi. Most likely you would be given this equation on the exam if it was needed.

Differentiating the above equation leads to N equations in N unknowns so it can be solved exactly. In particular, this equation can be solved for the case when the total assigned credibility Zi is distributed uniformly across the N years. Note that this method can produce credibilities which do not necessarily decline monotonically, and can even be negative. Negative weights allow more emphasis to be given to other years with an associated reduction in expected squared error.

The credibilities may also be restricted so they sum to 1 and hence place no weight on the grand mean. The resulting system of equations may be solved using Lagrange multipliers. In the paper, Mahler's results show the most recent year receives the greatest credibility. However, a monotonic decline is not guaranteed and nor are the weights necessarily all greater than zero.

At the end of Section 11, Mahler notes his results are reasonable for estimating the least squares credibility in baseball. However, the transfer of these results to another area is dependent on the covariance structure of the proposed dataset. Details are discussed in the appendix which is out of scope for the exam.

The three types of credibility considered in Mahler's paper can be viewed as two categories. Bühlmann/Bayesian credibility and limited fluctuation credibility attempt to eliminate large errors. Bühlmann/Bayesian credibility achieves this because large errors will have a disproportionate impact on the sum of square errors. Limited fluctuation credibility does this by minimizing the errors greater than a selected threshold.

However, Meyers/Dorweiler credibility may be viewed as concerned with the pattern of the errors. Large errors are not a problem as long as there is no pattern relating them to the experience rating modifications. Mahler concludes it's important to understand what any criterion tests and that there are hazards to solely relying on any one method.

Mahler's Ratemaking Example

At the end of the paper, Mahler gives a brief overview of an application to ratemaking. The idea is to combine the five most recent annual loss ratios to produce a rate level indication, i.e. [math]N=5[/math]. We're told to assume there are three years between the latest year of data and the future average date of loss, i.e. [math]\Delta=3[/math]. By assuming a covariance structure, placing no weight on the grand mean (because in ratemaking, what is our best guess for the grand mean?), and requiring the credibilities to sum to 1, there are N equations in N unknowns to solve. The equations in Mahler's example are:

  • [math]Z_1C(0)+Z_2C(1)+Z_3C(2)+Z_4C(3)+Z_5C(4)=C(7)+\frac{\lambda}{2} [/math]
  • [math]Z_1C(1)+Z_2C(0)+Z_3C(1)+Z_4C(2)+Z_5C(3)=C(6)+\frac{\lambda}{2} [/math]
  • [math]Z_1C(2)+Z_2C(1)+Z_3C(0)+Z_4C(1)+Z_5C(2)=C(5)+\frac{\lambda}{2} [/math]
  • [math]Z_1C(3)+Z_2C(2)+Z_3C(1)+Z_4C(0)+Z_5C(1)=C(4)+\frac{\lambda}{2} [/math]
  • [math]Z_1C(4)+Z_2C(3)+Z_3C(2)+Z_4C(1)+Z_5C(0)=C(3)+\frac{\lambda}{2} [/math]

This system can be solved using standard linear algebra techniques once the C(k) are known. The C(k) describe the covariance structure between times that are k years apart. Here's the data used in Mahler's example:

Separation in Years Covariance (0.00001)
0 130
1 60
2 55
3 50
4 45
5 40
6 35
7 30

We should read this as [math]C(1)=60\cdot 0.00001 =0.0006[/math] for example.

The example Mahler gives is a bit large to expect on the exam. If you solve the equations carefully you should get (in order) 11.6%, 13.4%, 17.3%, 23.8%, and 33.9% with the more recent data getting the larger weights.

If this is asked on the exam, it's more likely to be for two or three years. Go ahead and try solving the following problem:

An Example of Mahler in Ratemaking

Mahler concludes his work is directly applicable to ratemaking situations where the risk is being compared to a broader average, such as when experience rating individual risks, or calculating class or territory relativities. While in his baseball example the grand mean was always 50% (someone wins and someone loses almost always), this isn't necessarily true for insurance applications.

Mahler notes his estimation methods are always balanced. That is, when the dataset is not subdivided (by class or territory for instance), the estimate produces the average for the entire dataset.

Finally, (and perhaps most importantly) Mahler concludes when the distributional parameters are changing over time, less weight should be given to older years of data and it is more important to minimize the delay in receiving the data.

mini BattleQuiz 2 You must be logged in or this will not work.

Full BattleQuiz Excel Files Forum

You must be logged in or this will not work.