Skip to content

On contemporary mortality models for actuarial use

Stephen Richards’ and Angus Macdonald’s paper ‘On Contemporary Mortality Models for Actuarial Use’ (due to be discussed at the Institute of Actuaries on 24 October 2024) makes the case for the following in mortality experience analysis1:

  1. Use individual data if at all possible.

  2. Work in continuous time and use instantaneous mortality rates, i.e. μ rather than q.

  3. Consider mortality experience data as comprising a series of Bernoulli trials over infinitesimally small time periods.

The paper could be read as a polemic against actuaries who can’t help but think in terms of q and whose first instinct is to group all time and age-dependent data on annual grids. Which is fine by me – I agree with the thrust of the paper and, in particular, the above three points.

So, having welcomed the paper, I do have a few observations…

Observations

1. Are actuaries really migrating from q to µ?

The paper’s introduction states that ‘µx is usurping qx in actuarial work.’ While I’d like this to be true, I can’t help but wonder whether it’s wishful thinking.

Trying to switch the CMI Mortality Projections Model from q to µ2

As chair of the CMI’s Mortality Projections Committee, I oversaw the revision of its Projections Model in 2016. I was keen to remove all q-based aspects inherited from the previous approach, but – relevant in this context – all changes from q to µ were opposed by at least some actuaries.

The following two changes were objected to3, but did gain majority support in consultation and make it into the final version:

  • Changing the ‘long-term rate of mortality improvement’ (LTR) from q to µ-based.

  • Changing the projections part of the model to work in directly in terms of µ rather than q.

The only q-related aspect remaining was that the improvement factors output by the Model are expressed in terms of q-multipliers, but the committee itself felt that changing this to µ was a step too far4.

  • Many actuaries simply prefer q (in the same way that they prefer annually-compounded interest rates to the force of interest). See the boxout for a documented example.

  • Like it or not, q is the current lingua franca in the sense that ‘110% of S4PxA’ without further qualification means scale q by 110%5.

  • In my experience, most actuarial valuation systems work with q, data grouped by age and annual time grids6.

  • Even within specialised mortality experience analysis environments, where, as the paper notes, there are clear benefits to working in continuous time, I’ve encountered plenty of systems that use q.

2. Generalising survival likelihood factorisation

Section 3.3 of part II of the paper notes that survival likelihood can be factorised over sub-time intervals and that likelihoods for contiguous intervals can validly be multiplied together.

It’s worth highlighting that there is a more general concept at play here, which is that (a) the count of actual deaths, (b) the sum over individuals of integrals of mortality rate over exposure periods and (c) log-likelihood are all measures over experience data7. This means that they

  • give the same result when summed over a partitioned experience dataset regardless of how it is partitioned, and

  • in particular, when summing over multiple experience datasets, there is no need for experience time periods to be contiguous – the sole requirement is that experience datasets do not intersect.

The above are two points are intuitively well understood by practitioners8; my point is that the maths – suitably defined – aligns with this.

3. Likelihood vs log-likelihood

The technical part II of paper focuses on likelihood, i.e. probability, but I’d argue that log-likelihood is both more tractable and more fundamental. These are subjective concepts, so this is personal opinion, not a matter of right or wrong.

Let’s take tractability first.

  • The paper is littered9 with product signs, i.e. \(\prod\), arising from the use of probabilities. In contrast, working in term of log-likelihood results in standard integrals and makes linearity self-evident where present10. As a case in point, I didn’t find the continued use of the Volterra product integral, i.e.

    \[\prod_a^b\Big(1-f(t)\,\text{d}t\Big)\]

    enlightening. Given that they are technically equivalent, I’d suggest it’s simpler to explain survival probabilities once, then take logs and work in terms of standard integrals, i.e.

    \[-\int_a^b \! f(t) \,\text{d}t\]
  • In terms of what is actually practical (which is a form of tractability), implementing mortality experience analysis solely using probability will typically fail because of numerical underflow. So, if you believe that code should map to concepts, the concept you want is log-likelihood.

Turning to fundamentality:

  • A key rationale for fitting models by maximising log-likelihood and selecting between them using the AIC is information theory, the core unit of which is information content, which in turn is defined as (negative) log-likelihood.

  • The paper repeatedly refers to the concept of infinitesimal Bernoulli trials. I couldn’t agree more, but the usual mathematical treatment of summed infinitesimals requires something akin to dL to make sense, which works only if L is log-likelihood rather than plain likelihood.

  • Observation 2 above (that there is a more general concept at play than just being able to factorise the survival probabilities over contiguous time periods) is best expressed in terms of measures, which means using log-likelihood.

Conclusion

Stephen Richards’ and Angus Macdonald’s paper is to be welcomed for making the case for using continuous time and instantaneous rates in mortality experience analysis.

If it results in even one actuary thinking before diving in with a q-based approach, then that’s a good thing.


  1. The paper’s title suggests it covers all mortality modelling, but its sole focus is the analysis of portfolio mortality experience. So, for instance, mortality projection modelling is omitted entirely. This doesn’t detract from the paper’s contents – mortality experience analysis is an important topic in itself. 

  2. Technically the CMI Mortality Projections Model is defined in terms of m, not µ, but it is m rather than q that maps straightforwardly to µ

  3. See Q 6.1 and Q 6.2 in WP 93

  4. This is not a criticism – the CMI has an obligation to its users to consider the stability and consistency of its outputs. 

  5. And gloss over how to treat ages for which \(q_x\ge1\)

  6. There are valid practical arguments for, as well as against, this approach. 

  7. Log-likelihood is a signed measure

  8. For instance, no-one would bat an eyelid (in a technical sense) if mortality experience data are analysed excluding pandemic-affected periods, resulting in non-contiguous time periods. 

  9. A quick search suggests there are over 60 product signs, i.e. \(\prod\), in the paper. 

  10. It is for good reason that logarithms were originally described as mirifici