Your 5-year NCD customer has a small scrape. The repair is £250. They do not call you. They fix it themselves, quietly, and protect their 65% discount.

This is not a rounding error. It is a rational, optimal decision — one that your pricing model has no adjustment for.

The result is that your NCD relativities are estimated from claim data that is systematically missing small losses at high discount levels. The frequency model you rely on is fit on suppressed data. The NCD coefficients absorb the suppression signal and you call it good. Every UK motor pricing team we are aware of does this. Two papers published in January 2026 make clear how wrong it is.


The hunger for bonus

The phenomenon has a name in the actuarial literature: hunger for bonus (Jean Lemaire coined it in 1977 — “La Soif du Bonus”, ASTIN Bulletin). The concept is simple. Under a bonus-malus system, reporting a fault claim triggers a premium increase that lasts for years. If that future cost exceeds the claim amount, a rational policyholder does not claim.

The calculation is not difficult. Take a 5-year NCD holder paying £350 on a £1,000 base (65% discount). A fault claim steps them back to 3-year NCD (50% discount, £500). That is £150 per year more in premium. Rebuilding to 5-year NCD takes two years. Discounting at ~5%:

PV(penalty) ≈ £150 × (0.95 + 0.90) ≈ £278

The rational threshold: do not claim for losses below £278. Self-insure anything up to roughly three months of the premium saving.

Work through the same calculation at different points on the UK NCD ladder.

NCD years Typical premium (£1,000 base) Step-back to Annual increase Years to rebuild Rational threshold (approx.)
9 £300 (70%) 7 yr (65%) £50 2 ~£93
5 £350 (65%) 3 yr (50%) £150 2 ~£278
3 £500 (50%) 1 yr (30%) £200 2 ~£370
1 £700 (30%) 0 yr (0%) £300 1 ~£285

The pattern is not monotone. Thresholds peak in the middle of the ladder — around the 3-to-5-year range — and fall at both extremes. At 9-year NCD the absolute premium difference is small (the 70% cap means a step-back to 65% is only £50/year). At 1-year NCD the annual increase is large but the rebuild time is short — one claim-free year gets you back to 1-year NCD — so the present value of the penalty is lower than at mid-ladder.

The 3-year NCD holder faces the highest rational retention threshold: roughly £370 on a £1,000 base premium. That is an amount large enough to cover most minor own-damage incidents.

Lemaire’s full dynamic programming treatment accounts for the stochastic claim history during the rebuilding period and multiple future periods, so the real numbers differ from these back-of-envelope figures. But the direction is clear, and the magnitudes are not trivial. At 5-year NCD, the rational threshold exceeds a quarter of the annual base premium.


What this does to your frequency model

Here is the problem as a pricing actuary should see it.

True claim frequency at NCD class n is λ_n. But a policyholder in class n reports a claim only if its size exceeds the retention threshold b*_n. So observed frequency is:

λ_obs_n = λ_n × P(Y > b*_n)

If the severity distribution is Gamma with mean £1,500 and P(Y > £278) ≈ 0.65 for a 5-year NCD holder, then observed frequency is roughly 65% of true frequency. A 35% underestimate.

When you fit a Poisson GLM on reported claims with NCD as a categorical factor, you are fitting on λ_obs, not λ. The NCD coefficients you get back are descriptively accurate — they correctly predict future reported claim counts conditional on current NCD class. But they are structurally wrong as a measure of true risk. The 5-year NCD coefficient is too low because it has been estimated from data where 35% of claims never happened, from the model’s perspective.

The cross-subsidy runs from lower-NCD classes (whose observed frequency is closer to their true frequency — a 0-year NCD holder has a rational threshold near £0, because they have nothing to protect) toward the mid-ladder classes where suppression is heaviest. Your current 3-to-5-year NCD customers are likely undercharged relative to their actual risk. Drivers at the extremes — new to insurance or just past a claim — are cross-subsidising them.

The bias compounds another way: because small claims are suppressed, the claims that are reported skew toward larger losses. This inflates your observed mean severity at mid-ladder NCD classes. If your frequency and severity models interact, the severity distortion bleeds back.


The competitive layer: Liang et al. (2026)

This would be a contained, if underappreciated, calibration problem if it happened in isolation. But Liang, Zhang, Zhou, and Zou (arXiv:2601.12655, January 2026) show it is worse than that — because insurer competition turns it into a game with an equilibrium that most pricing teams are not solving for.

Their paper is the first analysis of strategic underreporting under oligopolistic competition. Prior work (including Lemaire’s) assumed a single insurer. Liang et al. add a second insurer and a policyholders’ switching choice, and show that the system has a Nash equilibrium with three interlocking conditions: each insurer’s premiums are profit-maximising given the other’s premiums; policyholders’ reporting strategy is optimal given both insurers’ premiums; and the stationary distribution of policyholders across NCD classes is consistent with the reporting strategy.

The key theorem (Theorem 3.1) establishes that the optimal reporting threshold is insurer-independent — it depends on the policyholder’s NCD class and the premium levels, but not on which specific insurer they are with. A 5-year NCD holder faces the same rational threshold at Insurer A as at Insurer B, as long as both charge similar premiums at that class. This simplifies the equilibrium considerably.

In their two-class numerical example (p₀ = 0.9, loss ~ Gamma(1.2, 0.0085), penalty ratio κ = 1.25, discount factor δ = 0.97), the Nash equilibrium has the more attractive insurer charging approximately 7% more than its competitor: θ₁ ≈ 35.83, θ₂ ≈ 33.45. The preferred insurer can exploit its competitive position in exactly the way you would expect: brand loyalty creates pricing power, and price-sensitivity governs how far premiums can diverge before the gap closes.

The practical implication is harder to digest. The standard UK workflow — fit GLM on reported claims, update relativities, repeat — does not converge to the Nash equilibrium. It converges to whatever fixed point the iterative refit finds, which is not the same thing. If you are not explicitly solving the three-way fixed point (premiums → reporting threshold → stationary distribution → expected profit → best-response premiums), your NCD rates are not in equilibrium. They are, in Liang et al.’s framework, a locally stable approximation to the wrong answer.

There is also an adverse selection dynamic. If one insurer corrects for the suppression bias and another does not, the correcting insurer prices mid-ladder NCD correctly (higher) while the naive insurer prices it cheap. Mid-ladder customers migrate to the cheaper book. The correcting insurer loses the good-looking business and keeps a portfolio that selected adversely on every other variable. The market dynamics push both insurers toward the naive equilibrium unless correction is adopted broadly.


What the FCA has and has not said

GIPP (PS21/5, effective January 2022) eliminated price walking — the renewal premium for existing customers must be no higher than the equivalent new business price through the same channel. The FCA’s 2025 evaluation paper (EP25/2) confirmed GIPP has held: the new-to-renewal gap has not returned, even as claims costs rose 49% between 2022 and 2024 (expected cost per policy from £92 to £138).

Consumer Duty, effective July 2023, requires firms to deliver fair value. The Price and Value outcome could, in principle, apply to NCD pricing methodology: if your NCD relativities are estimated from suppressed data and you are systematically undercharging mid-ladder customers while overcharging lower-NCD drivers, a fair-value argument is at least arguable. No FCA guidance exists on this point — the regulator has not addressed BMS calibration methodology, and there is no immediate regulatory obligation to correct for underreporting bias.

Protected NCD sits in more active regulatory territory. The FCA has signalled scrutiny of PNCD under Consumer Duty’s Consumer Understanding outcome, because the product is routinely misunderstood: consumers believe it protects their total premium, not just the discount percentage. After a fault claim, the NCD percentage is preserved but the base premium can be repriced upward — which can erase much of the apparent protection at first renewal. This is a live compliance question, separate from the pricing calibration issue.


What you can actually do about it

No off-the-shelf package handles the full correction pipeline — not in R, not in Python. The actuar package in R provides the credibility machinery; insurancerating has the GLM workflow with NCD as a factor; Loss Data Analytics (Chapter 12 at openacttexts.github.io) has worked BMS Markov chain examples. But the correction from observed to true frequency, calibrated to your own NCD ladder, has to be built.

We think a working first-pass is achievable in a few days.

Implement Lemaire’s algorithm for your own ladder. This is roughly 50 lines of NumPy. Inputs are your NCD discount percentages, your step-back rules, your average base premium by segment, and a severity distribution. Output is a retention table b*_n for each NCD class. The severity distribution you already have — use the Gamma parameters from your existing severity GLM, stratified by cover type.

Compute the censoring correction per class. For each NCD class, p_n = P(Y > b*_n) under your severity distribution. Compare λ_obs / p_n to λ_obs. If the correction factors are within 5% across all classes, the bias is not material for your portfolio and you can stop. Our guess is they will not be.

Re-examine your NCD relativities against corrected frequencies. You are not necessarily re-fitting the full GLM at this stage — you are asking whether the size of the bias is consistent with how much your current 5-year relativities differ from 0-year. If your observed 5-year relativity is 0.60 but the corrected frequency implies 0.68, that is meaningful. You are undercharging that segment by roughly 13%.

Look at your FNOL withdrawal data. Claims that are opened and then withdrawn — the policyholder called to register a claim, then settled it themselves — are your cleanest signal for calibrating the retention threshold empirically. If your claims system captures these with an indicative reserve or loss estimate at FNOL, the size distribution of withdrawals at each NCD class directly estimates b*_n. This data almost certainly exists somewhere in your claims system. Most teams have never looked at it this way.

The full competitive equilibrium from Liang et al. — where you jointly solve for your own premiums and competitor premiums given optimal policyholder behaviour — requires market-level competitor premium data at NCD class level. That means PCW aggregator data or industry sharing arrangements, and it is a larger project. But the single-insurer correction (the first three steps above) does not require any of that, and it quantifies the first-order bias from internal data alone.


The open questions

Liang et al. prove Nash equilibrium existence for two NCD classes. Most UK ladders have nine or more. The extension to N classes is conjectured but not proved, and numerical fixed-point approaches are needed for practical implementation.

The framework also assumes homogeneous policyholders. In practice, the interaction between a priori risk heterogeneity (the core pricing problem your GLM is trying to solve) and strategic reporting behaviour is unmodelled in the published literature. A high-frequency policyholder has a different reporting threshold calculation from a low-frequency one — their expected future premium is different, their rebuilding path is different. This is the problem Norberg’s credibility approach partially addresses, but the underreporting correction has not been integrated into the Bayesian updating framework.

The empirical question is also open: no published estimate of the hunger-for-bonus threshold exists for UK motor using UK data. Abbring, Chiappori, and Zavadil (SSRN, 2008) used Dutch longitudinal data and found evidence consistent with strategic non-reporting, but the Dutch BMS is more punitive than the UK standard and the numbers do not transfer directly.

Finally, telematics complicates the picture in a way the literature has not addressed. A connected car policy may detect an impact event, meaning the policyholder cannot suppress it cleanly. If your telematics portfolio has different NCD distributions from your standard book, your NCD relativities should not be shared across the two.


The bottom line

Policyholders with 3-to-5 years of NCD are rationally self-insuring losses up to several hundred pounds. Your observed claim frequency at those classes understates true frequency by a third or more. You are pricing those customers too cheaply relative to their actual risk, and your lower-NCD customers are funding the difference.

This is not a new finding — Lemaire documented it in 1977 — but the UK industry has not corrected for it, there is no regulatory requirement to do so, and no standard tool exists. Liang et al. (2026) show that the problem is structurally worse in a competitive market: the naive GLM re-fit does not converge to the right equilibrium.

The full correction is not a one-week project. Estimating the size of the bias is. If you have a pricing model and a severity distribution, you can quantify whether your NCD relativities have a material bias problem in a few days of work. We think you should.

Back to all articles