commit a392e8eab8f0987121cff20e5c875200210a6511 Author: Ralph pricing.frontier@gmail.com Date: Fri Mar 27 04:59:56 2026 +0000

blog: resolve merge conflicts + add benchmark write-ups

Resolve 3 merge conflicts in "Does X Actually Work" series (take fix branch).
Add 2 new posts: fairness audit on ausprivauto0405 and drift detection on freMTPL2.
All API references verified against library source code.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>

diff –git a/_posts/2026-03-27-fairness-audit-real-insurance-data-ausprivauto.md b/_posts/2026-03-27-fairness-audit-real-insurance-data-ausprivauto.md new file mode 100644 index 0000000..36924dc — /dev/null +++ b/_posts/2026-03-27-fairness-audit-real-insurance-data-ausprivauto.md @@ -0,0 +1,205 @@ +— +layout: post +title: “What a Real Fairness Audit Finds: Gender Bias Testing on 67,856 Motor Policies” +date: 2026-03-27 +categories: [fairness, pricing, techniques] +tags: [fairness-audit, proxy-discrimination, gender-bias, consumer-duty, equality-act-2010, fca, multicalibration, indirect-discrimination, insurance-fairness, motor, catboost] +description: “We ran insurance-fairness against ausprivauto0405 — a real Australian motor dataset with an explicit Gender field. Here is what FairnessAudit, MulticalibrationAudit, and IndirectDiscriminationAudit found, and what that means for how you audit UK pricing models.” +— + +Most fairness audit walkthroughs use toy data. An artificial dataset where gender neatly predicts claims, so the audit catches the problem, everyone looks clever, and you learn nothing about what happens on a real book. + +We ran insurance-fairness against a real published dataset — 67,856 Australian private motor insurance policies with an explicit Gender field — to see what these tools actually find when you throw messy real data at them. The results were instructive, and not in the way you might expect. + +A note before we start: ausprivauto0405 is Australian motor data from 2004-05, sourced from the CASdatasets R package (Dutang & Charpentier, 2024). We used it because it has a real Gender variable and real claim outcomes — not because it tells us anything about the UK market. Everything here is methodology validation. Do not treat any numbers in this post as benchmarks for Equality Act 2010 compliance or FCA Consumer Duty reporting. + +— + +## The setup + +The dataset has 67,856 policies after cleaning, with columns for vehicle value, vehicle age, vehicle body type, driver age, and claim occurrence. We split off 25% as a test set, then fitted a CatBoost classifier on the training set to predict claim occurrence — with Gender deliberately excluded from the model features. + +The point is to replicate production reality: your UK model does not use gender as a rating factor (it has been prohibited since Test-Achats in 2012 and the subsequent ABI guidance), but gender may still be correlated with the factors you do use. The question is whether that residual correlation creates discriminatory outcomes. + +```python +from insurance_fairness import (

Back to all articles