The Credibility Report
Edition 12: Neural Reserving, Embedding Regularization, and Model Risk
Neural network approaches to reserving, embedding regularization, and model risk.
What’s in this edition
Primary-source market updates (no aggregator links) plus the latest actuarial-relevant arXiv papers (score ≥ 15, last 14 days).
📰 Headlines (primary sources)
Legal system abuse, AI, and the cloud shape the 2026 insurance outlook
This is broad market framing rather than a single-line pricing signal, but it is useful because it bundles three pressure points that keep showing up together: litigation costs, AI-driven operational change, and cloud dependency. In governance terms, it is another reminder that technology and legal risk are now entangled, not separate workstreams.
Read source → • Triple-IWomen’s History Month: Kristen Martin’s full-circle journey to leading Utica National
This is a leadership profile rather than a market datapoint, but it still has value as a window into how insurer leadership narratives are being framed in 2026. For the publication, pieces like this work best when treated as culture-and-management context rather than hard underwriting signal.
Read source → • Triple-I🔬 Research Spotlight (arXiv)
Quantum Amplitude Estimation for Catastrophe Insurance Tail-Risk Pricing: Empirical Convergence and NISQ Noise Analysis
arXiv • Score: 37 • 2026-03-10
Classical Monte Carlo methods for pricing catastrophe insurance tail risk converge at order reciprocal root N, requiring large simulation budgets to resolve upper-tail percentiles of the loss distribution. This sample-sparsity problem can lead to AI models trained on impoverished tail data, producing poorly calibrated risk estimates where insolvency risk is greatest. Quantum Amplitude Estimation (QAE), following Montanaro, achieves convergence approaching order reciprocal N in oracle queries - a quadratic speedup that, at scale, would enable high-resolution tail estimation within practical budgets. We validate this advantage empirically using a Qiskit Aer simulator with genuine Grover amplification. A complete pipeline encodes fitted lognormal catastrophe distributions into quantum oracles via amplitude encoding, producing small readout probabilities that enable safe Grover amplification with up to k=16 iterations. Seven experiments on synthetic and real (NOAA Storm Events, 58,028 records) data yield three main findings: an oracle-model advantage, that strong classical baselines win when analytical access is available, and that discretisation, not estimation, is the current bottleneck.
Open paper →A stochastic SIR model for cyber contagion: application to granular growth of firms and to insurance portfolio
arXiv • Score: 28 • 2026-03-16
This work evaluates the impact of contagious cyber-events, over a finite horizon, on firms' financial health and on a cyber insurance portfolio. Our approach builds on key empirical findings from economics and cybersecurity. In economics, firm size and growth-rate distributions are non-Gaussian and exhibit heavy tails. In cybersecurity, contagion dynamics strongly depend on firm size and environmental conditions. To capture these features, we propose a stochastic multi-group SIR model coupled with a granular model of firm growth. This framework allows us to quantify the financial impact of cyber-attacks on firms' revenues and on the insurer's portfolio. In the model, the arrival time and duration of cyber-attacks are driven by a combination of a Cox process and a Bernoulli random variable. The Cox process represents external contagion, with an intensity given by the force of infection derived from the stochastic SIR dynamics. The Bernoulli component captures contagion originating from an infected sister or subsidiary firm. Environmental variability enables stochastic scenario generation and the computation of aggregate exceedance probabilities, a standard metric in catastrophe modeling that provides insurers with immediate insight into the financial severity of an event. We apply the framework to the LockBit ransomware attacks observed between May and July 2024. For a portfolio of 2,929 firms located in Ile-de-France, the model predicts that, with 50% probability, the insurer will need to compensate losses equivalent to up to two days of revenue over a 100-day cyber incident.
Open paper →Robust Investment-Driven Insurance Pricing and Liquidity Management
arXiv • Score: 28 • 2026-03-19
This paper develops a dynamic equilibrium model of the insurance market that jointly characterizes insurers' underwriting, investment, recapitalization, and dividend policies under model uncertainty and financial frictions. Competitive insurers maximize shareholder value under a subjective worst-case probability measure, giving rise to liquidity-driven underwriting cycles and flight-to-quality behavior. While an equilibrium typically fails to exist in such dynamic liquidity management framework with external financial investment, we show that incorporating model uncertainty restores equilibrium existence under plausible parameter conditions. Moreover, the model uncovers a novel relationship between the correlation of insurance and financial market risks and the equilibrium insurance price: negative loadings may emerge when insurance gains and financial returns are positively correlated, contrary to conventional intuition.
Open paper →Pricing Derivatives under Self-Exciting Dynamics: A Finite-Difference and Transform Approach
arXiv • Score: 27 • 2026-03-13
We consider the pricing of derivatives written on accumulated marks, such as weather derivatives or aggregate loss claims, using a self-exciting marked point process. The jump intensity mean-reverts between events and increases at jump times by an amount proportional to the mark. The resulting state process, where the variable $U_t$ accumulates jump magnitudes, is a piecewise deterministic Markov process (PDMP). We derive the discounted pricing equation as a backward partial integro-differential equation (PIDE) in two spatial dimensions. To overcome the dimensionality, we propose an exponential (Laplace/Fourier) transform in the accumulated mark variable, which diagonalizes the translation operator and reduces the pricing problem to a family of one-dimensional PIDEs in the intensity variable along a Bromwich contour. For Gamma-mixture mark laws (under actuarial or Esscher-tilted measures), the nonlocal jump term is efficiently approximated by generalized Gauss--Laguerre quadrature. We solve the reduced PIDEs backward in time using a monotone IMEX finite difference scheme (implicit upwind drift and discounting, explicit jump operator) and recover option prices via numerical inversion. We provide a rigorous, term-by-term global error bound covering time and space discretization, quadrature, interpolation, and boundary effects, supported by numerical experiments and Monte Carlo benchmarks.
Open paper →How Proxy Race Distorts Regression-Based Fairness Audits
arXiv • Score: 26 • 2026-03-17
Proxy-based race inference is increasingly used to conduct fairness assessments when protected-class data are unavailable or legally restricted -- most prominently in U.S. fair-lending enforcement, and now explicitly contemplated in emerging insurance regulation, including Colorado's draft SB21-169 testing framework and New York's Insurance Circular Letter No. 7. Despite this growing regulatory relevance, little is known about how standard regression-based discrimination analyses behave when race is measured with error through proxies such as Bayesian Improved Surname Geocoding (BISG) or Bayesian Improved First Name and Surname Geocoding (BIFSG). This paper studies the consequences of using proxy-imputed race as a categorical regressor in regression-based fairness assessments. Treating proxy race as a categorical covariate subject to misclassification, we show that proxy-based coefficients become weighted mixtures of true group effects, systematically shrinking estimated disparities toward the majority group -- even when overall classification accuracy is high. Empirically, using a linked North Carolina voter-insurance dataset with self-reported race and ZIP-level auto insurance premiums, we demonstrate two mechanisms through which it distorts inference: (i) the intrinsic mixing of group effects implied by misclassification, and (ii) structured errors that vary with ZIP-level racial composition and socioeconomic conditions and remain correlated with pricing residuals after controls. As a result, regression-based disparity estimates can be attenuated or amplified relative to analogous analyses based on self-reported race. Our findings caution against treating proxy race as a plug-in substitute in regulatory testing and highlight design implications for proxy-based audit frameworks in insurance and other high-stakes domains.
Open paper →Robust Investment-Driven Insurance Pricing under Correlation Ambiguity
arXiv • Score: 25 • 2026-03-19
As insurers increasingly behave like financial intermediaries and actively participate in capital markets, understanding the dependence structure between insurance and financial risks becomes crucial for insurers' operations. This paper studies dynamic equilibrium insurance pricing when insurers face ambiguity about the correlation between insurance and financial risks and optimally choose underwriting and investment strategies under worst-case beliefs. Correlation ambiguity can generate multiple equilibrium regimes. Contrary to conventional intuition, we find ambiguity does not necessarily increase insurance prices nor reduce insurers' utility.
Open paper →A Portfolio-Anchored Frequency-Severity Risk Index for Trip and Driver Assessment Using Telematics Signals
arXiv • Score: 23 • 2026-03-16
In this paper, we propose a novel frequency-severity joint trip-level risk index that combines the frequency of abnormal driving patterns with a severity component reflecting how extreme such behavior is relative to a portfolio-level baseline. Severity is quantified through an inverse-probability penalty that increases with the rarity of observed tail extremes, rather than being interpreted as a claim size. Based on high-frequency telematics data, we construct a multi-scale representation of longitudinal acceleration using the maximal overlap discrete wavelet transform (MODWT), which preserves localized driving patterns across multiple time scales. To capture severity as tail rarity, we model the portfolio distribution using a Gaussian-Uniform mixture with a layered tail structure, where Gaussian components describe typical driving behavior and the tail is partitioned into ordered severity layers that reflect increasing extremeness. We develop a likelihood-based estimation procedure that makes inference feasible for this mixture model. The resulting severity layers are then used to construct multi-layer tail counts (MLTC) at the trip level, which are modeled within a Poisson-Gamma framework to yield a closed-form posterior risk index that jointly reflects frequency and severity. This conjugate structure naturally supports sequential updating, enabling the construction of dynamically evolving driver-level risk profiles. Using the UAH-DriveSet controlled dataset, we demonstrate that the proposed index enables reliable discrimination across behavioral driving states, identification of high-risk trips, and coherent ranking of drivers, yielding a purely behavior-driven risk measure suitable for actuarial ratemaking and potentially mitigating fairness concerns associated with traditional covariates.
Open paper →Gradient Boosting for Spatial Panel Models with Random and Fixed Effects
arXiv • Score: 19 • 2026-03-15
Due to the increase in data availability in urban and regional studies, various spatial panel models have emerged to model spatial panel data, which exhibit spatial patterns and spatial dependencies between observations across time. Although estimation is usually based on maximum likelihood or generalized method of moments, these methods may fail to yield unique solutions if researchers are faced with high-dimensional settings. This article proposes a model-based gradient boosting algorithm, which enables estimation with interpretable results that is feasible in low- and high-dimensional settings. Due to its modular nature, the flexible model-based gradient boosting algorithm is suitable for a variety of spatial panel models, which can include random and fixed effects. The general framework also enables data-driven model and variable selection as well as implicit regularization where the bias-variance trade-off is controlled for, thereby enhancing accuracy of prediction on out-of-sample spatial panel data. Monte Carlo experiments concerned with the performance of estimation and variable selection confirm proper functionality in low- and high-dimensional settings while real-world applications including non-life insurance in Italian districts, rice production in Indonesian farms and life expectancy in German districts illustrate the potential application.
Open paper →Association of Progressive PPFE and Mortality in Lung Cancer Screening Cohorts
arXiv • Score: 18 • 2026-03-10
Open paper →Feynman-Kac Derivatives Pricing on the Full Forward Curve
arXiv • Score: 17 • 2026-03-12
Open paper →Starting Off on the Wrong Foot: Pitfalls in Data Preparation
arXiv • Score: 15 • 2026-03-18
Open paper →Until next time—stay credible.
— The Credibility Report
Edition 012 | Prepared March 22, 2026 (UTC)