Processing math: 100%
De Virgilis, Marco, and Giulio Carnevale. 2025. “Applications of Gaussian-Inverse Wishart Process Regression Models in Claims Reserving.” Variance 18 (June).
Download all (62)
  • Figure 1. Covariance functions.
  • Figure 2. Sample draws.
  • Figure 3. GPR: Prior functions.
  • Figure 4. GPR: Posterior functions.
  • Figure 5. Graphical representation of claims data.
  • Figure 6. Graphical representation of claims data.
  • Figure 7. Triangle data.
  • Figure 8. EMC PA—total reserve distribution.
  • Figure 9. EMC OL—total reserve distribution.
  • Figure 10. CC WC—total reserve distribution.
  • Figure 11. PRI MM—total reserve distribution.
  • Figure 12. HNI CA—total reserve distribution.
  • Figure 13. EMC PA—correlation distribution.
  • Figure 14. EMC OL—correlation distribution.
  • Figure 15. CC WC—correlation distribution.
  • Figure 16. PRI MM—correlation distribution.
  • Figure 17. HNI CA—correlation distribution.
  • Figure 18. Observed claim amounts development patterns.
  • Figure 19. Observed claim LR development patterns.
  • Figure 20. Estimated claim amounts development patterns.
  • Figure 21. Estimated claim LR development patterns.
  • Figure 22. Loss development patterns.
  • Figure 23. EMC PA—parameters posterior distributions.
  • Figure 24. EMC PA—parameters posterior distributions.
  • Figure 25. EMC PA—parameters posterior distributions.
  • Figure 26. EMC PA—parameters posterior distributions.
  • Figure 27. EMC OL—parameters posterior distributions.
  • Figure 28. EMC OL—parameters posterior distributions.
  • Figure 29. EMC OL—parameters posterior distributions.
  • Figure 30. EMC OL—parameters posterior distributions.
  • Figure 31. CC WC—parameters posterior distributions.
  • Figure 32. CC WC—parameters posterior distributions.
  • Figure 33. CC WC—parameters posterior distributions.
  • Figure 34. CC WC—parameters posterior distributions.
  • Figure 35. PRI MM—parameters posterior distributions.
  • Figure 36. PRI MM—parameters posterior distributions.
  • Figure 37. PRI MM—parameters posterior distributions.
  • Figure 38. PRI MM—parameters posterior distributions.
  • Figure 39. HNI CA—parameters posterior distributions.
  • Figure 40. HNI CA—parameters posterior distributions.
  • Figure 41. HNI CA—parameters posterior distributions.
  • Figure 42. HNI CA—parameters posterior distributions.
  • Figure 43. EMC PA—correlation posterior distributions.
  • Figure 44. EMC PA—correlation posterior distributions.
  • Figure 45. EMC PA—correlation posterior distributions.
  • Figure 46. EMC PA—correlation posterior distributions.
  • Figure 47. EMC OL—correlation posterior distributions.
  • Figure 48. EMC OL—correlation posterior distributions.
  • Figure 49. EMC OL—correlation posterior distributions.
  • Figure 50. EMC OL—correlation posterior distributions.
  • Figure 51. CC WC—correlation posterior distributions.
  • Figure 52. CC WC—correlation posterior distributions.
  • Figure 53. CC WC—correlation posterior distributions.
  • Figure 54. CC WC—correlation posterior distributions.
  • Figure 55. PRI MM—correlation posterior distributions.
  • Figure 56. PRI MM—correlation posterior distributions.
  • Figure 57. PRI MM—correlation posterior distributions.
  • Figure 58. PRI MM—correlation posterior distributions.
  • Figure 59. HNI CA—correlation posterior distributions.
  • Figure 60. HNI CA—correlation posterior distributions.
  • Figure 61. HNI CA—correlation posterior distributions.
  • Figure 62. HNI CA—correlation posterior distributions.

Abstract

Gaussian processes are stochastic processes based on the normal distribution (i.e., collections of normal random variables indexed by a mathematical set). In the context of probability theory and statistics, these processes are well-known and well-behaved objects that have been extensively explored and used. On the other hand, Gaussian process regression (GPR) is a relatively lesser known procedure based on Gaussian stochastic processes that can be implemented in the context of machine learning for both regression and classification problems.

GPR can be defined as a supervised nonparametric machine learning technique stemming from the Bayesian field. This is a relatively novel technique in the field of machine learning and statistics, only used in some geostatistics applications under the name of kriging. Moreover, its use is nearly unknown in the actuarial sciences.

We propose a novel procedure based on the inverse Wishart distribution that has not been explored in the context of actuarial modeling. This affords us the advantage of exploring the full correlation between claim amounts, observed and expected, a piece of crucial information in many reserving and capital requirements analyses.

Additionally, our goal is to provide actuarial practitioners with an easy explanation of GPR workings alongside a worked example and comparisons with traditional methodologies. In a world where the relevance of advanced analytics is ever growing, GPR models can represent another powerful instrument in the actuary’s toolkit.

A main feature of GPR is the ability to fit functions on a set of observations and produce predictions with uncertainty intervals around them. Because of this feature, GPR techniques can be ideal candidates to extend traditional stochastic reserving techniques to meet the ever-growing need of practitioners, regulators, and rating agencies to be able to quantify reserve variability.

We believe this work represents a starting point for further research while offering an initial understanding of the topic to anyone interested. Furthermore, we provide all the data and tools needed to replicate the results shown. For this reason, we adopt only open-source software and publicly available datasets. In particular, we use the publicly available NAIC Schedule P dataset to collect loss triangle and premium data. The models’ implementation is performed in R, and the code of the project is made available through a GitHub repository.