Hostname: page-component-89b8bd64d-z2ts4 Total loading time: 0 Render date: 2026-05-10T10:41:39.978Z Has data issue: false hasContentIssue false

Telematics combined actuarial neural networks for cross-sectional and longitudinal claim count data

Published online by Cambridge University Press:  14 February 2024

Francis Duval*
Affiliation:
Département de mathématiques, Université du Québec à Montréal, Montréal, H2X 3Y7, Québec, Canada
Jean-Philippe Boucher
Affiliation:
Département de mathématiques, Université du Québec à Montréal, Montréal, H2X 3Y7, Québec, Canada
Mathieu Pigeon
Affiliation:
Département de mathématiques, Université du Québec à Montréal, Montréal, H2X 3Y7, Québec, Canada
*
Corresponding author: Francis Duval; Email: francisduval37@gmail.com
Rights & Permissions [Opens in a new window]

Abstract

We present novel cross-sectional and longitudinal claim count models for vehicle insurance built upon the combinedd actuarial neural network (CANN) framework proposed by Wüthrich and Merz. The CANN approach combines a classical actuarial model, such as a generalized linear model, with a neural network. This blending of models results in a two-component model comprising a classical regression model and a neural network part. The CANN model leverages the strengths of both components, providing a solid foundation and interpretability from the classical model while harnessing the flexibility and capacity to capture intricate relationships and interactions offered by the neural network. In our proposed models, we use well-known log-linear claim count regression models for the classical regression part and a multilayer perceptron (MLP) for the neural network part. The MLP part is used to process telematics car driving data given as a vector characterizing the driving behavior of each insured driver. In addition to the Poisson and negative binomial distributions for cross-sectional data, we propose a procedure for training our CANN model with a multivariate negative binomial specification. By doing so, we introduce a longitudinal model that accounts for the dependence between contracts from the same insured. Our results reveal that the CANN models exhibit superior performance compared to log-linear models that rely on manually engineered telematics features.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2024. Published by Cambridge University Press on behalf of The International Actuarial Association
Figure 0

Table 1. Variables of the contract dataset

Figure 1

Figure 1. Number of contracts per vehicle.

Figure 2

Table 2. Extract from the telematics dataset. Dates are displayed in the yyyy-mm-dd format. The actual VINs have been hidden for privacy purposes

Figure 3

Table 3. Data partitioning

Figure 4

Figure 2. CANN architecture for the Poisson specification. The MLP’s preactivation output value $a_1^{(4)}$ is added to the log-linear model’s preactivation output value $\langle \boldsymbol{x}, \boldsymbol{\beta} \rangle$ before being transformed with the softplus function $\zeta(\cdot)$. The resulting $\mu$ value is compared to the ground truth y using Poisson deviance loss. The architecture shown employs a 3-hidden-layer MLP but can be customized with any number of layers.

Figure 5

Algorithm 1: Parameter estimation procedure – Poisson CANN model

Figure 6

Figure 3. CANN architecture for the negative binomial specification. The MLP’s preactivation output value $a_1^{(4)}$ is added to the log-linear model’s preactivation output value $\langle \boldsymbol{x}, \boldsymbol{\beta} \rangle$ before being transformed with the softplus function $\zeta(\cdot)$ to obtain the $\mu$ value of the negative binomial distribution. The $\phi$ value is obtained by transforming a real-valued parameter $w_\phi$ through the softplus function. The resulting parameters $\mu$ and $\phi$ are then compared to the ground truth y using negative binomial deviance loss. The architecture shown employs a 3-hidden-layer MLP but can be customized with any number of layers.

Figure 7

Figure 4. CANN architecture for the MVNB specification. The MLP’s preactivation output value $a_1^{(4)}$ is added to the log-linear model’s preactivation output value $\langle \boldsymbol{x}, \boldsymbol{\beta} \rangle$ before being transformed with the softplus function $\zeta(\cdot)$ to obtain the $\mu$ value of the negative binomial distribution of Equation (3.17). The $\phi$ value is obtained by transforming a real-valued parameter $w_\phi$ through the softplus function. To obtain $\alpha$, the sum of past claims $\Sigma^{(y)}$ is added to the $\phi$ parameter, while for $\gamma$, the sum of past $\mu$ values $\Sigma^{(\mu)}$ is added to the same $\phi$ parameter. The resulting distribution parameters $\mu$, $\alpha,$ and $\gamma$ are then compared to the ground truth y using negative binomial deviance loss. The architecture shown employs a 3-hidden-layer MLP but can be customized with any number of layers.

Figure 8

Algorithm 2: Parameter estimation procedure – MVNB CANN model

Figure 9

Table 4. Handcrafted telematics features extracted from the telematics dataset.

Figure 10

Table 5. Optimal CANN models’ performance on the validation set

Figure 11

Table 6. Performance of the baseline model on the testing set

Figure 12

Table 7. Performance comparison of the CANN models and their corresponding log-linear benchmark model on the testing set. The percentages denote the improvement in the scoring rule compared to the baseline model

Figure 13

Table 8. Comparison of the sum of predicted values and the sum of actual values for CANN models using telematics

Figure 14

Figure 5. Importance scores of the 20 most important variables obtained for the MVNB CANN model.

Figure 15

Table A.1. Telematics inputs names translation

Supplementary material: File

Duval et al. supplementary material 1

Duval et al. supplementary material
Download Duval et al. supplementary material 1(File)
File 267 Bytes
Supplementary material: File

Duval et al. supplementary material 2

Duval et al. supplementary material
Download Duval et al. supplementary material 2(File)
File 269 Bytes
Supplementary material: File

Duval et al. supplementary material 3

Duval et al. supplementary material
Download Duval et al. supplementary material 3(File)
File 267 Bytes
Supplementary material: File

Duval et al. supplementary material 4

Duval et al. supplementary material
Download Duval et al. supplementary material 4(File)
File 319 Bytes
Supplementary material: File

Duval et al. supplementary material 5

Duval et al. supplementary material
Download Duval et al. supplementary material 5(File)
File 297 Bytes
Supplementary material: File

Duval et al. supplementary material 6

Duval et al. supplementary material
Download Duval et al. supplementary material 6(File)
File 97.4 KB
Supplementary material: File

Duval et al. supplementary material 7

Duval et al. supplementary material
Download Duval et al. supplementary material 7(File)
File 6.3 KB
Supplementary material: File

Duval et al. supplementary material 8

Duval et al. supplementary material
Download Duval et al. supplementary material 8(File)
File 326.5 KB
Supplementary material: File

Duval et al. supplementary material 9

Duval et al. supplementary material
Download Duval et al. supplementary material 9(File)
File 60.3 KB
Supplementary material: File

Duval et al. supplementary material 10

Duval et al. supplementary material
Download Duval et al. supplementary material 10(File)
File 475.4 KB
Supplementary material: File

Duval et al. supplementary material 11

Duval et al. supplementary material
Download Duval et al. supplementary material 11(File)
File 479.5 KB