Skip to main navigation menu Skip to main content Skip to site footer

Articles

Vol. 2 No. 2 (2022): Blockchain Technology and Distributed Systems

Predictive Modelling for Insurance Pricing: A Comparative Analysis of Machine Learning Approaches

Published
06-07-2022

Abstract

This research paper delves into the realm of predictive modelling for insurance pricing, focusing on a comparative analysis of various machine learning approaches. The study scrutinizes the accuracy, interpretability, and scalability of these methods to offer insights into their effectiveness within the insurance domain. Leveraging extensive datasets, diverse machine learning algorithms are examined, including decision trees, random forests, gradient boosting machines, neural networks, and support vector machines. Each approach is rigorously evaluated based on its predictive performance, transparency of decision-making processes, and computational efficiency. Through empirical analysis and statistical comparisons, this paper illuminates the strengths and limitations of different techniques, providing valuable guidance for insurance companies seeking optimal pricing strategies.

References

  1. Breiman, Leo. "Random forests." Machine learning 45.1 (2001): 5-32.
  2. Friedman, Jerome H. "Greedy function approximation: A gradient boosting machine." Annals of statistics (2001): 1189-1232.
  3. Hastie, Trevor, Robert Tibshirani, and Jerome Friedman. The elements of statistical learning: data mining, inference, and prediction. Springer Science & Business Media, 2009.
  4. Quinlan, J. Ross. "C4. 5: programs for machine learning." Morgan Kaufmann, 2014.
  5. Bishop, Christopher M. "Pattern recognition and machine learning." springer, 2006.
  6. James, Gareth, et al. "An introduction to statistical learning." springer, 2013.
  7. R Core Team. R: A language and environment for statistical computing. R Foundation for Statistical Computing, 2021.
  8. Cherkassky, Vladimir, and Filip Mulier. "Learning from data: concepts, theory, and methods." John Wiley & Sons, 2007.
  9. Chen, Tianqi, and Carlos Guestrin. "Xgboost: A scalable tree boosting system." Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2016.
  10. Freund, Yoav, and Robert E. Schapire. "A decision-theoretic generalization of on-line learning and an application to boosting." Journal of computer and system sciences 55.1 (1997): 119-139.
  11. Schapire, Robert E. "The strength of weak learnability." Machine learning 5.2 (1990): 197-227.
  12. Lecun, Yann, Yoshua Bengio, and Geoffrey Hinton. "Deep learning." nature 521.7553 (2015): 436-444.
  13. Hinton, Geoffrey, et al. "Deep neural networks for acoustic modelling in speech recognition: The shared views of four research groups." IEEE Signal processing magazine 29.6 (2012): 82-97.
  14. Kim, Edward Y., et al. "Forecasting insurance loss payments using machine learning." Variance 13.1 (2019): 25-57.
  15. Cohen, Israel, and Naftali Tishby. "Information theoretic considerations concerning the design of convex functions for binary classification." Machine learning 30.1 (1998): 11-47.
  16. Hall, Peter, and Joel L. Horowitz. "Methodology and convergence rates for functional linear regression." The Annals of Statistics 35.1 (2007): 70-91.
  17. Hastie, Trevor, et al. "Imputing missing data with the EM algorithm." The Journal of Machine Learning Research 8 (2007): 1623-1657.
  18. Kuhn, Max, and Kjell Johnson. Applied predictive modelling. Springer, 2013.
  19. Tibshirani, Robert. "Regression shrinkage and selection via the lasso." Journal of the Royal Statistical Society: Series B (Methodological) 58.1 (1996): 267-288.
  20. Witten, Ian H., Eibe Frank, and Mark A. Hall. Data mining: practical machine learning tools and techniques. Morgan Kaufmann, 2016

Most read articles by the same author(s)