In-context learning enhanced credibility transformer


Loading...

Date

2025-01-01

Publication Type

Journal Article

ETH Bibliography

Citations

Web of Science:
Altmetric

Data

Rights / License

Abstract

The starting point of our network architecture is the Credibility Transformer which extends the classical Transformer architecture by a credibility mechanism to improve model learning and predictive performance. This Credibility Transformer learns credibilitised classification tokens that serve as learned representations of the original input features. In this paper we present a new paradigm that augments this architecture by an in-context learning mechanism, i.e., we increase the information set by a context batch consisting of similar instances. This allows the model to enhance the classification token representations of the instances by additional in-context information and fine-tuning. We empirically verify that this in-context learning enhances predictive accuracy by adapting to similar risk patterns. Moreover, this in-context learning also allows the model to generalise to new instances which, e.g., have feature levels in the categorical covariates that have not been present when the model was trained-for a relevant example, think of a new vehicle model which has just been developed by a car manufacturer.

Publication status

Editor

Book title

Journal / series

SOUTH AFRICAN ACTUARIAL JOURNAL

Volume

25

Pages / Article No.

Publisher

Event

Edition / version

Methods

Software

Geographic location

Date collected

Date created

Subject

Transformer; credibility transformer; in-context learning; attention layer; foundation model; insurance pricing

Organisational unit

Notes

Funding

Related publications and datasets