Claim your CPD points
ChatGPT has captured public attention with its impressive ability to mimic human language and engage in conversations on various topics. It reached 100 million monthly users just two months after public release.
The area is developing quickly. GPT-4, released on 14 March 2023, now passes the US Bar - and is estimated to beat ~90% of all those sitting the exam. The previous model GPT-3.5 performed at the 10% percentile and failed the exam.
But what does this technology mean for actuaries and their work? In this article series, we explore some of the commercial implications for LLMs, and some of the ethical and professional challenges that generative AI poses for actuaries.
LLMs process vast amounts of natural language data with the goal of generating text that is indistinguishable from human writing and can be fine-tuned for specific tasks. They are based on a deep neural network architecture that uses the concept of "attention", which they then use to predict the next word in a sentence, given the previous words. The more data the models are trained on, the more accurate and sophisticated their language abilities become.
The concept of "attention" was first proposed by Google researchers, originally for language translation, to learn the relationships between words in a text . A summary of the mathematical concepts is explained in this video.
LLMs like those powering ChatGPT bring a number of unique benefits that make them suitable for commercial applications.
While these abilities are impressive, it is important to recognise current limitations. LLMs have the tendency to 'hallucinate', which means that responses may be factually incorrect but be delivered with the same degree of certainty. Results are ultimately unpredictable - the same prompt can generate different responses, so ensuring a consistent level of quality is not guaranteed. Training inputs are drawn from the web and may reflect biases that are present. The technology has obvious implications for cheating at school or university, given its good performance on tasks. The abilities can also be used for nefarious purposes - such as hackers using the tools to write malware more efficiently.
There are a wide range of use cases for the commercial applications of LLMs. Below lists a few of them, starting from the more actuarially relevant examples:
Overall, LLMs have many commercial applications and are becoming increasingly popular as businesses leverage the power of natural language processing (NLP) to improve customer experience, increase efficiency, and gain insights from data. Going one level lower, how do these actually help actuaries with their day-to-day work?
In a recent real-life example, one of the co-authors of this article was handed over a piece of R script without much documentation. As some actuaries may be able to relate, seeking explanations or documents for R functions over the internet may be time-consuming and typically, community sites such as Stack Overflow and R-bloggers offer only knowledge sharing in function silos (i.e. for specific function or syntax). This author was particularly impressed by ChatGPT which provided not only a comprehensive summary but also bugs and suggested fixes after simply throwing a chunk of the R script which contains multiple R functions into a ChatGPT prompt.
This offers a glimpse of the applications of LLMs for actuaries which will be discussed in detail in Part 2 of the article series (to be published next week). Overall, Data Science Actuaries should be looking to be part of how these models are designed, built, implemented and governed within organisations.
Although ChatGPT has received significant media attention, it is not the only LLM-powered solution available. The field of LLMs, and Generative AI more broadly, is developing quite quickly, with many competing models and improvements released within a number of months:
The future state does look promising but what does this mean for actuaries?
Using the topical GPT-4 as an example, it's powered by a multi-modal LLM in that it handles not only natural language but also image. In Open-AI's live demonstration for the release of GPT-4, it was shown that GPT-4 was able to turn images of hand drawings into functional websites in a matter of minutes. Its visual capabilities also allowed it to perform well in some of the standard tests (e.g. GPT-4 passed the US Bar exam as mentioned earlier, and scored in the 99th percentile in the Biology Olympiad). For actuaries specialised in insurance operations, a future state which could potentially be reached by these LLMs is to underwrite policyholders in real-time by simply requesting images or photographs.
How can we possibly keep up? Independent researchers have put enormous effort into quickly investigating and experimenting with ChatGPT v3.5 to understand its political correctness, biases, and areas most at risk of hallucination, and to refine "prompt engineering" techniques to produce the most effective results. With "ChatGPT Plus" based on GPT 4.0, how much of those research findings remain relevant now? And how much would remain relevant in a year?
It is easy to get caught up in the excitement, but as actuaries and data specialists, we need to be aware of the risks of integrating these new technologies into our lives and the way we work. As with the outputs of any data-driven tool, results need to be critically evaluated with careful consideration of the potential impact on stakeholders.
The capabilities and risks of LLMs will only continue to increase, but with proper care and attention, they can unlock new possibilities for innovation and progress in our field. Stay in the loop with Actuaries Digital as in the upcoming weeks, we will explore applications, especially for actuaries, and discuss the risks with using LLMs in more detail.