Eleanor Brodie, senior manager of data science at LexisNexis Risk Solutions, sits down with Insurance Times to talk all things data
How much has changed for the data science team at LexisNexis Risk Solutions in the past year, both operationally and in terms of focus as a direct consequence of the pandemic?
Aside from the adjustment to homeworking, which has entailed lots of video calls with customers and colleagues, the past year has certainly fuelled the appetite for more and more data, attributes and scores. Insurance providers are more hungry now than they have ever been to evaluate new data-driven solutions to better understand and segment their customers.
A significant challenge for insurance providers is any sudden change in consumer behaviour. We all do things we don’t even think about as part of our daily routines. Shopping for insurance, driving to work and going to the supermarket are just a few activities that just happened until the pandemic took hold.
With very little notice, insurance providers needed to augment and expedite their ability to service their prospects and customers virtually. This includes pre-fill solutions, data-driven underwriting, pricing applications and contactless claims processing. Fortunately, we have been developing these solutions for years, so we have been in a good place to help insurance providers interact effectively with their customers and also understand how risks have changed as a direct consequence of the pandemic.
What are the biggest misconceptions you encounter about data for insurance risk assessment?
A big misconception in the industry is that big data and analytics driven products will replace human capital. Data-driven products allow insurance companies to streamline their existing processes and are meant to complement their existing workflow.
Human judgement and expertise will always be required to set the foundation, to ensure risk is accurately priced and aligns with each company’s business strategy. However, data-driven insights can assist with the decision-making process.
Another misunderstanding about data science is that there is a simple formula where all data is put into a magic funnel that draws out the desired outcome.
But, before any predictive model can be built or nugget of insight can be gleaned, data generally needs to be enriched, filtered and structured - this process relies heavily on the quality of the initial data sources and how they are modelled.
How can brokers and insurers put themselves in the strongest position to use data as part of the customer journey – from quoting to claim?
Look for data enrichment solutions that can be pulled into existing processes seamlessly and efficiently. Calling out for data from multiple data sources is inefficient and can lead to a poor customer experience, so look for platform solutions that offer the widest choice of data. You may not want that whole choice today, but you do want the flexibility to expand in the future.
Consider joining data sharing initiatives which will help you gain a view of the industry’s experience of an individual, vehicle or location. Shared motor policy history data has already established links between policy behaviour and claims losses. The next development will focus on shared claims data.
Also, conquer the challenge of creating a single customer view. Many insurance providers have gone through merger and acquisition activity over the past few years, which has made customer database management a real headache. It means you can end up with multiple records for the same customer.
When you are able to pull those disparate records together through linking and matching technology to create one consolidated view, you open up cross-sell and upsell opportunities based on a much clearer understanding of the customer. It also helps you gain the most value from data enrichment solutions.
Finally, for insurance providers to truly maximise and leverage the opportunities new data sources will bring through data enrichment at point of quote and renewal, it’s worth going back to their initial data sources.
This can start with refreshing the initial data model. Refreshing the data will identify behaviours used in rating the risk, how it’s changed and how it should be adapted for the current market. Without this crucial first step, adding in new data could be duplicating effort and capturing behaviours that existing models may already capture.
It is also worth viewing new data sources as possible replacements for existing data sets, rather than an add-on to the current data used. Insurance providers need to look at the incremental benefits a new data source will bring.
As the industry continues to rise to the challenge of pricing in a highly competitive market, maximising the opportunities of its data models to create more intelligible insights and a strategic advantage over competitors is vital.
Data is often called ‘the new gold’, but how do you assess the market opportunity and bring data sets to insurers and brokers that are satisfying an insurance market need?
We start by building an analytical prototype to validate an idea from our colleagues in the home, motor or commercial teams. These ideas come from the constant dialogue we have with customers on the pain points they need solved. Our new Covid-19 attributes are a good example of this – we could see the value the market could gain from understanding changes in motor policy behaviour during the first lockdown versus changes outside of that time.
Once we’ve proved the concept, we start the product development work, which may leverage millions of data points. We then create the final specs for technology to implement. As we near implementation, we develop the attributes or inputs into the solution and our audit team works with technology to ensure the final product performs as expected.
As the product is being developed, we give our customers the opportunity to test the solution on their own data. This might be through actionable insight studies or we may perform retrospective validation tests through our batch team.
Closer to launch, we look at any required regulatory documents on the solution inputs, outputs and overall performance. Then, following launch, we monitor the attributes and scores to ensure they continue to perform as expected.
It’s a well-oiled process and can be hugely rewarding when you find a new data attribute that you know can solve a pain point for the market.
Given the increasing demand for data scientists across many industries, what makes working with insurance data attractive to candidates and how do you attract and nurture this talent?
Competition for data scientists is indeed hot and while there are more colleges delivering great talent into the marketplace, I think demand will exceed supply for some time yet.
I might be biased, but we do have a lot more to offer good data scientists. We have what they dream of - good quality data and lots of it.
Our data scientists work with literally hundreds of millions, and often billions, of records to solve our customers’ problems. Where other companies are limited in the breadth and depth of their data and the ability to pull it all together in a commercially viable application, this is what we do every day and that is exciting to a candidate.
Our Data Science Rotational Programme (DSRP) sees graduates from a wide array of disciplines, including but not limited to mathematics, statistics, computer science, data science, physics, financial math, actuarial science and engineering, join LexisNexis for a two-year cycle through four different teams.
This experience provides a robust hands-on journey from data access, data analysis and model building through to model implementation. Right now, we have six DSRP team members in this programme globally and we typically hire three new positions each year.