Predictive models in banks, lenders, insurance companies and others are notorious for being extremely slow to produce and launch. This is the result of many reasons, none of them directly related to the technical details of building a predictive model. Some of the reasons are:
- Absence of direct linkage (code) between research model and the deployed model (often written in different languages!)
- Absence of a unified data lake holding both the training data as well as the new data coming from production.
- The need for full data lineage detailing the source and usage of every piece of data
- The need to go through a set of regulatory procedures and approvals
These reasons (and more) make the modeling effort in these organizations a one-off project where each iteration or versions of the same model feels like a whole new modelling project. This in turn leads to shockingly slow modelling iterations. How slow? Some organisations require as much as six months to build and sometimes even much longer to launch a new model!
Times like this, when a world pandemic is warping the financial universe as we know it, create an enormous stress on the modelling teams. They are required to tweak, validate and launch models in an ever increasing speed and agility. This is the enormous challenge they are facing.
Iterating on models
A model should be as lively and as fresh as any other production scale application code. Yes, it’s better to look at a model as a code project and learn from decades of experience gained in the software industry.
A model in production should be a direct product of the model assembly line. This means that if the model (function) disappears magically, the team can reproduce the exact same model in a few seconds with the click of a button. And the line continues backwards: the same model can be re-examined in the research environment without any additional fuss. New feature engineering ideas can be ideated and tested in production in minutes, not months! Even going all the way back to the data sources used for modelling – it should be darn easy to remove or add new data points and test the outcome without too much trouble.
Such iterations are simply impossible today in the reality of a heterogeneous modelling environment. Often, the pipeline is made up of five or even more different tools which does not necessarily talk to each other.
The end to end modelling experience
When models live in an end to end modelling environment, iterations measured in minutes and hours, become a reality. The entire stack needed to build, validate and deploy a model is centered around a single point of knowledge.
We can take this even further and close the loop between the “ends”. This creates a feedback loop which translates the lessons learnt from production data, automatically fed to the system as a new data source. Imagine that! a Shadow Modelling process running continuously and effortlessly in the background providing you with a current view of the potential of new models against the reality of production data.
Shadow Modelling
The EyeOnRisk platform is an ML -based credit risk focused, end to end platform for smarter, faster risk model management. It is specifically designed to boost credit risk management processes and is built with the bank’s processes and compliance needs in mind.
The platform encapsulates the model data flow. This lets you transfer models from research to production at an unparalleled speed. At the same time, this encapsulation allows for easier external data enrichment. Close monitoring of the data sources feeding the model while tracking data skews and suggests continual improvements.
Contact us to learn more
About the EyeOnRisk platform
The EyeOnRisk platform is an ML -based credit risk focused, end to end platform for smarter, faster risk model management. It is specifically designed to boost credit risk management processes and is built with the bank’s processes and compliance needs in mind.
The platform encapsulates the model data flow. This lets you transfer models from research to production at an unparalleled speed. At the same time, this encapsulation allows for easier external data enrichment. Close monitoring of the data sources feeding the model while tracking data skews and suggests continual improvements.
Contact us to learn more