A leading non-bank financial institution partnered with Evolve AI Labs to modernise their credit decisioning infrastructure—approving 10% more creditworthy applicants while reducing bad debt rates and cutting model development time in half.
$1M
ROI
10%
uplift in acceptance rate
When a non-bank lender competing against major banking institutions approached us, they faced what seemed like a straightforward modernisation challenge: their credit scorecards were conservative and slow to update. But after working through their underwriting data, we uncovered something more fundamental.
The real constraint wasn't just scorecard performance—it was their entire approach to credit decisioning. Traditional scorecards required three months to rebuild and redeploy, making it impossible to respond to shifting market conditions or evolving consumer behaviours. Every percentage point of approval rate they left on the table represented millions in foregone revenue, yet their rigid development cycles prevented the rapid iteration needed to capture this opportunity.
Their existing scorecards exhibited a common pattern we see across lending institutions: they were optimised for risk avoidance rather than risk discrimination. This meant rejecting borderline applicants who would have been profitable customers—a particularly costly approach for a lender whose competitive advantage came from serving creditworthy near-prime borrowers that traditional banks often decline.
The post-COVID spending reset had amplified this problem. Consumer credit patterns had fundamentally shifted, but the lender's scorecards—built on pre-pandemic assumptions—couldn't adapt quickly enough. They were caught between regulatory requirements for explainable decisions and the need for models sophisticated enough to identify profitable lending opportunities in a transformed market.
Rather than accepting the three-month development cycle as fixed, we rebuilt their scorecard development process from first principles. Instead of the traditional approach of building a handful of models over several months, we constructed an infrastructure that could generate and evaluate 100+ machine learning models in three weeks.
This wasn't about brute-force computation—it was about understanding which assumptions in their existing process were actually regulatory requirements versus inherited practices. We separated the explainability requirements (regulatory constraints) from the development velocity constraints (self-imposed limitations).
We deployed Generalised Additive Models (GAMs) as the production solution—not because they were the most accurate in isolation, but because they solved the credit risk trilemma: maintaining regulatory explainability while achieving superior predictive performance compared to linear models, all without introducing the opacity that comes with gradient boosting or neural network approaches.
The GAMs provided something their previous scorecards couldn't: the ability to understand non-linear relationships in credit behaviour while generating the feature importance plots and partial dependence explanations that compliance officers require. This mattered because the lender operated in a regulatory environment where every declined application could face scrutiny.
The breakthrough came from incorporating unobserved applicant data through reject inference. Their existing scorecards suffered from a fundamental selection bias: they were trained only on customers who had been approved, creating a blind spot for potentially profitable borrowers who matched the profile of previously rejected applicants.
By reconstructing the likely outcomes for rejected applicants based on similar approved customers, we identified patterns that traditional scorecards consistently missed. This was particularly valuable for near-prime applicants—the exact segment where this lender had competitive advantages over larger banks.
We implemented DataRobot's MLOps capabilities to transform scorecard development from a quarterly project into a continuous improvement process. The infrastructure includes:
This infrastructure addresses something we see across financial services: the gap between proof-of-concept success and production reliability. Most lending institutions can build accurate models in testing environments—the challenge is maintaining that performance while meeting enterprise requirements for governance, compliance, and operational stability.
After deploying similar solutions across multiple lenders, we've identified three systemic issues that prevent most institutions from achieving these results:
Development Cycles Optimised for Risk Avoidance, Not Learning Velocity
Most lenders treat scorecard development as a high-stakes, low-frequency event. This creates perverse incentives where teams avoid experimentation because each model takes months to develop. The result is conservative scorecards that prioritise defending the status quo over discovering better approaches.
We inverted this dynamic by making model development so fast and cheap that experimentation became the default. When you can test a hypothesis in days rather than months, the risk calculus fundamentally changes.
Training Data That Encodes Historical Biases
Traditional scorecards are trained only on approved applicants, which means they're optimised to replicate past decisions rather than improve on them. This creates a self-reinforcing cycle where the model learns to approve applicants similar to those previously approved, leaving profitable opportunities undiscovered.
Reject inference breaks this cycle by forcing the model to reason about what would have happened with declined applicants, revealing patterns that aren't visible in the approved-only training set.
Explainability as an Afterthought Rather Than a Design Constraint
Many lending institutions treat model explainability as a documentation exercise—something to retrofit after selecting the most accurate model. This leads to a false choice between performance and transparency.
By selecting GAMs as the modelling approach from the beginning, we designed explainability into the architecture rather than bolting it on afterwards. This meant the compliance requirements actually helped constrain the solution space toward more interpretable, maintainable models.
The new scorecards approved 10% more applicants while reducing bad debt rates. For a lending institution processing tens of thousands of applications annually, this represented substantial incremental revenue with no increase in credit losses.
The improvement came from better risk discrimination—the models could identify profitable near-prime borrowers that traditional scorecards reflexively rejected. This aligned perfectly with the lender's competitive positioning: serving creditworthy customers that larger institutions overlooked.
Model development cycles decreased from three months to six weeks, and the infrastructure supports even faster iteration for smaller refinements. This speed advantage compounds over time as the lender can respond to market changes, test new variables, and optimise for emerging customer segments while competitors are still updating their quarterly scorecards.
The MLOps infrastructure provides lasting value beyond the initial deployment:
Automated retraining as new data accumulates, ensuring models stay current with market conditions real-time performance monitoring that identifies degradation before it impacts business metrics champion-challenger testing that validates improvements without risking production stability complete audit trails that satisfy regulatory examinations while enabling rapid iteration
This infrastructure converts scorecard maintenance from a periodic project requiring senior resources into an automated process with exception-based human oversight.
Interested in understanding how assumption-aware architecture could apply to your credit decisioning process?
We build a custom solutions as well as provide AI implementation consultation to help you maximise business revenue, reduce costs and add operational efficiency
Speak to an expert