Machine Learning: Challenges, Lessons, and Opportunities in Credit Risk Modeling
Thanks to rapid increases in data availability and computing power, machine learning now plays a vital role in both technology and business. Machine learning contributes significantly to credit risk modeling applications. Using two large datasets, we analyze the performance of a set of machine learning methods in assessing credit risk of small and mediumsized borrowers, with Moody’s Analytics RiskCalc model serving as the benchmark model. We find the machine learning models deliver similar accuracy ratios as the RiskCalc model. However, they are more of a “black box” than the RiskCalc model, and the results produced by machine learning methods are sometimes difficult to interpret. Machine learning methods provide a better fit for the nonlinear relationships between the explanatory variables and default risk. We also find that using a broader set of variables to predict defaults greatly improves the accuracy ratio, regardless of the models used.
Introduction
Machine learning is a method of teaching computers to parse data, learn from it, and then make a determination or prediction regarding new data. Rather than handcoding a specific set of instructions to accomplish a particular task, the machine is “trained” using large amounts of data and algorithms to learn how to perform the task. Machine learning overlaps with its lowerprofile sister field, statistical learning. Both attempt to find and learn from patterns and trends within large datasets to make predictions. The machine learning field has a long tradition of development, but recent improvements in data storage and computing power have made them ubiquitous across many different fields and applications, many of which are very commonplace. Apple’s Siri, Facebook feeds, and Netflix movie recommendations all rely upon some form of machine learning. One of the earliest uses of machine learning was within credit risk modeling, whose goal is to use financial data to predict default risk.
When a business applies for a loan, the lender must evaluate whether the business can reliably repay the loan principal and interest. Lenders commonly use measures of profitability and leverage to assess credit risk. A profitable firm generates enough cash to cover interest expense and principal due. However, a moreleveraged firm has less equity available to weather economic shocks. Given two loan applicants – one with high profitability and high leverage, and the other with low profitability and low leverage – which firm has lower credit risk? The complexity of answering this question multiplies when banks incorporate the many other dimensions they examine during credit risk assessment. These additional dimensions typically include other financial information such as liquidity ratio, or behavioral information such as loan/trade credit payment behavior. Summarizing all of these various dimensions into one score is challenging, but machine learning techniques help achieve this goal.
The common objective behind machine learning and traditional statistical learning tools is to learn from data. Both approaches aim to investigate the underlying relationships by using a training dataset. Typically, statistical learning methods assume formal relationships between variables in the form of mathematical equations, while machine learning methods can learn from data without requiring any rulesbased programming. As a result of this flexibility, machine learning methods can better fit the patterns in data. Figure 1 illustrates this point.
Figure 1 Statistical model vs. machine learning
Source: Moody's Analytics
In this simulated example, the first chart shows the actual distribution of data points with respect to X and Y, while the points in red are classified as defaults. One can relate this to a geographical map, where the X axis is longitude, and the Y axis is latitude. The areas in red represent highrisk demographics, where we see a higher default rate. As expected, a linear statistical model cannot fit this complex nonlinear and nonmonotonic behavior. The random forest model, a widely used machine learning method, is flexible enough to identify the hot spots because it is not limited to predicting linear or continuous relationships. A machine learning model, unconstrained by some of the assumptions of classic statistical models, can yield much better insights that a human analyst could not infer from the data. At times, the prediction contrasts starkly with traditional models.
A machine learning model, unconstrained by some of the assumptions of classic statistical models, can yield much better insights that a human analyst could not infer from the data.
Machine Learning Approaches
Now let’s look at three different machine learning algorithms: artificial neural networks, random forest, and boosting.
Artificial Neural Networks
An artificial neural network (ANN) is a mathematical simulation of a biological neural network. Its simple form is shown in Figure 2. In this example, there are three input values and two output values. Different transformations link the input values to a hidden layer, and the hidden layer to the output values. We use a backpropagation algorithm to train the ANNs on the underlying data. ANNs can easily handle the nonlinear and interactive effects of the explanatory variables due to the presence of many hidden layers and neurons.
Figure 2 Artificial neural network
Source: Moody's Analytics
Random Forest
Random forests combine decision tree predictors, such that each tree depends on the values of a random vector sampled independently, and with the same distribution. A decision tree is the most basic unit of the random forest. In a decision tree, an input is entered at the top and, as it traverses down the tree, the data is bucketed into smaller and smaller subsets. In the example shown in Figure 3, the tree determines probability of default based on three variables: firm size; the ratio of earnings before interest, tax, depreciation, and amortization (EBITDA) to interest expense; and the ratio of current liabilities to sales. Box 1 contains the initial dataset in which 39% of the firms are defaulters and 61% are nondefaulters. Firms with EBITDAtointerest expense ratios less than 2.4 go into Box 2. Box 2, accounting for 33% of the data, is 100% composed of defaulters. Its orange color indicates higher default risk, whereas the blue color indicates lower default risk. The random forest approach combines the predictions of many trees, and the final decision is based on the average of the output of the underlying independent decision trees. In this exercise, we use the bootstrap aggregation of several trees as an advancement to a simple treebased model.^{1}
Figure 3 Random forest
Source: Moody's Analytics
Boosting
Boosting is similar to random forest, but the underlying decision trees are weighted based on their performance. Consider the parable of the blind men and the elephant, in which the men are asked to touch different parts of the elephant and then construct a full picture. The blind men are sent in six different batches. The first group is led to randomly selected spots, and each person’s (partial) description is evaluated on how well it matches the actual description. This group happens to give an accurate description of only the trunk, while description of the rest of the body is inaccurate. The incomplete sections are noted, and when the second batch of blind men is led into the room, they are steered to these parts. This process is repeated for the remaining batches. Finally, the descriptions are combined additively by weighting them according to their accuracy and, in this case, the size of the body parts as well. This final description – the combination – describes the elephant quite well.
In boosting, each decision tree is similar to a group of blind men, and the description of the elephant is synonymous to the prediction problem being solved. If a tree misclassifies defaulters as nondefaulters or vice versa, the subsequent trees will put more weight on the misclassified observations. This idea of giving misclassified areas additional weight (or direction while sending in a new group) is the difference between random forests and boosting.
Moody’s Analytics RiskCalc Model
The RiskCalc model produces expected default probabilities for private firms by estimating the impact of a set of risk drivers. It utilizes a generalized additive model (GAM) framework, in which nonlinear transformations of each risk driver are assigned weights and combined into a single score. A link function then maps the combined score to a probability of default.
The RiskCalc model delivers robust performance in predicting private firm defaults. But how does it compare to other machine learning techniques? We use the three popular machine learning methods to develop new models using the RiskCalc sample as a training set. We seek to answer the following questions: Do the machine learning models outperform the RiskCalc model’s GAM framework in default prediction? What are the challenges we face when using the machine learning methods for credit risk modeling? Which model is most robust? Which model is easiest to use? And what can we learn from the alternative models?
Results
Data Description
To analyze the performance of these three approaches, we consider two different datasets. The first dataset comes from the Moody’s Analytics Credit Research Database (CRD) which is also the validation sample for the RiskCalc US 4.0 corporate model. It utilizes only firm information and financial ratios. The second dataset adds behavioral information, which includes credit line usage, loan payment behavior, and other loan type data. This information comes from the loan accounting system (LAS), collected as part of the CRD. We want to test for additional default prediction power using the machine learning techniques and the GAM approach with both datasets. Figure 4 shows the summary of the two datasets.
Figure 4 Data information
Source: Moody's Analytics
Model Performance
For both datasets, we use the GAM model’s rank ordering ability as the benchmark. We measure rank ordering ability using the accuracy ratio (AR) statistic. Figure 5 shows the set of explanatory variables.
Figure 5 Input variable descriptions for the PD models
Source: Moody's Analytics
CrossValidation
Because machine learning offers a high level of modeling freedom, it tends to overfit the data. A model overfits when it performs well on the training data but does not perform well on the evaluation data. A standard way to find outofsample prediction error is to use kfold crossvalidation (CV). In a kfold CV, the dataset is divided into k subsets. One of the k subsets is used as the test set, and the other k1 subsets are combined to form a training set. This process is repeated k times. If the accuracy ratio, a measure of model performance, is high for the training sample relative to the test sample, it indicates overfitting. In this case, we impose more constraints on the model and repeat crossvalidation until the results are satisfactory. In this example, we use a fivefold cross validation. Figure 6 reports the average AR across the five trials.
Figure 6 Model performance
Source: Moody's Analytics
We observe that machine learning models outperform the GAM model by 2 to 3 percentage points for both datasets. The accuracy ratio improves by 8 to 10 percentage points when we add loan behavioral information, regardless of the modeling approach. Credit line usage and loan payment information complement financial ratios and significantly enhance the models’ ability to predict defaults.
Where Machine Learning Excels
Machine learning methods are particularly powerful in capturing nonlinear relationships. Let’s take a closer look at the EBITDAtointerestexpense ratio. Intuitively, this ratio has a nonlinear relationship with default risk. In Figure 7, we divide the ratio into 50 percentiles and calculate the average values of predicted probability of default (PD) and the actual default rate. We plot this with the ratio percentiles on the xaxis and the default rate (in %) on yaxis. The default rate decreases as the ratio of EBITDA to interest expense increases. However, on the lefthand side, there is an inflection point where the EBIDTA becomes negative. When EBITDA is negative, as the interest expense decreases making the ratio more negative, the default risk should decrease. From the graph, we observe that the machine learning method of boosting provides a more accurate prediction of the actual default rate than the GAM model, especially on the lefthand side. We observe this similar behavior from the plots of other ratios, as well. Hence, we observe modest prediction improvement for machine learning methods.
Figure 7 Comparing machine learning and GAM PD levels for different values of EBITDA to interest expense
Source: Moody's Analytics
Overfitting Problem
Despite the use of crossvalidation to minimize overfitting, machine learning models may still produce results that are difficult to interpret and defend. Figure 8 shows two cases in which the PD determined by the boosting method differs significantly from the PD determined by the GAM approach.
Figure 8 Overfitting of machine learning algorithms
Source: Moody's Analytics
In case 1, a company with a negative return on assets (ROA), a low cashtoassets ratio, and a high debttodebtplusequity ratio is classified as safe, with an implied rating of A3. Intuitively, this firm’s PD should reflect a higher level of risk, as predicted by GAM. Similarly in case 2, a firm with high EBITDA to interest expense, high ROA, and high retained earnings is categorized as Caa/C using the boosting method. In both cases, the complex nature of the underlying algorithm makes it difficult to explain the boosting method’s nonintuitive PD. The RiskCalc model’s results, based on the GAM model, are much more intuitive and easier to explain.
Summary
This exercise analyzes the performance of three machine learning methods using the RiskCalc software’s GAM model as a benchmark. The machine learning approaches deliver comparable accuracy ratios as the GAM model. Compared to the RiskCalc model, these alternative approaches are better equipped to capture the nonlinear relationships common to credit risk.At the same time, the predictions made by the approaches are sometimes difficult to explain due to their complex “black box” nature. These machine learning models are also sensitive to outliers, resulting in an overfitting of the data and counterintuitive predictions. Additionally, and perhaps more interestingly, we find that expanding the dataset to include loan behavioral variables improves predictive power by over 10 percentage points for all modeling methods.
While the approaches we study all have their merits and have comparable accuracy levels, we believe that to improve default prediction accuracy and to expand the field of credit risk modeling in general, efforts should focus on the data dimension. Besides financial statement and loan payment behavioral data, additional information such as transactional data, social media data, geographical information, and other data can potentially add a tremendous amount of insight. We must gather more varied, nonconventional data to further refine and improve our approaches to assessing risk.
Sources
1 Breiman, Leo. “Random Forests.” Machine Learning, volume 45, issue 1. October 2001.
Featured Experts
Douglas Dwyer
Credit risk modeling expert; machine learning researcher; quantitative and statistical data analyst
Rama Sankisa
Risk metric architect; data scientist; quantitative researcher; macroeconomic and exchange rate model developer
John Baer
Credit risk expert; data analytics and technology process consultant; former M&A advisor for private equity and corporate clients
As Published In:
Examines the role of disruptive technologies in the financial sector and how firms can improve their practices to remain competitive.
Related Articles
Whitepaper
Sovereign & SizeAdjusted EDFImplied Rating Template (for Private Firms)RiskCalc™ EDF™ (Expected Default Frequency) values and agency ratings are widely used credit risk measures. RiskCalc EDF values typically measure default risk for private companies, while agency ratings are only available for rated companies. A RiskCalc EDF value measures a company's standalone credit risk based on financial statement information, while an agency rating considers qualitative factors such as Business Profile, Financial Policy, external support, and countryrelated risks. Moody's Analytics new Sovereign & SizeAdjusted EDFImplied Rating Template combines RiskCalc EDF values with additional factors to provide a rating comparable to agency ratings for private companies. The new template applies to RiskCalc EDF values across numerous geographies and regulatory environments. With the new template, users can generate a rating more comparable to an agency rating than RiskCalc EDF values or EDFimplied ratings. Analyzing data from 3,900+ companies in 60+ countries, we find that sovereign rating and total asset size, in addition to EDF value, have a statistically significant impact on an agency rating — our quantitative template incorporating these three variables reliably estimates agency ratings in a robust fashion. 
Whitepaper
Identifying AtRisk Names in Your Private Firm Portfolio — RiskCalc Early Warning ToolkitThis report outlines a practical approach for using RiskCalc EDF credit measures to effectively monitor large portfolios of private firms and to proactively identify atrisk names. The RiskCalc Early Warning Toolkit Excel addin is an easy to use, yet comprehensive tool that allows users to focus costly and scarce resources on a highly targeted selection of the most atrisk names in their portfolios. This research for private firms compliments previous research on Early Warning Toolkit for public firms. The Early Warning Toolkit identifies atrisk names within a private firm portfolio well before default, using a number of different EDFrelated risk metrics. 
Moody's Analytics Webinar: Improve Your Lending with Scores from RiskCalc™ Small BusinessMoody’s Analytics recently released the RiskCalc™ Small Business solution as part of our ongoing efforts to help lenders better understand and serve their small business clients. 
WebinaronDemand
Improve Your Lending with Scores from RiskCalc™ Small BusinessMoody’s Analytics recently released the RiskCalc™ Small Business solution as part of our ongoing efforts to help lenders better understand and serve their small business clients. 
Presentation
Improve Your Lending with Scores from RiskCalc™ Small BusinessImprove Your Lending with Scores from RiskCalc™ Small Business 
Whitepaper
Features of a Lifetime PD Model: Evidence from Public, Private, and Rated FirmsWith the new CECL and IFRS 9 requirements, this document formally investigates and summarizes the term structure properties consistently seen across public, private, and rated firms. 
WebinaronDemand
Applications of Alternative Data in Credit DecisioningWith an immense amount of available data generated worldwide within the last two years, the next evolution of banking analytics will include information from a variety of open and closed sources. 
Presentation
Applications of Alternative Data in Credit DecisioningIn this webinar, a panel of research and data scientist experts across Moody's Analytics discuss social data in probability of default modeling, closed and open data for location scoring, and text analytics for credit risk. 
Whitepaper
Combining Financial and Behavioral Information to Predict Defaults for Small and MediumSized Enterprises: A Dynamic Weighting ApproachThis note presents the first tool that assesses borrowers’ credit risk using a scientific method that leverages both financial and behavioral information.
September 2017
WebPage
Alessio Balduini, Dr. Douglas Dwyer, Sara Gianfreda, Reeta Hemminki, Lucia Yang, Dr. Janet Zhao

Whitepaper
Moody's Analytics RiskCalc Transfer Pricing SolutionTax authorities monitor crossborder, intercompany loan and financing transactions to curb tax avoidance and require arm's length pricing for such transactions. At the core of arm's length pricing is the process of understanding the creditworthiness of a borrower and identifying a typical interest rate charged to borrowers with comparable credit ratings. The Moody's Analytics RiskCalc Transfer Pricing Excel Template provides a consistent, analytical solution to the arm's length transfer pricing process. This document explains the methodology behind this tool. 