scispace - formally typeset
Search or ask a question
Journal ArticleDOI

Bankruptcy prediction for credit risk using neural networks: A survey and new results

01 Jul 2001-IEEE Transactions on Neural Networks (IEEE)-Vol. 12, Iss: 4, pp 929-935
TL;DR: Inspired by one of the traditional credit risk models developed by Merton (1974), it is shown that the use of novel indicators for the NN system provides a significant improvement in the (out-of-sample) prediction accuracy.
Abstract: The prediction of corporate bankruptcies is an important and widely studied topic since it can have significant impact on bank lending decisions and profitability. This work presents two contributions. First we review the topic of bankruptcy prediction, with emphasis on neural-network (NN) models. Second, we develop an NN bankruptcy prediction model. Inspired by one of the traditional credit risk models developed by Merton (1974), we propose novel indicators for the NN system. We show that the use of these indicators in addition to traditional financial ratio indicators provides a significant improvement in the (out-of-sample) prediction accuracy (from 81.46% to 85.5% for a three-year-ahead forecast).

Summary (1 min read)

Introduction

  • In this study the authors focus only on the corporate bankruptcy prediction problem.
  • It can help in accurately assessing thecredit risk of bank loan portfolios.
  • The focus of this article is on the empirical approach, especially the use of NNs.

A. Early Empirical Approaches

  • Altman uses the classical multivariate discriminant analysis technique (MDA).
  • It is based on applying the Bayes classification procedure, under the assumption that the two classes have Gaussian distributions with equal covariance matrices.
  • These particular financial ratios have been widely used as inputs, even for NNs and other nonlinear models.
  • Ohlson introduced the logistic regression approach (LR) to the bankruptcy prediction problem.

B. Neural-Network (NN) Approaches

  • Research studies on using NNs for bankruptcy prediction started in 1990, and are still active now.
  • One of the first studies to apply NNs to the bankruptcy prediction problem was the work by Odom and Sharda [33].
  • The total assets gives some indication of the size of the firm.
  • Boritz and Kennedy [9] (see also [10]) compared between a number of techniques, including different NN training procedures, LR and MDA, using the indicators chosen by Altman, and those chosen by Ohlman.
  • Their method obtains better results than NN and MDA for Type II error, but worse results for Type I error.

C. A Brief Review of the Structural Approach

  • One of the earlier and commonly used methods is the assetbased approach by Merton [31] (developed further by Longstaff and Schwartz [28]).
  • This model views a firm’s equity as an option on the firm (held by the shareholders) to either repay the debt of the firm when it is due, or abandon the firm without paying the obligations.
  • This model has been successfully developed into a successful commercial product by KMV Corporation.
  • Another approach, by Jarrow and Turnbull [20], models default as a point process, where the time-varying hazard function for each credit class is estimated from the credit spreads.
  • J.P. Morgan’s CreditMetrics product [14] is based on modeling changes in the credit quality ratings.

D. Challenges for the NN Prediction Models

  • For portfolio credit risk estimation, this is essential in order to compute the loss level.
  • As such, indicators obtained from the stock price can be beneficial especially in long horizon default forecast.
  • Using an initial prescreening procedure based on individual indicator prediction accuracy and correlation matrix, and then a subsequent cross-validation procedure to narrow down the choice, the authors select the best five or six inputs from this pool of indicators.
  • He also received theYoung Investigator Award from the International Neural Network Society, in 1996.

Did you find this useful? Give us your feedback

Content maybe subject to copyright    Report

IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 12, NO. 4, JULY 2001 929
Bankruptcy Prediction for Credit Risk Using Neural
Networks: A Survey and New Results
Amir F. Atiya, Senior Member, IEEE
Abstract—The prediction of corporate bankruptcies is an
important and widely studied topic since it can have signifi-
cant impact on bank lending decisions and profitability. This
work presents two contributions. First we review the topic of
bankruptcy prediction, with emphasis on neural-network (NN)
models. Second, we develop an NN bankruptcy prediction model.
Inspired by one of the traditional credit risk models developed
by Merton, we propose novel indicators for the NN system. We
show that the use of these indicators in addition to traditional
financial ratio indicators provides a significant improvement in
the (out-of-sample) prediction accuracy (from 81.46% to 85.5%
for a three-year-ahead forecast).
Index Terms—Asset-based model, bankruptcy prediction, cor-
porate distress, corporate failure prediction, credit risk, default
prediction, financial ratios, financial statement data, multilayer
networks.
I. INTRODUCTION
B
ANKRUPTCY prediction has long been an important and
widely studied topic. The main impact of such research is
in bank lending. Banks need to predict the possibility of default
of a potential counterparty before they extend a loan. This can
lead to sounder lending decisions, and therefore result in sig-
nificant savings. In this study we focus only on the corporate
bankruptcy prediction problem. For the consumer bankruptcy
prediction problem, there is likewise an extensive amount of re-
search, but the reader is referred to [19], [40], and [52] for a
review of this topic.
To get an idea about the potential impact of the bankruptcy
prediction problem, we note that the volume of outstanding debt
to corporations in the United States is about $5 trillion. An
improvement in default prediction accuracy of just a few per-
centage points can lead to savings of tens of billions of dol-
lars. In addition to avoiding potentially troubled obligors, the
research can also benefit in other ways. It can help in estimating
a fair value of the interest rate of a loan (that reflects the cred-
itworthiness of the counterparty). It can help in accurately as-
sessing the credit risk of bank loan portfolios. The credit risk
problem is essentially the computation of the loss level, which
is defined as the level for which there is a probability of 1% that
the loss incurred in the portfolio will exceed that level in a par-
ticular time period. Credit risk has been the subject of much re-
search activity, especially after realizing its practical necessity
after a number of high profile bank failures in Asia. As a re-
Manuscript received February 13, 2001; revised March 16, 2001 and March
25, 2001.
The author is with the California Institute of Technology, Pasadena, CA
91125 USA (e-mail: amir@caltech.edu).
Publisher Item Identifier S 1045-9227(01)05006-8.
sult, the regulators are acknowledging the need and are urging
the banks to utilize cutting edge technology to assess the credit
risk in their portfolios. Measuring the credit risk accurately also
allows banks to engineer future lending transactions, so as to
achieve targeted return/risk characteristics. The other benefit
of the prediction of bankruptcies is for accounting firms. If an
accounting firm audits a potentially troubled firm, and misses
giving a warning signal (say a “going concern” opinion), then it
faces costly lawsuits.
The traditional approach for banks for credit risk assessment
is to produce an internal rating, which takes into account var-
ious quantitative as well as subjective factors, such as leverage,
earnings, reputation, etc., through a scoring system [48]. The
problem with this approach is of course the subjective aspect
of the prediction, which makes it difficult to make consistent
estimates. Some banks, especially smaller ones, use the ratings
issued by the standard credit rating agencies, such as Moody’s
and Standard & Poor’s. The problem with these ratings is that
they tend to be reactive rather than predictive (for the agencies
to change a rating of a debt, they usually wait until they have
a considerably high confidence/evidence to support their deci-
sion). There is a need, therefore, to develop fairly accurate quan-
titative prediction models that can serve as very early warning
signals for counterparty defaults.
There are two main approaches to loan default/bankruptcy
prediction. The first approach, the structural approach, is based
on modeling the underlying dynamics of interest rates and firm
characteristics and deriving the default probability based on
these dynamics. The second approach is the empirical or the
statistical approach. Instead of modeling the relationship of
default with the characteristics of a firm, this relationship is
learned from the data. The focus of this article is on the empir-
ical approach, especially the use of NNs. In the next section
we give a review on this approach. To give a flavor about the
structural approach, it is also very briefly reviewed in the next
section. Section III presents some results of simulations that
we have performed, where we introduce novel inputs that lead
to considerable improvement in prediction accuracy. Section
IV is the summary and conclusion of this paper.
II. A R
EVIEW OF BANKRUPTCY PREDICTION MODELS
A. Early Empirical Approaches
The pioneers of the empirical approach are Beaver [7],
Altman [2], and Ohlson [34]. Beaver was one of the first re-
searchers to study the prediction of bankruptcy using financial
statement data. However, his analysis is very simple in that
it is based on studying one financial ratio at a time and on
1045–9227/01$10.00 © 2001 IEEE

930 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 12, NO. 4, JULY 2001
developing a cutoff threshold for each ratio. The approaches by
Altman and Ohlson are essentially linear models that classify
between healthy/bankrupt firms using financial ratios as inputs.
Altman uses the classical multivariate discriminant analysis
technique (MDA). It is based on applying the Bayes classifi-
cation procedure, under the assumption that the two classes
have Gaussian distributions with equal covariance matrices.
The covariance matrix and the class means are estimated from
the training set. Altman used the following financial ratios as
inputs:
1) working capital/total assets;
2) retained earnings/total assets;
3) earnings before interest and taxes/total assets;
4) market capitalization/total debt;
5) sales/total assets.
These particular financial ratios have been widely used as in-
puts, even for NNs and other nonlinear models. They are de-
scribed in more detail in the next subsection.
Ohlson introduced the logistic regression approach (LR) to
the bankruptcy prediction problem. It is essentially a linear
model with a sigmoid function
at the
output (it is thus similar to a single-neuron network). Because
the output is in between 0 and 1, the model has a nice proba-
bilistic interpretation. Ohlson used a novel set of financial ratios
as inputs. Both the MDA model and the LR model have been
widely used in practice and in many academic studies. They
have been standard benchmarks for the loan default prediction
problem.
B. Neural-Network (NN) Approaches
Research studies on using NNs for bankruptcy prediction
started in 1990, and are still active now. There are a number of
reasons why a nonlinear approach would be superior to a linear
approach. It can be argued that there are saturation effects in
the relationships between the financial ratios and the prediction
of default. For example, if the earnings/total assets changes
say by an amount of 0.2, from
0.1 to 0.1, it would have a far
larger effect (on the prediction of default) than it would if that
ratio changes from say 1.0 to 1.2. One can also argue that there
are multiplicative factors as well. For example, the potential for
default for a firm with negative cash flow gets more amplified if
it has large liabilities. The reason is that highly leveraged firms
have a harder time borrowing money to finance their deficits.
As will be seen from the review below, NNs have generally
outperformed the other existing methods. Currently, several
of the major commercial loan default prediction products are
based on NNs. For example, Moody’s Public Firm Risk Model
[32] is based on NNs as the main technology. Many banks have
also developed and are using proprietary NN default prediction
models.
The following is a review of the NN bankruptcy prediction
studies. There has been also a number of other review papers.
For example, Vellido et al. [52] survey the use of NNs in busi-
ness applications. This survey includes a section on bankruptcy
prediction. Also, the survey of Wong et al. [56] on NNs in busi-
ness applications includes some references on the bankruptcy
prediction problem. Dimitras et al. [15] provide a survey on
the classical empirical approaches. Zhang et al. [58] include in
their paper a nice review of existing work on NN bankruptcy
prediction. The majority of the NN approaches to default pre-
diction use multilayer networks. Since this is the dominant ap-
proach, henceforth when we mention NNs we mean multilayer
networks.
One of the first studies to apply NNs to the bankruptcy pre-
diction problem was the work by Odom and Sharda [33]. Odom
and Sharda used Altman’s financial ratios (described above) as
inputs to the NN, and applied their method, as well as MDA as
a comparison, to a number of bankrupt and solvent US firms,
where the data used for the bankrupt firms are from the last fi-
nancial statement before declaring bankruptcy. They considered
128 firms, and performed severalexperimentswhere they varied
the proportion of bankrupt/healthy firms in the training set. The
NN achieved a Type I correct classificationaccuracyinthe range
of 77.8% to 81.5% (depending on the training setup), and a Type
II accuracy in the range of 78.6% to 85.7%. The corresponding
results for MDA were in the range of 59.3% to 70.4% for Type I
acuracy, and in the range of 78.6% to 85.7% for Type II acuracy.
Let us now discuss why the particular indicators of [33]
(which are the same as Altman’s indicators) have been chosen.
Most other studies use indicators similar in nature, and the
analysis presented will somewhat apply to these studies as well.
A company’s total assets consists of current assets and long
term assets. The total assets gives some indication of the size
of the firm. Therefore it is frequently used as a normalizing
factor (like in indicators 1,2,3,5 of Altman’s indicators). The
current assets can or will typically be turned into money
fairly fast. The firm’s liabilities consists of current liabilities
and long term debt. The current liabilities include short term
loans (less than one year due), accounts payable, taxes due,
etc. The working capital is current assets minus the current
liabilities. It is an indication of the ability of the firm to pay
its short term obligations. If it is too negative, the company
might default on some payments. The firm’s total assets is
financed by a) the total liabilities and b) the shareholders’
equity [therefore the name “balance sheet,” since the total
assets have to exactly equal the sum of the two items in a) and
b)]. The shareholders’ equity consists of the capital raised in
share offerings and the retained earnings. The retained earnings
means the accumulation of the firm’s earnings since the firm’s
inception. The shareholders’ equity is also called the book
value of the firm. Even though it is based on the historical costs
(plus adjustments through depreciation/amortization) of the
firm’s assets and liabilities, rather than market values, it has
been a very useful indicator in assessing the financial health
of a firm. Retained earnings is a related and similarly useful
indicator. The firm’s earnings is also an important indicator.
Highly negative earnings indicate that the firm is losing its
competitiveness, and that geopardizes its survival. Another
related, widely used indicator is the cash flow. It is less prone
than earnings to management manipulation. In addition, it
measures directly the ability of the firm to generate cash to
retire debt. The rationale behind Altman’s fourth indicator is
the following. The firm can issue and sell new shares in the
market to repay its debt. A large market capitalization (relative
to the total debt) indicates a high capacity to perform that.

ATIYA: BANKRUPTCY PREDICTION FOR CREDIT RISK USING NEURAL NETWORKS 931
Finally, the firm’s sales is an indication of the health of its
business. However, this indicator is probably the least effective
among the five Altman indicators, because sales to total assets
can vary a lot from industry to industry.
Tam and Kiang [46], [47] considered the problem of bank
failure prediction. They compared between several methods:
MDA, LR, K-nearest neighbor (KNN), ID3 (a decision tree clas-
sification algorithm), single-layer network, and multilayer net-
work. For the case of one-year-ahead, the multilayer network
was the best, while for the case of two-year-ahead, LR was the
best. When they used a leave-one-out procedure instead of a
hold-out sample, the multilayer network was the clear winner
(for both forecast horizons). KNN and ID3 were almost always
inferior to the other methods.
Salchenberger et al. [41] considered the problem of pre-
dicting thrift failures. They compared NN with LR. The NN
significantly outperformed the LR. For example for 18-months
ahead prediction the LR achieves 83.3–85.4% accuracy (de-
pending on some threshold), whereas the NN achieves 91.7%.
Coats and Fant [12] compared between NN and MDA. They
obtained a classification accuracy in the range of 81.9% to
95.0% for the NN (depending on the horizon: from three-years
ahead to less than a year-ahead), and in the range of 83.7% to
87.9% for the MDA (also depending on the horizon).
Kerling and Poddig [23] compared NN with MDA for a data-
base of French firms for a three-year-ahead forecast. The NN
achieved a prediction accuracy in the range of 85.3–87.7% com-
pared to 85.7% for MDA. Kerling tested several cross-validation
procedures and early-stopping procedures in a follow-through
study [22].
Altman et al. [3] applied NN and MDA to a large database
of 1000 Italian firms for one-year ahead prediction. The com-
parison yielded no decisive winner, though MDA was slightly
better.
Boritz and Kennedy [9] (see also [10]) compared between a
number of techniques, including different NN training proce-
dures, LR and MDA, using the indicators chosen by Altman,
and those chosen by Ohlman. The results of the comparison are
inconclusive.
Fernandez and Olmeda [17] compared NN with MDA, LR,
MARS and C4.5 (two well known methods that are based on the
CART decision tree algorithm) on Spanish banks (no horizon
is specified). The NN obtained 82.4% accuracy compared with
61.8–79.4% for the competing techniques.
Alici [1] used principal component analysis and self-orga-
nizing maps for the input selection phase, together with a skele-
tonization step for the NN. He achieved an accuracy in the range
of 69.5% to 73.7% (dependinding on some parameter variation),
compared with 65.6% for MDA and 66.0% for LR for a data-
base of UK firms (no horizon is mentioned).
Leshno and Spector [27] used a novel NN architectures
containing cross-terms and cosine terms, and achieved pre-
diction accuracy for the two-years-ahead case in the range of
74.2–76.4% (depending on the order of the network), compared
with 72% for the linear perceptron network.
Lee et al. [25] propose hybrid models. Specifically, they
tested combinations of the models MDA, ID3, self-organizing
maps, and NN. They applied their study to the problem of
default prediction of Korean firms.
Back et al. [4] propose the use of genetic algorithms for input
selection, to be used in conjunction with multilayer networks.
They applied their method to data covering the periods one to
three years before the bankruptcy, where it obtains significant
improvement over MDA and LR.
Kiviluoto [24] use self-organizing maps on an extensive data-
base of Finnish firms (horizon is not specified), and show that it
obtains comparable results to MDA and learning vector quanti-
zation (in the range from 81% to 86%). Kaski et al. (this issue
[60]) developed a novel self-organizing map procedure based
on the Fisher metric, and applied it also to a number of Finnish
firms.
Zhang et al. [58] compared between NN and LR, and
employed a five-fold cross-validation procedure, on a sample
of manufacturing firms (horizion is not specified). They used
Altman’s five financial ratios plus the ratio current assets/cur-
rent liabilities as inputs to the NN. The NN significantly
outperformed LR with accuracy of 88.2% versus 78.6%.
Piramuthu et al. [37] developed a technique to construct sym-
bolic features, to be inputed to a multilayer network. They ap-
plied their technique to a collection of Belgian firms (no fore-
cast horizon is mentioned), where they obtained an accuracy of
82.9% versus 76.1% for the nontransformed input case. They
applied it also to a problem of one- and two-year ahead default
prediction for US banks. They get superior results, and signif-
icantly outperform the nontransformed input case. Piramuthu
[36] applies a similar input selection technique in conjunction
with decision tree classifiers.
Martinelli et al. [29] compared between two decision tree al-
gorithms, C4.5 and CN2, and NN on a database of Brazilian
firms. C4.5 outperform the other methods.
Yang et al. [57] used probabilistic NNs (PNNs) [45], which
essentially implement the Bayes classification rule. They
tested it on firms in the oil sector. The results were mixed:
PNN tied with the multilayer networks, but with a particular
preprocessing step MDA was the best.
McKee and Greenstein [30] developed a method based on
decision trees and applied it to a number of US firms for one
year ahead forecast. Their method obtains better results than NN
and MDA for Type II error, but worse results for Type I error.
Fan and Palaniswami [59] propose the use of support
vector machines (SVMs) for predicting bankruptcies among
Australian firms, and compared it with NN, MDA and learning
vector quantization (LVQ). SVM obtained the best results
(70.35%–70.90% accuracy depending on the number of inputs
used), followed by NN (66.11%–68.33%), followed by LVQ
(62.50%–63.33%), followed by MDA (59.79%–63.68%).
These reviewed papers are just a sample of what has been
done on the topic of NN default prediction. There are many
other studies (e.g., [5], [6], [11], [13], [16], [18], [21], [26], [35],
[38], [39], [42], [43], [44], [49], [50], [51], [53]), but for space
considerations they are not reviewed here.
C. A Brief Review of the Structural Approach
One of the earlier and commonly used methods is the asset-
based approach by Merton [31] (developed further by Longstaff

932 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 12, NO. 4, JULY 2001
and Schwartz [28]). This model views a firm’s equity as an op-
tion on the firm (held by the shareholders) to either repay the
debt of the firm when it is due, or abandon the firm without
paying the obligations. The probability of default can be de-
rived by modeling the market value of the firm as a geometric
Brownian motion. What makes that model successful is its re-
liance on the equity market as an indicator, since it can be ar-
gued that the market capitalization of the firm (together with the
firm’s liabilities) reflect the solvency of the firm. This model
has been successfully developed into a successful commercial
product by KMV Corporation.
Another approach, by Jarrow and Turnbull [20], models de-
fault as a point process, where the time-varying hazard func-
tion for each credit class is estimated from the credit spreads.
The CreditRisk+ product, developed by Credit Suisse Financial
Products, is also based on the same concept of modeling default
as a Poisson process.
Wilson [54], [55] proposed a discrete-time dynamical model,
whereby the default probabilities are a function of macro-eco-
nomic variables. J.P. Morgan’s CreditMetrics product [14] is
based on modeling changes in the credit quality ratings. By
modeling the “rating migrations,” one can obtain an estimate
for the probability of default. Several other models have been
proposed. For a more detailed review of structural credit risk
models refer to Crouhy
et al. [14].
D. Challenges for the NN Prediction Models
In spite of the success of NN models, there are a number of
open issues that should desirably be addressed by the research
community. Even though a prediction of the default event is by
itself very useful, an estimate of the default probabilty is very
desirable. For portfolio credit risk estimation, this is essential
in order to compute the loss level. (As described in the intro-
duction section, the loss level is the level for which there is a
probability of 1% that the loss incurred in the portfolio will ex-
ceed that level in a particular time period.) Also, typically banks
have several prediction systems in place. They make a lending
decision based on the combination of these predictions. Having
a probability of default rather than a (binary) prediction of de-
fault is valuable for them. Even though there are some objective
function measures that achieve that, such as cross-entropy error
function [8], our experience with this objective function has not
been very favorable.
The other open issue is to consider macroeconomic indicators
as inputs to the NN. The prevailing economic conditions (as
well as the current interest rates) can have a significant effect
on the probability of bankruptcy. There are very few studies
that consider these factors in conjunction with NN models. This
should therefore be a recommended study.
III. T
HE DEVELOPED BANKRUPTCY PREDICTION MODEL
In this work we introduce a novel set of indicators that can
be used in addition to the financial ratios and lead to significant
improvement in prediction accuracy. These indicators are ex-
tracted from the stock price of the firm. (We are inspired here by
Merton’s asset-based model, described in Section II-C, which
is based on information extracted from the equity markets.) It
is well known that the equity markets are very-early predictors
of shortfalls (or improvements) in the performance of a firm. A
problem faced by a firm will typically be reflected in the stock
price well before it shows up in its balance sheet and income
statement. As such, indicators obtained from the stock price can
be beneficial especially in long horizon default forecast. Ex-
amples of indicators tested are: volatility, change in volatility,
change in price, absolute price, price-cashflow ratio, etc. We de-
scribe the developed model below.
To test the comparative advantage of stock-price-based indi-
cators, we have developed two systems: one system based on fi-
nancial ratios alone (financial ratio system), and another based
on financial ratios and price-based indicators (financial ratio
and equity-based system). We will not compare here with linear
models such as MDA and LR, because that is not the objective
of the paper, and because there are so many previous studies
that have performed such a comparison (see Section II-C). The
NN is designed to predict default three-years-ahead, so it gives
a fairly long-horizon forecast. Each developed system consists
of two stages: the input selection stage, and the NN application
stage. We have considered a pool of about 120 candidate inputs
(financial statement data, ratios, stock price data, and transfor-
mations of these). Using an initial prescreening procedure based
on individual indicator prediction accuracy and correlation ma-
trix, and then a subsequent cross-validation procedure to narrow
down the choice, we select the best five or six inputs from this
pool of indicators. For the financial ratio system the chosen in-
dicators were:
1) book value/total assets BV/TA;
2) cashflow/total assets CF/TA;
3) rate of change of cashflow per share ROC(CF);
4) gross operating income/total assets GOI/TA;
5) return on assets ROA.
For the financial ratio and equity-based system the chosen indi-
cators are:
1) book value/total assets BV/TA;
2) cashflow/total assets CF/TA;
3) price/cashflow ratio P/CF;
4) rate of change of stock price ROC(P);
5) rate of change of cashflow per share ROC(CF);
6) stock price volatility VOL.
To test the system, we have collected historical data from de-
faultedand from solvent US firms.The defaulted firmdata cover
the period spanning 1 month to 36 months before the bank-
ruptcy event. (median time-to-default is 13 months). Note that
we have selected the solvent firms randomly (from among all
solvent firms), so the choice covers the whole spectrum from
healthy to border-line firms in order to avoid any selection bias.
We have considered 716 solvent firms and 195 defaulted firms.
We have performed the prediction for the defaulted firms at two
or three instants before default. The number of data points then
became 1160 (444 defaulted and 716 solvent). We note that
the size of the data set is quite large compared to the majority
of bankruptcy prediction studies. To our knowledge, only the
work by Altman et al.[3] uses a comparable size data set (1000
firms). The in-sample set consists of 491 data points, while the

ATIYA: BANKRUPTCY PREDICTION FOR CREDIT RISK USING NEURAL NETWORKS 933
TABLE I
R
ESULTS FOR THE NEURAL NETWORK DEFAULT PREDICTION MODEL:FINANCIAL
RATIO MODEL
TABLE II
R
ESULTS FOR THE NEURAL NETWORK
DEFAULT PREDICTION MODEL:FINANCIAL RATIO AND EQUITY-BASED MODEL
out-of-sample set consists of 669 data points. In case of mul-
tiple prediction instants for one firm, the firm’s data are either
all in the in-sample set or all in the out-of-sample set in order
to avoid bias. Note that we maintained a fixed ratio of number
of defaulted data points/number of solvent data points for both
in-sample and out-of-sample set (about 62%). The in-sample
data-set is used for the design of the input selection stage and
the NN design, while the out-of-sample set is reserved for the
final test of the system. Using the repeated random partitioning
procedure for the in-sample set into training set and validation
set, and repeated training and validation for the different parti-
tions, we determined optimal valuesof the different network and
learning parameters and performed the input selection. Based
on this tuning approach we selected a network of size 2 hidden
nodes. Since training a network till death for highly noisy ap-
plications can introduce some overfitting, we have used early
stopping. The best number of iterations is determined with the
help of the validation set to be 100.
Tables I and II show the results for both systems, along with
a break-down according to time till default. For the financial
ratio system we obtained a prediction accuracy of 84.52% for
the in-sample set, and 81.46% for the out-of-sample set. For
the financial ratio and equity-based system we obtained a pre-
diction accuracy of 89.41% for the in-sample set, and 85.50%
for the out-of-sample set. One can see that it outperforms by a
full 4 percentage points the financial ratio system, indicating the
value of indicators extracted from the equity markets. Note also
that its edge gets better for long horizon forecasts. It can clas-
sify significantly better data points that correspond to a large
time before default (for example more than 18 months). An ex-
planation of this observation is that financial statement data tend
to be lagging, since all the figures are reported by its book value.
Also, the stock market is highly predictive. It reflects qualitative
factors such as business conditions and insider information that
trickle through the market.
Table III showsthe correlation matrix for all 8 indicators used.
Of particular interest is the uniformly negative correlation of the
volatility indicator (VOL) with the other indicators. This makes
it a particularly useful indicator as part of the group, since it
might add discriminating powernot there in the other indicators.
IV. S
UMMARY AND CONCLUSION
In this article we reviewed the problem of bankruptcy predic-
tion using NNs. From the many studies existing in the literature,
it can be seen that NNs are generallymore superior to other tech-
niques. Once this is established, the logical next step for the re-
search community is to improve further the performance of NNs
for this application, perhaps through better training methods,
better architecture selection, or better inputs. It is this latter im-
provement aspect that we have addressed in the second half of

Citations
More filters
Journal ArticleDOI
TL;DR: A comprehensive review of the work done, during the 1968-2005, in the application of statistical and intelligent techniques to solve the bankruptcy prediction problem faced by banks and firms is presented.

978 citations


Cites background or methods from "Bankruptcy prediction for credit ri..."

  • ...[7] Firms USA 1160 BPNN with novel indicators 3 years [8] Firms Finnish 74 DA/DA, LA/LA, GA/BPNN, DA/ BPNN, LA/BPNN 1986–1989...

    [...]

  • ...Atiya [7] reviewed the applications of NN in bankruptcy prediction and developed an NN....

    [...]

  • ...[1] Cash flow/total liabilities, current assets/current liabilities (current ratio), inventories turnover, net income/total sales, net income/total assets, net working capital/total assets, owners equity/total assets, (total borrowings + bonds payable)/total assets [2] Net income/total assets (return on assets (ROA)), net loan losses/adjusted assets, net loan losses/total loan, (net loan losses + provision for loan losses)/income, non-performing loans/total assets [5] Common equity/total capital (capitalization), cumulative profitability, debt services, stability of earnings, roa, liquidity, size [6] Debt cost, debt quality, growth, indebtedness, share of labour costs, short-term liquidity, size, turnover of assets [7] Book value/total assets, cash flow/total assets, gross operating income/total assets, ROA, price/cash flow, rate of change of cash flow per share (ROC), rate of change of stock price, stock price volatility [8] Cash/current liabilities, cash/net sales, cash/total assets, cash flow/current liabilities, cash flow/total assets, cash flow/total debt, current ratio, current assets/net sales, current assets/total assets, current liabilities/equity, earnings before interest and taxes (EBIT)/total interest payments, equity/fixed assets, equity/net sales, inventory/net sales, long-term debt/equity, market value of equity/book value of debt, net income/total assets, net quick assets/inventory, net sales/total assets, operating income/total assets, quick assets/current liabilities, quick assets/net sales, quick assets/total assets, rate of return/common stock holders, retained earnings/total assets, return on stock, total debt/equity, total debt/total assets, working capital/equity, working capital/net sales, working capital/total assets...

    [...]

Journal ArticleDOI
TL;DR: The results demonstrate that the accuracy and generalization performance of SVM is better than that of BPN as the training set size gets smaller, and the several superior points of the SVM algorithm compared with BPN are investigated.
Abstract: This study investigates the efficacy of applying support vector machines (SVM) to bankruptcy prediction problem. Although it is a well-known fact that the back-propagation neural network (BPN) performs well in pattern recognition tasks, the method has some limitations in that it is an art to find an appropriate model structure and optimal solution. Furthermore, loading as many of the training set as possible into the network is needed to search the weights of the network. On the other hand, since SVM captures geometric characteristics of feature space without deriving weights of networks from the training data, it is capable of extracting the optimal solution with the small training set size. In this study, we show that the proposed classifier of SVM approach outperforms BPN to the problem of corporate bankruptcy prediction. The results demonstrate that the accuracy and generalization performance of SVM is better than that of BPN as the training set size gets smaller. We also examine the effect of the variability in performance with respect to various values of parameters in SVM. In addition, we investigate and summarize the several superior points of the SVM algorithm compared with BPN.

728 citations


Cites background or methods from "Bankruptcy prediction for credit ri..."

  • ...Keywords: Support vector machines; Bankruptcy prediction...

    [...]

  • ...…such as neural networks (NNs) can be an alternative method for classification problems to which traditional statistical method have long been applied (Atiya, 2001; Barniv, Agarwal, & Leach, 1997; Bell, 1997; Boritz & Kennedy, 1995; Charalambous, 0957-4174/$ - see front matter q 2004 Elsevier Ltd....

    [...]

  • ...…non-linear relationships in the data set, they have been studied extensively in the fields of financial problems including bankruptcy prediction (Atiya, 2001; Barniv et al., 1997; Bell, 1997; Boritz & Kennedy, 1995; Charalambous et al., 2000; Etheridge & Sriram, 1997; Fletcher & Goss, 1993;…...

    [...]

  • ...In this study, we show that the proposed classifier of SVM approach outperforms BPN to the problem of corporate bankruptcy prediction....

    [...]

Journal ArticleDOI
TL;DR: This paper extensively elaborates on the application of (1) univariate analysis, (2) risk index models, (3) multivariate discriminant analysis, and (4) conditional probability models, such as logit, probit and linear probability models.
Abstract: Over the last 35 years, business failure prediction has become a major research domain within corporate finance. Numerous corporate failure prediction models have been developed, based on various modelling techniques. The most popular are the classic cross-sectional statistical methods, which have resulted in various ‘single-period’ or static models, especially multivariate discriminant models and logit models. To date, there has been no clear overview and discussion of the application of classic statistical methods to business failure prediction. Therefore, this paper extensively elaborates on the application of (1) univariate analysis, (2) risk index models, (3) multivariate discriminant analysis, and (4) conditional probability models in corporate failure prediction. In addition, because there is no clear and comprehensive analysis in the existing literature of the diverse problems related to the application of these methods to the topic of corporate failure prediction, this paper brings together all problem issues and enlarges upon each of them. It discusses all problems related to: (1) the classical paradigm (i.e. the arbitrary definition of failure, non-stationarity and data instability, sampling selectivity, and the choice of the optimisation criteria); (2) the neglect of the time dimension of failure; and (3) the application focus in failure prediction modelling. Further, the paper elaborates on a number of other problems related to the use of a linear classification rule, the use of annual account information, and neglect of the multidimensional nature of failure. This paper contributes towards a thorough understanding of the features of the classic statistical business failure prediction models and their related problems.

691 citations

Journal ArticleDOI
TL;DR: This paper investigates the performance of a single classifier as the baseline classifier to compare with multiple classifier and diversified multiple classifiers by using neural networks based on three datasets and suggests it is better to consider these three classifier architectures to make the optimal financial decision.
Abstract: Bankruptcy prediction and credit scoring have long been regarded as critical topics and have been studied extensively in the accounting and finance literature. Artificial intelligence and machine learning techniques have been used to solve these financial decision-making problems. The multilayer perceptron (MLP) network trained by the back-propagation learning algorithm is the mostly used technique for financial decision-making problems. In addition, it is usually superior to other traditional statistical models. Recent studies suggest combining multiple classifiers (or classifier ensembles) should be better than single classifiers. However, the performance of multiple classifiers in bankruptcy prediction and credit scoring is not fully understood. In this paper, we investigate the performance of a single classifier as the baseline classifier to compare with multiple classifiers and diversified multiple classifiers by using neural networks based on three datasets. By comparing with the single classifier as the benchmark in terms of average prediction accuracy, the multiple classifiers only perform better in one of the three datasets. The diversified multiple classifiers trained by not only different classifier parameters but also different sets of training data perform worse in all datasets. However, for the Type I and Type II errors, there is no exact winner. We suggest that it is better to consider these three classifier architectures to make the optimal financial decision.

479 citations


Cites background or methods from "Bankruptcy prediction for credit ri..."

  • ...Keywords: Bankruptcy prediction; Credit scoring; Neural networks; Classifier ensembles...

    [...]

  • ...…intelligence and machine learning techniques (e.g. artificial neural networks (ANN), decision trees (DT), support vector machines (SVM), etc.) have been used to solve the above financial decision-making problems (e.g. Atiya, 2001; Huang, Chen, Hsu, Chen, & Wu, 2004; Lee, Chiu, Chou, & Lu, 2006)....

    [...]

  • ...Thus, predicting bankruptcy timely and correctly has become great importance for financial institutions (Atiya, 2001; Zhang, Hu, Patuwo, & Indro, 1999)....

    [...]

Journal ArticleDOI
TL;DR: This paper applied machine learning techniques to construct nonlinear nonparametric forecasting models of consumer credit risk, which significantly improved the classification rates of credit-card-holder delinquencies and defaults with linear regression R-squared's of forecasted/realized delinquencies of 85%.
Abstract: We apply machine-learning techniques to construct nonlinear nonparametric forecasting models of consumer credit risk. By combining customer transactions and credit bureau data from January 2005 to April 2009 for a sample of a major commercial bank's customers, we are able to construct out-of-sample forecasts that significantly improve the classification rates of credit-card-holder delinquencies and defaults, with linear regression R-squared's of forecasted/realized delinquencies of 85%. Using conservative assumptions for the costs and benefits of cutting credit lines based on machine-learning forecasts, we estimate the cost savings to range from 6% to 25% of total losses. Moreover, the time-series patterns of estimated delinquency rates from this model over the course of the recent financial crisis suggests that aggregated consumer-credit risk analytics may have important applications in forecasting systemic risk.

428 citations


Cites methods from "Bankruptcy prediction for credit ri..."

  • ...Related approaches have been used by Atiya (2001), Shin et al. (2005), and Min et al. (2005), in which artificial neural networks (ANN) and support vector machines (SVM) were applied to the problem of predicting corporate bankruptcies....

    [...]

References
More filters
Book ChapterDOI
TL;DR: The chapter discusses two important directions of research to improve learning algorithms: the dynamic node generation, which is used by the cascade correlation algorithm; and designing learning algorithms where the choice of parameters is not an issue.
Abstract: Publisher Summary This chapter provides an account of different neural network architectures for pattern recognition. A neural network consists of several simple processing elements called neurons. Each neuron is connected to some other neurons and possibly to the input nodes. Neural networks provide a simple computing paradigm to perform complex recognition tasks in real time. The chapter categorizes neural networks into three types: single-layer networks, multilayer feedforward networks, and feedback networks. It discusses the gradient descent and the relaxation method as the two underlying mathematical themes for deriving learning algorithms. A lot of research activity is centered on learning algorithms because of their fundamental importance in neural networks. The chapter discusses two important directions of research to improve learning algorithms: the dynamic node generation, which is used by the cascade correlation algorithm; and designing learning algorithms where the choice of parameters is not an issue. It closes with the discussion of performance and implementation issues.

13,033 citations

Journal ArticleDOI
TL;DR: In this article, the American Finance Association Meeting, New York, December 1973, presented an abstract of a paper entitled "The Future of Finance: A Review of the State of the Art".
Abstract: Presented at the American Finance Association Meeting, New York, December 1973.(This abstract was borrowed from another version of this item.)

11,225 citations

Journal ArticleDOI
TL;DR: In this paper, a set of financial and economic ratios are investigated in a bankruptcy prediction context wherein a multiple discriminant statistical methodology is employed, and the data used in the study are limited to manufacturing corporations, where an initial sample of sixty-six firms is utilized to establish a function which best discriminates between companies in two mutually exclusive groups: bankrupt and nonbankrupt firms.
Abstract: ACADEMICIANS SEEM to be moving toward the elimination of ratio analysis as an analytical technique in assessing the performance of the business enterprise. Theorists downgrade arbitrary rules of thumb, such as company ratio comparisons, widely used by practitioners. Since attacks on the relevance of ratio analysis emanate from many esteemed members of the scholarly world, does this mean that ratio analysis is limited to the world of \"nuts and bolts\"? Or, has the significance of such an approach been unattractively garbed and therefore unfairly handicapped? Can we bridge the gap, rather than sever the link, between traditional ratio \"analysis\" and the more rigorous statistical techniques which have become popular among academicians in recent years? The purpose of this paper is to attempt an assessment of this issue-the quality of ratio analysis as an analytical technique. The prediction of corporate bankruptcy is used as an illustrative case.' Specifically, a set of financial and economic ratios will be investigated in a bankruptcy prediction context wherein a multiple discriminant statistical methodology is employed. The data used in the study are limited to manufacturing corporations. A brief review of the development of traditional ratio analysis as a technique for investigating corporate performance is presented in section I. In section II the shortcomings of this approach are discussed and multiple discriminant analysis is introduced with the emphasis centering on its compatibility with ratio analysis in a bankruptcy prediction context. The discriminant model is developed in section III, where an initial sample of sixty-six firms is utilized to establish a function which best discriminates between companies in two mutually exclusive groups: bankrupt and non-bankrupt firms. Section IV reviews empirical results obtained from the initial sample and several secondary samples, the latter being selected to examine the reliability of the discriminant

10,737 citations

Journal ArticleDOI
TL;DR: In this paper, the authors present some empirical results of a study predicting corporate failure as evidenced by the event of bankruptcy, and the methodology is one of maximum likelihood estimation of the so-called conditional logit model, in which the data set used in this study is from the seventies (1970-76).
Abstract: This paper presents some empirical results of a study predicting corporate failure as evidenced by the event of bankruptcy. There have been a fair number of previous studies in this field of research; the more notable published contributions are Beaver [1966; 1968a; 1968b], Altman [1968; 1973], Altman and Lorris [1976], Altman and McGough [1974], Altman, Haldeman, and Narayanan [1977], Deakin [1972], Libby [1975], Blum [1974], Edmister [1972], Wilcox [1973], Moyer [1977], and Lev [1971]. Two unpublished papers by White and Turnbull [1975a; 1975b] and a paper by Santomero and Vinso [1977] are of particular interest as they appear to be the first studies which logically and systematically develop probabilistic estimates of failure. The present study is similar to the latter studies, in that the methodology is one of maximum likelihood estimation of the so-called conditional logit model. The data set used in this study is from the seventies (1970-76). I know of only three corporate failure research studies which have examined data from this period. One is a limited study by Altman and McGough [1974] in which only failed firms were drawn from the period 1970-73 and only one type of classification error (misclassification of failed firms) was analyzed. Moyer [1977] considered the period 1965-75, but the sample of bankrupt firms was unusually small (twenty-seven firms). The

5,244 citations

Journal ArticleDOI
TL;DR: In this article, the authors focus on the use of ratios as predictors of failure, defined as the inability of a firm to pay its financial obligations as they mature, and demonstrate that a firm is said to have failed when any of the following events have occurred.
Abstract: At the turn of the century, ratio analysis was in its embryonic state. It began with the development of a single ratio, the current ratio,' for a single purpose-the evaluation of credit-worthiness. Today ratio analysis involves the use of several ratios by a variety of users-including credit lenders, credit-rating agencies, investors, and management.2 In spite of the ubiquity of ratios, little effort has been directed toward the formal empirical verification of their usefulness. The usefulness of ratios can only be tested with regard to some particular purpose. The purpose chosen here was the prediction of failure, since ratios are currently in widespread use as predictors of failure. This is not the only possible use of ratios but is a starting point from which to build an empirical verification of ratio analysis. "Failure" is defined as the inability of a firm to pay its financial obligations as they mature. Operationally, a firm is said to have failed when any of the following events have occurred: bankruptcy, bond default, an overdrawn bank account, or nonpayment of a preferred stock dividend.3 A "financial ratio" is a quotient of two numbers, where both num-

4,210 citations


"Bankruptcy prediction for credit ri..." refers background in this paper

  • ...The pioneers of the empirical approach are Beaver [7],...

    [...]

  • ...The pioneers of the empirical approach are Beaver [7], Altman [2], and Ohlson [34]....

    [...]

  • ...Beaver was one of the first researchers to study the prediction of bankruptcy using financial statement data....

    [...]

Frequently Asked Questions (11)
Q1. What are the main indicators for the financial ratio system?

For the financial ratio system the chosen indicators were:1) book value/total assets BV/TA; 2) cashflow/total assets CF/TA; 3) rate of change of cashflow per share ROC(CF); 4) gross operating income/total assets GOI/TA; 5) return on assets ROA. 

For the financial ratio and equity-based system the chosen indicators are:1) book value/total assets BV/TA; 2) cashflow/total assets CF/TA; 3) price/cashflow ratio P/CF; 4) rate of change of stock price ROC(P); 5) rate of change of cashflow per share ROC(CF); 6) stock price volatility VOL. 

Once this is established, the logical next step for the research community is to improve further the performance of NNs for this application, perhaps through better training methods, better architecture selection, or better inputs. 

For the financial ratio system the authors obtained a prediction accuracy of 84.52% for the in-sample set, and 81.46% for the out-of-sample set. 

Wilson [54], [55] proposed a discrete-time dynamical model, whereby the default probabilities are a function of macro-economic variables. 

In this work the authors introduce a novel set of indicators that can be used in addition to the financial ratios and lead to significant improvement in prediction accuracy. 

For the financial ratio and equity-based system the authors obtained a prediction accuracy of 89.41% for the in-sample set, and 85.50% for the out-of-sample set. 

Fan and Palaniswami [59] propose the use of support vector machines (SVMs) for predicting bankruptcies among Australian firms, and compared it with NN, MDA and learning vector quantization (LVQ). 

Using an initial prescreening procedure based on individual indicator prediction accuracy and correlation matrix, and then a subsequent cross-validation procedure to narrow down the choice, the authors select the best five or six inputs from this pool of indicators. 

They applied their technique to a collection of Belgian firms (no forecast horizon is mentioned), where they obtained an accuracy of 82.9% versus 76.1% for the nontransformed input case. 

The corresponding results for MDA were in the range of 59.3% to 70.4% for Type The authoracuracy, and in the range of 78.6% to 85.7% for Type II acuracy.