Next Article in Journal
Bank Crisis Boosts Bitcoin Price
Previous Article in Journal
A New Proxy for Estimating the Roughness of Volatility
Previous Article in Special Issue
Evaluation of Weather Yield Index Insurance Exposed to Deluge Risk: The Case of Sugarcane in Thailand
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Artificial Intelligence Techniques for Bankruptcy Prediction of Tunisian Companies: An Application of Machine Learning and Deep Learning-Based Models

1
International Finance Group Tunisia Lab, Faculty of Management and Economic Sciences of Tunis, University of Tunis El Manar, Tunis 2092, Tunisia
2
Applied Economics and Simulation, Faculty of Management and Economic Sciences of Mahdia, University of Monastir, Rue Ibn Sina Hiboun, Mahdia 5111, Tunisia
3
Laboratory of Engineering Mathematics (LR01ES13), Tunisia Polytechnic School, University of Carthage, Tunis 2078, Tunisia
4
Department of Advanced Sciences and Technologies, National School of Advanced Sciences and Technologies of Borj Cedria, University of Carthage, Hammam-Chott 1164, Tunisia
*
Author to whom correspondence should be addressed.
J. Risk Financial Manag. 2024, 17(4), 132; https://doi.org/10.3390/jrfm17040132
Submission received: 5 January 2024 / Revised: 10 March 2024 / Accepted: 14 March 2024 / Published: 22 March 2024

Abstract

:
The present paper aims to compare the predictive performance of five models namely the Linear Discriminant Analysis (LDA), Logistic Regression (LR), Decision Trees (DT), Support Vector Machine (SVM) and Random Forest (RF) to forecast the bankruptcy of Tunisian companies. A Deep Neural Network (DNN) model is also applied to conduct a prediction performance comparison with other statistical and machine learning algorithms. The data used for this empirical investigation covers 25 financial ratios for a large sample of 732 Tunisian companies from 2011–2017. To interpret the prediction results, three performance measures have been employed; the accuracy percentage, the F1 score, and the Area Under Curve (AUC). In conclusion, DNN shows higher accuracy in predicting bankruptcy compared to other conventional models, whereas the random forest performs better than other machine learning and statistical methods.

1. Introduction

Predicting bankruptcy has always been of great importance and a huge challenge for banks and lending institutions. Therefore, financial analysts and credit experts look for the best techniques that can help them in decision making. For a long time, the traditional approaches have been widely used for bankruptcy prediction. These techniques are based on the financial ratios analysis, statistical models, and expert judgment. However, these models have limitations in predicting bankruptcy accurately (Hamdi 2012; Altman et al. 1994; Hamdi and Mestiri 2014).
Over recent years, several research studies have been focused on bankruptcy forecasting using artificial intelligence and machine learning models. The research paper of Ravi Kumar and Ravi (2007) summarizes existing researches on bankruptcy prediction studies using statistical and intelligence techniques during 1968–2005. For the same objective, Gergely (2015) has also presented a rich bibliographic review. He summarizes the short evolution of bankruptcy prediction and presents the main critiques made on modeling process for bankruptcy prediction. Furthermore, the author announces avenues of future research recommended in these studies. More recently, a systematic literature was presented by Clement (2020) to predict bankruptcy. His review was conducted based on published papers between 2016 and 2020. In the same context, Kuizinienė et al. (2022) present another systematic review covering 232 research studies spanning from 2017 to February 2022 that use artificial intelligence techniques to identify financial distress.
A more advanced model is applied in this study, specifically, the concept of deep learning. For more details about deep learning approaches refer to the studies of Deng and Yu (2014) and LeCun et al. (2015). Deep learning approaches have been extensively employed in the field of computer vision (Kamruzzaman and Alruwaili 2022), speech recognition (Roy et al. 2021), natural language programming (Xie et al. 2018), and medical image analysis (Suganyadevi et al. 2022). However, few are the studies which have been focused on the use of deep learning in finance (Qu et al. 2019).
This study is organized as follows: Section 2 provides a pertinent literature review related to bankruptcy prediction. Section 3 presents the different statistical and artificial intelligence techniques applied in this work. The data used are identified in Section 4. The Section 5 is devoted to the empirical investigation to predict the bankruptcy of Tunisian companies. And finally, the conclusion of this research study is presented in Section 6.

2. Related Literature

In past decades, the discriminant approach (Beaver 1966; Altman 1968; Deakin 1972) and the logistic regression method (Ohlson 1980; Pang 2006) were the two well-known and most popular statistical methods for predicting corporate bankruptcy. More recently, Mestiri and Hamdi (2013) used the logistic regression with random effect to predict the credit risk of Tunisian banks. For bankruptcy prediction, several more developed methods have been employed. Some authors apply the decision trees method (Aoki and Hosonuma 2004; Zibanezhad et al. 2011; Begović and Bonić 2020), some others utilize various machine learning techniques such as genetic algorithm (Shin and Lee 2002; Kim and Han 2003; Davalos et al. 2014), support vector machine (Shin et al. 2005; Härdle et al. 2005; Dellepiane et al. 2015) and random forest (Joshi et al. 2018; Ptak-Chmielewska and Matuszyk 2020; Gurnani et al. 2021). Recently, several comparative analyses of machine learning models have been carried out to predict bankruptcy (Narvekar and Guha 2021; Park et al. 2021; Bragoli et al. 2022; Máté et al. 2023; Martono and Ohwada 2023).
As a matter of fact, with the invasion of the artificial intelligence modeling algorithms since the 1990s in diverse domains, artificial neural networks were the most famous and well-used machine learning tool to predict financial distress (Odom and Sharda 1990; Atiya 2001; Anandarajan et al. 2004; Hamdi 2012; Aydin et al. 2022). However, despite the good forecasting results observed by applying this tool, deep learning models are the most applied today. This comes down to the ability of deep learning approach to overcome some limitations by training the neural network which includes a significant number of hidden layers, such as the vanishing gradient, overfitting problem and the computational load (Kim 2017).
Until now, few are the works which have been focused on applying deep learning models to predict bankruptcy. Addo et al. (2018) used seven methods (LR, RF, boosting approach and 4 deep learning models) to predict loan default probability. Based on AUC and RMSE performance criteria, they concluded that the gradient boosting model outperforms the other models in solving the binary classification problem. In another study, Hosaka (2019) proposed a convolutional neural network to forecast the bankruptcy of Japanese firms. This model is specifically effective for image recognition, therefore the author has converted the financial ratios in order to train and test the network. The prediction performance results showed higher performance with the use of deep neural network compared to other employed tools.
For the same purpose, Noviantoro and Huang (2021) used machine learning as well as deep learning approaches to predict bankruptcy of Taiwanese companies between 1999 and 2009. They compared the best prediction performance of decision tree, random forest, k-nearest neighbour algorithm, support vector machine, artificial neural network, Naïve bayes, logistic regression, rule induction and deep neural network. To evaluate the classifier’s performance of these models, they computed the accuracy rate, F score and AUC of each technique. They found that random forest demonstrated the highest accuracy and AUC, as well as the highest F score, and this was followed by the deep learning approach.
Very recently, Shetty et al. (2022) utilized deep neural network, extreme gradient boosted tree and support vector machine in order to predict the bankruptcy of 3728 Belgian firms for the period from 2002 to 2012. The authors concluded that the use of these different techniques yields roughly the same bankruptcy prediction accuracy rate of approximately 82–83%. Elhoseny et al. (2022) applied an adaptive whale optimization algorithm combined with deep learning (AWOA-DL) to predict bankruptcy. They evaluated the ability of the proposed new approach, to predict the failure of any company compared to logistic regression, the RBF Network, the teaching-learning-based optimization-DL (TLBO-DL) and the deep neural network. The empirical results show that the new deep learning-based approach (AWOA-DL) allows better predictions. More recently, Ben Jabeur and Serret (2023) proposed a Fuzzy Convolutional Neural Networks (FCNN) to predict corporate financial distress. They used eight evaluation measures in order to compare the performance of the new adopted method to other traditional and machine learning techniques. They found that the combined new approach outperforms traditional methods. In another study, Noh (2023) tested the accuracy performance of Long Short-Term Memory (LSTM), Logistic Regression (LR), K-Nearest Neighbour (k-NN), Decision Tree (DT), and Random Forest (RF) models for corporate bankruptcy prediction. On the basis of five performance measures, the author concluded that the proposed technique can enhance the prediction accuracy by using a small sample of an unbalanced financial dataset.
Table 1 provides a literature review summary of the main research studies that apply deep learning to predict bankruptcy.

3. Statistical, Machine Learning and Deep Learning Techniques

3.1. Linear Discriminant Analysis (LDA)

Ronald Fisher (1933) pioneered work on discriminant analysis. In his work, he developed a statistical technique for defaults prediction, by developing a linear combination of quantitative predictor variables. The output of LDA is a score that classifies data observations between the good and bad classes.
Score = i = o p a i X i
Score = a0 + a1X1 + a2X2 + …… + apXp
where ai: are the weights associated with the quantitative input variables Xi.
The study of Altman (1968) is considered as the reference work that uses the LDA to classify default and health companies based on five financial ratios.

3.2. Logistic Regression (LR)

LR is a statistical method used for binary classification tasks (e.g., 0 or 1, bad or good, health or default, etc.). Corresponding to Ohlson (1980), the outcome of the LR model can be written as:
P ( y = 1 | X ) = s i g m o i d ( z ) = 1 1 + e x p ( z )
where   P ( y = 1 | X ) is the probability of y being 1, given the input variables X, z is a linear combination of X: z = a 0 + a 1 X 1 + a 2 X 2 + + a p X p .
Where a 0 is the intercept term, a 1 , a 2 , , a p are the weights, and X 1 , X 2 , , X p are the inputs.

3.3. Decision Trees (DT)

DTs proceed recursively partitioning the data into subsets based on the values of the input variables, with each partition represented by a branch in the tree (Quinlan 1986). The function of DTs is aimed at training a sequence of binary decisions that can be utilized to forecast the value of the output for a new observation. In the tree, each decision node corresponds to a test of value for one of the input variables, and the branches correspond to the possible outcomes of the test. The leaves of the tree denote the predicted values of the output variable for each combination of input values. For each step, the algorithm identifies the input variable that provides the best split of the data into two subsets which are as homogeneous as possible in relation to the output variable. The quality of a split is typically measured using information gain or Gini impurity, which quantifies the reduction in uncertainty about the output variable achieved by the split.
Decision trees are typically not formulated in terms of mathematical equations, but rather as a sequence of logical rules that describe how the input variables are used to predict the output variable. However, the splitting standard utilized to select the best split at each decision node can be expressed mathematically. Suppose having a dataset with n observations and m input variables, denoted by X 1 , X 2 , , X p , and a binary output variable y that takes values in 0.1. Let S be a subset of the data at a particular decision node, and let p i be the part of observations in S that belong to class i. The Gini impurity of S is calculated as follows:
G ( S ) = 1 i ( p i ) 2
The Gini impurity measures the probability of misclassifying an observation in S if randomly assign it to a class corresponding to the observations proportion for each class. (Gelfand et al. 1991). A small value of G(S) indicates that the observations in S are well-separated by the input variables.
To split the data at a decision node, consider all possible splits of each input variable into two subsets, and choose the split that minimizes the weighted sum of the Gini impurities of the resulting subsets. The weighted sum is given by:
Δ G = G ( S ) ( | S 1 | | S | ) · G ( S 1 ) ( | S 2 | | S | ) · G ( S 2 )
where S 1 and S 2 are the subsets of S resulting from the split, and | S 1 | and | S 2 | are their respective sizes. The split with the smallest value of Δ G is chosen as the best split. The decision tree algorithm proceeds recursively, splitting the data at each decision node based on the best split, until a stopping criterion is met, such as reaching a maximum depth or minimum number of observations at a leaf node.

3.4. Support Vector Machine (SVM)

SVM is a supervised learning model used for classification, regression, and outlier detection, developed by Vapnik (1998). The basic idea of this technique is to determine the best separating hyperplane between two classes in a given dataset. The mathematical formulation of SVM is divided into two parts: optimization problem and decision function (Hearst et al. 1998).
Given a training set ( x i , y i ) where x i is the ith input vector and y i is the corresponding output: y i = ( 1 , 1 ) . Then, SVM seeks to find the best separating hyperplane defined by:
w · x + b = 0
where w is the weight vector, b is the bias term, and x is the input vector.
SVM algorithm aims to determine the optimal w and b that maximize the margin between two classes. The margin is the distance between the hyperplane and the nearest data point from either class. Then, SVM optimization problem can be formulated as:
minimize   1 2 w 2 + C i = 1 n ξ i   subject   to   y i ( w t x i + b ) 1 ξ i   and   ξ i 0
where | | w | | 2 is the L2-norm of the weight vector, C is a hyperparameter that controls the tradeoff between maximizing the margin and minimizing the classification error, ξ i is the slack variable that allows for some misclassifications, and the two constraints enforce that all data points lie on the correct side of the hyperplane with a margin of at least 1 ξ i .
The optimization problem can be solved by using convex optimization methods, for example the quadratic programming. Once the optimization problem is solved, the decision function can be defined as:
f ( x ) = s i g n ( w · x + b )
where sign is the sign function that returns +1 or −1 depending on the sign of the argument. The decision function takes an input vector x and returns its predicted class label based on whether the output of the hyperplane is positive or negative. For more details about the optimization process, refer to (Chang and Lin 2011; Cristianini and Shawe-Taylor 2000; Gunn 1998).
Thereafter, SVM finds the best separating hyperplane by solving an optimization problem that maximizes the margin between the two classes, subject to constraints that ensure all data points are correctly classified with a margin of at least 1 ξ i . The decision function then predicts the class label of new data points based on the output of the hyperplane.

3.5. Random Forests (RF)

RF is an ensemble of learning algorithm. It is a type of ensemble learning algorithm, developed by Breiman (2001), which combines multiple decision trees to make predictions. The algorithm is called “random” because it uses random subsets of the features and random samples of the data to build the individual decision trees. The data is split into training and testing sets. The training set is used to build the model, and the testing set is used to evaluate its performance. At each node of a decision tree, the algorithm selects a random subset of the features to consider when making a split. This helps to reduce overfitting and increase the diversity of the individual decision trees.
A decision tree is built using the selected features and a subset of the training data. The tree is grown until it reaches a pre-defined depth or until all the data in a node belongs to the same class. Suppose having a dataset with n observations and p features. Let X be the matrix of predictor variables and Y be the vector of target variables.
To build an RF model, start by creating multiple decision trees using a bootstrap sample of the real data. This means that we randomly sample n observations from the dataset with replacement to create a new dataset, and for k times this process is repeated to create k bootstrap samples. For each bootstrap sample, we then create a decision tree using random subsets of p features. For each node of the tree, we select the optimal feature and threshold value to divide the data based on a criterion, for example; the information gain or Gini impurity. We repeat the mentioned steps k times to create k decision trees. To make a prediction for a new observation, we pass it through each of the k decision trees and therefore obtain k predictions. For more details about the technical analysis of random forests, see Biau (2012).

3.6. Deep Neural Network (DNN)

DNN is an enhanced version of the conventional artificial neural network with at least two hidden layers (Schmidhuber 2015). Figure 1 illustrates the standard architecture of deep neural network.
To fully understand how DNN works, a thorough knowledge of the basics of artificial neural network is then necessary. For more information, readers can look at the studies of Walczak and Cerpa (2003) and Zou et al. (2008). According to Addo et al. (2018), the DNN output is computed as:
y ( t ) = k = 1 L f ( w k + x k ( t ) ) + ε ( t )
where Wk is the matrix weights of the layer, Xk (k = 1, …, L) is the total number of sequence of real values called events during an epoch and f is the activation function.

4. Data

A series of financial ratios was calculated using balance sheets and income statements of 732 firms from different sectors of activity for the period between 2011–2017. A total of 4925 credit files, provided by a private Tunisian bank, constitute the database used in this empirical study. Table 2 presents the input ratios.
In our research study, the same financial ratios considered by the previous works (Hamdi 2012; Mestiri and Hamdi 2013; Hamdi and Mestiri 2014) are used and demonstrated a high prediction accuracy in predicting bankruptcy of Tunisian firms. We excluded only one non-significant ratio (Raw stock/Total assets) in our empirical investigation.
On the other hand, the estimated output (Y) can be written as binary values:
Y = { 1     f o r     d e f a u l t     f i r m 0     f o r     h e a l t h y     f i r m
Following this classification criterion, the out-of-sample test is composed of 488 healthy companies and 244 are bankrupt companies.

5. Empirical Investigation

5.1. Predictive Performance Measures

There are several criteria that can be utilized to compare and evaluate the predictive ability of the employed techniques including accuracy rate, F1 score and AUC.

5.1.1. Accuracy Rate

The accuracy rate is the most famous performance metric, deduced from the confusion matrix (see Table 3) and calculated following this formula:
A c c u r a c y   r a t e = ( T 0 + T 1 ) ( T 0 + F 1 ) + ( F 0 + T 1 )

5.1.2. F1 Score

The F1 score is also computed from the confusion matrix. The value of F1 score varies between 0 and 1, since 1 is the best possible score. The model can correctly identify positive and negative cases with a high F1-score, meaning that the model has high precision and high recall.
  F 1   score = 2 ( P r e c i s i o n R e c a l l ) ( P r e c i s i o n + R e c a l l )
where R e c a l l = T 0 T 0 + F 0 and P r e c i s i o n = T 0 T 0 + F 1 .

5.1.3. AUC

Area Under Curve (AUC) is a synthetic indicator derived from the ROC curve. This curve is a graphical indicator utilized to assess the model forecasting accuracy (Pepe 2000; Vuk and Curk 2006). Specificity and sensitivity are the two relevant indicators on which ROC curve is based (see Zweig and Campbell 1993 and Mestiri and Hamdi 2013 for further details). This curve is characterized by the 1-specificity rate on the x axis and by sensitivity on the y axis. Where
S e n s i t i v i t y = T r u e   p o s i t i v e   r a t e = T 0 P o s i t i v e s = T 0 T 0 + F 1
and
S p e c i f i c i t y = T r u e   n e g a t i v e   r a t e = T 1 N e g a t i v e s = T 1 T 1 + F 0
Moreover, AUC measure reflects the quality of the model classification between heath and default firms. In the ideal case, AUC is equal to 1, i.e., the model makes it possible to completely separate all the positives from the negatives, without false positives or false negatives.

5.2. Results &Discussion

Table 4 presents the empirical results of the accuracy rate, F1 score and AUC criteria used to judge the classifier’s performance of the applied methods.
According to Table 4, the deep neural network significantly outperforms other techniques. DNN shows the highest accuracy rate with 93.6% whereas 88.2% for RF and 85.8% for LR. The lowest rate of prediction accuracy was found by the use of DT (74.3%). For the same objective to assess the predictive ability of the proposed algorithms, F1-score equal to 0.964 proves DNN’s ability to identify with a great precision healthy companies from bankrupt companies. Since 1 is the best desired F1 score, DNN reaches the highest score while F1 score values were equal to 0.933, 0.922, 0.910, 0.890 and 0.838 for RF, LR, SVM, LDA and DT, respectively.
Another graphical indicator was also used to evaluate the quality of classification of the models under study, is the ROC curve (see Figure 2). The AUC measure is deduced from this curve. A model with AUC value near to unity shows high quality of classification between health and default firms. Based on Table 4, the AUC of DNN yields 0.888. In the second rank, RF was found with AUC equals to 0.815. The RL and ADL models present the worst classification results as the AUC is 0.633 and 0.574, respectively, in the testing sample.
Similar conclusions were provided by Hosaka (2019). The study’s findings indicate that the convolutional neural network has better prediction performance than statistical and conventional machine learning methods. Furthermore, the work of Efron (1975) proved the robustness of the LR model compared to the LDA. Barboza et al. (2017) obtained similar results in predicting bankruptcy of North American firms. Their empirical findings indicate that RF is the most accurate prediction model compared to LR and ADL. They found that RF reaches 87% accuracy, whereas LR reach 69%and LDA reach 50%.
As a final conclusion, the ability of DNN outperforms the traditional statistical models and the conventional machine learning techniques in forecasting bankruptcy. In the second rank, RF has a significantly higher prediction accuracy compared to other employed techniques. Based on our empirical investigation, the DNN can be considered as the best technique to detect a company’s financial distress and therefore can help to make managerial decisions.
In our empirical study, we have used 20% of the sample (985 firms) as a test data set in order to check the prediction accuracy and classifier’s quality of the models. The type of deep neural network used in our study is a recurrent neural network with three hidden layers. Nodes per layer are 200,100,40,1(‘output’ layer). Activation function is ReLU and Loss function is binary cross entropy. The output unit is Sigmoid. Backpropagation training algorithm was used and a stopping criteria equal to 10−3 was set.

6. Conclusions

There are considerable consequences of a company’s financial default on several financial and economic actors such as investors, creditors, managers, shareholders, financial analysts, auditors, employees and government. Prediction bankruptcy has become of great importance and concern. By developing accurate bankruptcy prediction techniques, many advantages and benefits can be achieved, such as cost reduction and rapidity in recovery and credit file analysis, gaining time and better reimbursement monitoring of loan files. The machine learning models are widely used and applied in the literature of bankruptcy prediction. These models demonstrate performance in terms of prediction accuracy which explains our choice to adopt these models and compare them with the deep learning approach. The main contribution of this present work is to identify the appropriate model able to predict financial distress with high precision in the Tunisian context.
Statistical, machine learning and deep learning models such as the ADL, LR, DT, SVM, RF and DNN are applied to predict the financial distress of 732 Tunisian companies from different activity sectors. The empirical findings showed that DNN is a highly suitable tool for studying financial distress in Tunisian credit institutions. Compared to past work, this study is distinguished from other references in predicting bankruptcy that employed an interesting number of input features (25 ratios) as well as a large sample of firms in training phase (3940 ≈ 80% of total sample of firms). Wilson and Sharda (1994) used only five ratios (same input ratios employed by Altman 1968) to predict the bankruptcy of 169 firms. The machine learning models applied in their work are the shallow neural network and multi-discriminant analysis. In a related study, Chen (2011) utilized a set of eight selected features as inputs of machine learning models and an evolutionary computation approach was used for predicting business failure of 200 Taiwanese companies. To forecast the bankruptcy of Korean construction companies, Heo and Yang (2014) used a total of 2762 samples and 12 ratios for training several models such as adaptive boosting with DT, SVM, DT and ANN. For future research studies, we can apply hybrid learning techniques by combining the DNN with other machine learning model which can provide higher performance than when using a single model. In this context and for the same purpose to forecast bankruptcy, Ben Jabeur and Serret (2023) utilized the fuzzy convolutional neural networks. The present work as well as previous research supports the idea that artificial intelligence models perform better than traditional methods. However, it will be interesting for further research to diversify the data sources and not only use standard financial ratio data, by adding miscellaneous textual data (e.g., news, companies’ public report, notes and comments from experts, auditors’ reports and managements’ statements) that can enhance the forecasting accuracy of financial distress (Mai et al. 2019; Matin et al. 2019). Furthermore, it is of great interest to integrate sector diversification as an input variable to predict company default and to subsequently study the impact of changing industry on the accuracy of predictions. Another concern that should be studied in the future, is the occurrence of several recent crises such as the COVID-19 crisis. It is interesting to apply artificial intelligence models to investigate the crisis impact on the performance of financial distress prediction methods (Sabir et al. 2022).

Author Contributions

Conceptualization, M.H. and S.M.; methodology, M.H.; software, S.M. and M.H.; validation, M.H., S.M. and A.A.; formal analysis, M.H., S.M. and A.A.; investigation, M.H., S.M. and A.A.; resources, M.H., S.M. and A.A.; data curation, M.H., S.M. and A.A.; writing—original draft preparation, M.H.; writing—review and editing, M.H.; visualization, M.H., S.M. and A.A.; supervision, M.H, A.A.; project administration, M.H., S.M. and A.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Addo, Peter Martey, Dominique Guegan, and Bertrand Hassani. 2018. Credit Risk Analysis Using Machine and Deep Learning Models. Risks 6: 38. [Google Scholar] [CrossRef]
  2. Altman, Edward I. 1968. Financial ratios, discriminant analysis, and the prediction of corporate bankruptcy. Journal of Finance 23: 589–609. [Google Scholar] [CrossRef]
  3. Altman, Edward I., Giancarlo Marco, and Franco Varetto. 1994. Corporate distress diagnosis: Comparisons using linear discriminant analysis and neural networks (the Italian experience). Journal of Banking and Finance 18: 505–29. [Google Scholar] [CrossRef]
  4. Anandarajan, Murugan, Picheng Lee, and Asokan Anandarajan. 2004. Bankruptcy Prediction Using Neural Networks. Business Intelligence Techniques 11: 117–32. [Google Scholar]
  5. Aoki, Shigeo, and Yukio Hosonuma. 2004. Bankruptcy Prediction Using Decision Tree. In The Application of Econophysics. Tokyo: Springer, pp. 299–302. [Google Scholar]
  6. Atiya, Amir. 2001. Bankruptcy Prediction for Credit Risk Using Neural Networks: A Survey and New Results. IEEE Transactions on Neural Networks 12: 929–35. [Google Scholar] [CrossRef] [PubMed]
  7. Aydin, Nezir, Nida Sahin, Muhammet Deveci, and Dragan Pamucar. 2022. Prediction of financial distress of companies with artificial neural networks and decision trees models. Machine Learning with Applications 10: 100432. [Google Scholar] [CrossRef]
  8. Barboza, Flavio, Herbert Kimura, and Edward Altman. 2017. Machine learning models and bankruptcy prediction. Expert Systems with Applications 83: 405–17. [Google Scholar] [CrossRef]
  9. Beaver, William H. 1966. Financial ratios as predictors of failure. Journal of Accounting Research 4: 71–111. [Google Scholar] [CrossRef]
  10. Begović, Sanja Vlaović, and Ljiljana Bonić. 2020. Developing a model to predict corporate bankruptcy using decision tree in the Republic of Serbia. Facta Universitatis, Series: Economics and Organization 17: 127–39. [Google Scholar] [CrossRef]
  11. Ben Jabeur, Sami, and Vanessa Serret. 2023. Bankruptcy prediction using fuzzy convolutional neural networks. Research in International Business and Finance 64: 101–844. [Google Scholar] [CrossRef]
  12. Biau, Gérard. 2012. Analysis of a random forests model. The Journal of Machine Learning Research 13: 1063–95. [Google Scholar]
  13. Bragoli, Daniela, Camilla Ferretti, Piero Ganugi, Giovanni Marseguerra, Davide Mezzogori, and Francesco Zammori. 2022. Machine-learning models for bankruptcy prediction: Do industrial variables matter? Spatial Economic Analysis 17: 156–77. [Google Scholar] [CrossRef]
  14. Breiman, Leo. 2001. Random Forests. Machine Learning 45: 5–32. [Google Scholar] [CrossRef]
  15. Chang, Chih-Chung, and Chih-Jen Lin. 2011. LIBSVM: A library for support vector machines. ACM Transactions on Intelligent Systems and Technology 2: 1–27. [Google Scholar] [CrossRef]
  16. Chen, Mu-Yen. 2011. Bankruptcy prediction in firms with statistical and intelligent techniques and a comparison of evolutionary computation approaches. Computers and Mathematics with Applications 62: 4514–24. [Google Scholar] [CrossRef]
  17. Clement, Claudiu. 2020. Machine learning in bankruptcy prediction—A review. Journal of Public Administration, Finance and Law 17: 178–96. [Google Scholar]
  18. Cristianini, Nello, and John Shawe-Taylor. 2000. An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods. Cambridge: Cambridge University Press, pp. 1–189. [Google Scholar]
  19. Davalos, Sergio, Fei Leng, Ehsan Feroz, and Zhiyan Cao. 2014. Designing an if-then rules-based ensemble of heterogeneous bankruptcy classifiers: A genetic algorithm approach. Intelligent Systems in Accounting, Finance and Management 21: 129–53. [Google Scholar] [CrossRef]
  20. Deakin, Edward B. 1972. A discriminant analysis of predictors of business failure. Journal of Accounting Research 10: 167–79. [Google Scholar] [CrossRef]
  21. Dellepiane, Umberto, Michele Di Marcantonio, Enrico Laghi, and Stefania Renzi. 2015. Bankruptcy Prediction Using Support Vector Machines and Feature Selection during the Recent Financial Crisis. International Journal of Economics and Finance 7: 182–95. [Google Scholar] [CrossRef]
  22. Deng, Li, and Dong Yu. 2014. Deep Learning: Methods and Applications. Foundations and Trends in Signal Processing 7: 197–387. [Google Scholar] [CrossRef]
  23. Efron, Bradley. 1975. The efficiency of logistic regression compared to normal discriminant analysis. Journal American Statistical Society 7: 892–98. [Google Scholar] [CrossRef]
  24. Elhoseny, Mohamed, Noura Metawa, Gabro Sztano, and Ibrahim M. El-hasnony. 2022. Deep Learning-Based Model for Financial Distress Prediction. Annals of Operations Research 11: 1–23. [Google Scholar] [CrossRef] [PubMed]
  25. Fisher, Ronald. 1933. The use of multiple measurements in taxonomic problems. Annals of Eugenics 7: 179–88. [Google Scholar] [CrossRef]
  26. Gelfand, Saul B., C. S. Ravishankar, and Edward J. Delp. 1991. An iterative growing and pruning algorithm for classification tree design. IEEE Transaction on Pattern Analysis and Machine Intelligence 13: 163–74. [Google Scholar] [CrossRef]
  27. Gergely, Fejér-Király. 2015. Bankruptcy Prediction: A Survey on Evolution, Critiques, and Solutions. Acta Universitatis Sapientiae, Economics and Business 3: 93–108. [Google Scholar]
  28. Gunn, Steve. 1998. Support Vector Machines for Classification and Regression. Technical Report. Southampton: University of Southampton. [Google Scholar]
  29. Gurnani, Ishika, Febryan Stefanus Tandian, and Maria Susan Anggreainy. 2021. Predicting Company Bankruptcy Using Random Forest Method. Paper presented at IEEE 2nd International Conference on Artificial Intelligence and Data Sciences (AiDAS), IPOH, Malaysia, September 8–9; pp. 1–5. [Google Scholar]
  30. Hamdi, Manel. 2012. Prediction of Financial Distress for Tunisian Firms: A Comparative Study between Financial Analysisand Neuronal Analysis. Business Intelligence Journal 5: 374–82. [Google Scholar]
  31. Hamdi, Manel, and Sami Mestiri. 2014. Bankruptcy Prediction for Tunisian Firms: An Application of Semi-Parametric Logistic Regression and Neural Networks Approach. Economics Bulletin 34: 133–43. [Google Scholar]
  32. Härdle, Wolfgang Karl, Rouslan Moro, and Dorothea Schäfer. 2005. Predicting Bankruptcy with Support Vector Machines. In Statistical Tools for Finance and Insurance. Berlin and Heidelberg: Springer. [Google Scholar]
  33. Hearst, Marti A., Susan T. Dumais, Edgar Osuna, John Platt, and Bernhard Scholkopf. 1998. Support vector machines. IEEE Intelligent System 13: 18–28. [Google Scholar] [CrossRef]
  34. Heo, Junyoung, and Jin Yong Yang. 2014. AdaBoost based bankruptcy forecasting of Korean construction companies. Applied Soft Computing 24: 494–99. [Google Scholar] [CrossRef]
  35. Hosaka, Tadaaki. 2019. Bankruptcy prediction using imaged financial ratios and convolutional neural networks. Expert Systems with Applications 117: 287–99. [Google Scholar] [CrossRef]
  36. Joshi, Shreya, Rachana Ramesh, and Shagufta Tahsildar. 2018. A Bankruptcy Prediction Model Using Random Forest. Paper presented at IEEE Second International Conference on Intelligent Computing and Control Systems (ICICCS), Madurai, India, June 14–15. [Google Scholar]
  37. Kamruzzaman, M. M., and Omar Alruwaili. 2022. AI-based computer vision using deep learning in 6G wireless networks. Computers and Electrical Engineering 102: 108233. [Google Scholar] [CrossRef]
  38. Kim, Myoung-Jong, and Ingoo Han. 2003. The Discovery of Experts’ Decision Rules from Qualitative Bankruptcy Data Using Genetic Algorithms. Expert Systems with Application 25: 637–46. [Google Scholar] [CrossRef]
  39. Kim, Phil. 2017. MATLAB Deep Learning: With Machine Learning, Neural Networks and Artificial Intelligence, 1st ed. New York: A Press Book. 151p. [Google Scholar]
  40. Kuizinienė, Dovilė, Tomas Krilavičius, Robertas Damaševičius, and Rytis Maskeliūnas. 2022. Systematic Review of Financial Distress Identification using Artificial Intelligence Methods. Applied Artificial Intelligence 36: 2138124. [Google Scholar] [CrossRef]
  41. Kumar, Ravi, and Vadlamani Ravi. 2007. Bankruptcy prediction in banks and firms via statistical and intelligent techniques—A review. European Journal of Operational Research 180: 1–28. [Google Scholar] [CrossRef]
  42. LeCun, Yann, Yoshua Bengio, and Geoffrey Hinton. 2015. Deep learning. Nature 521: 436–44. [Google Scholar] [CrossRef] [PubMed]
  43. Mai, Feng, Shaonan Tian, Chihoon Lee, and Ling Ma. 2019. Deep learning models for bankruptcy prediction using textual disclosures. European Journal of Operational Research 274: 743–58. [Google Scholar] [CrossRef]
  44. Martono, Niken Prasasti, and Hayato Ohwada. 2023. Financial Distress Model Prediction Using Machine Learning: A Case Study on Indonesia’s Consumers Cyclical Companies. In Machine Learning and Principles and Practice of Knowledge Discovery in Databases. Cham: Springer, vol. 1753. [Google Scholar]
  45. Matin, Rastin, Casper Hansen, Christian Hansen, and Pia Mølgaard. 2019. Predicting distresses using deep learning of text segments in annual reports. Expert Systems with Applications 132: 199–208. [Google Scholar] [CrossRef]
  46. Máté, Domicián, Hassan Raza, and Ishtiaq Ahmad. 2023. Comparative Analysis of Machine Learning Models for Bankruptcy Prediction in the Context of Pakistani Companies. Risks 11: 176. [Google Scholar] [CrossRef]
  47. Mestiri, Sami, and Manel Hamdi. 2013. Credit Risk Prediction: A Comparative Study between Logistic Regression and Logistic Regression with Random Effects. International Journal of Management Science and Engineering Management 7: 200–4. [Google Scholar] [CrossRef]
  48. Narvekar, Aditya, and Debashis Guha. 2021. Bankruptcy prediction using machine learning and an application to the case of the COVID-19 recession. Data Science in Finance and Economics 1: 180–95. [Google Scholar] [CrossRef]
  49. Noh, Seol-Hyun. 2023. Comparing the Performance of Corporate Bankruptcy Prediction Models Based on Imbalanced Financial Data. Sustainability 15: 4794. [Google Scholar] [CrossRef]
  50. Noviantoro, T., and J. P. Huang. 2021. Comparing machine learning algorithms to investigate company financial distress. Review of Business, Accounting & Finance 1: 454–79. [Google Scholar]
  51. Odom, Marcus D., and Ramesh Sharda. 1990. A neural network model for bankruptcy prediction. Paper presented at the International Joint Conference on Neural Networks, San Diego, CA, USA, June 17–21; Alamitos: IEEE Press, vol. 2, pp. 163–68. [Google Scholar]
  52. Ohlson, James A. 1980. Financial Ratios and the Probabilistic Prediction of Bankruptcy. Journal of Accounting Research 18: 109–31. [Google Scholar] [CrossRef]
  53. Pang, Su-Juan. 2006. Application of Logistic Regression Model in Credit Risk Analysis. Mathematics in Practice and Theory 9: 129–37. [Google Scholar]
  54. Park, Min Sue, Hwijae Son, Chongseok Hyun, and Hyung Ju Hwang. 2021. Explainability of Machine Learning Models for Bankruptcy Prediction. IEEE Access 9: 124887–99. [Google Scholar] [CrossRef]
  55. Pepe, Margaret Sullivan. 2000. Receiver operating characteristic methodology. Journal of the American Statistical Association 95: 308–11. [Google Scholar] [CrossRef]
  56. Ptak-Chmielewska, Aneta, and Anna Matuszyk. 2020. Application of the random survival forests method in the bankruptcy prediction for small and medium enterprises. Argumenta Oeconomica 1: 127–42. [Google Scholar] [CrossRef]
  57. Qu, Yi, Pei Quan, Minglong Lei, and Yong Shi. 2019. Review of bankruptcy prediction using machine learning and deep learning techniques. Procedia Computer Science 162: 895–9. [Google Scholar] [CrossRef]
  58. Quinlan, J. Ross. 1986. Induction of decision trees. Machine Learning 1: 81–106. [Google Scholar] [CrossRef]
  59. Roy, Tanmoy, Marwala Tshilidzi, and Snehashish Chakraverty. 2021. Chapter 12—Speech emotion recognition using deep learning. In New Paradigms in Computational Modeling and Its Applications. Cambridge: Academic Press, pp. 177–87. [Google Scholar]
  60. Sabir, Zulqurnain, Muhammad Asif Zahoor Raja, Sharifah E. Alhazmi, Manoj Gupta, Adnène Arbi, and Isa Abdullahi Baba. 2022. Applications of artificial neural network to solve the nonlinear COVID-19 mathematical model based on the dynamics of SIQ. Journal of Taibah University for Science 16: 874–84. [Google Scholar] [CrossRef]
  61. Schmidhuber, Jürgen. 2015. Deep learning in neural networks: An overview. Neural Networks 61: 85–117. [Google Scholar] [CrossRef] [PubMed]
  62. Shetty, Shekar, Mohamed Musa, and Xavier Brédart. 2022. Bankruptcy Prediction Using Machine Learning Techniques. Journal of Risk and Financial Management 15: 35. [Google Scholar] [CrossRef]
  63. Shin, Kyung-Shik, Taik Soo Lee, and Hyun-jung Kim. 2005. An Application of Support Vector Machines in Bankruptcy Prediction Model. Expert Systems and Applications 28: 127–35. [Google Scholar] [CrossRef]
  64. Shin, Kyung-Shik, and Yong-Joo Lee. 2002. A genetic algorithm application in bankruptcy prediction modeling. Expert Systems with Applications 23: 321–28. [Google Scholar] [CrossRef]
  65. Suganyadevi, S., V. Seethalakshmi, and Krishnasamy Balasamy. 2022. A review on deep learning in medical image analysis. International Journal of Multimedia Information Retrieval 11: 19–38. [Google Scholar] [CrossRef]
  66. Vapnik, Vladimir Naumovich, and Vlamimir Vapnik. 1998. The Nature of Statistical Learning Theory. New York: Springer. [Google Scholar]
  67. Vuk, Miha, and Tomaz Curk. 2006. Roc curve, lift chart and calibration plot. Organization Science 3: 89–108. [Google Scholar] [CrossRef]
  68. Walczak, Steven, and Narciso Cerpa. 2003. Artificial Neural Networks. In Encyclopedia of Physical Science and Technology, 3rd ed. Cambridge: Academic Press. [Google Scholar]
  69. Wilson, Rick L., and Ramesh Sharda. 1994. Bankruptcy Prediction Using Neural Networks. Decision Support Systems 11: 545–57. [Google Scholar] [CrossRef]
  70. Xie, Ying, Linh Le, Yiyun Zhou, and Vijay V. Raghavan. 2018. Chapter 10—Deep Learning for Natural Language Processing. In Handbook of Statistics. Amsterdam: Elsevier, vol. 38, pp. 317–28. [Google Scholar]
  71. Zibanezhad, Elahe, Daryush Foroghi, and Amirhassan Monadjemi. 2011. Applying decision tree to predict bankruptcy. Paper presented at IEEE International Conference on Computer Science and Automation Engineering, Shanghai, China, June 10–12; pp. 165–69. [Google Scholar]
  72. Zou, Jinming, Yi Han, and Sung-Sau So. 2008. Overview of Artificial Neural Networks. In Artificial Neural Networks. Totowa: Humana Press, vol. 458, pp. 14–22. [Google Scholar]
  73. Zweig, Mark, and Gregory Campbell. 1993. Receiver-operating characteristic (ROC) plots: A fundamental evaluation tool in clinical medicine. Clinical Chemistry 39: 561–77. [Google Scholar] [CrossRef]
Figure 1. The Standard architecture of DNN.
Figure 1. The Standard architecture of DNN.
Jrfm 17 00132 g001
Figure 2. ROC curve for the five machine learning models and DNN.
Figure 2. ROC curve for the five machine learning models and DNN.
Jrfm 17 00132 g002
Table 1. A summary of literature review on bankruptcy prediction using deep learning.
Table 1. A summary of literature review on bankruptcy prediction using deep learning.
Author(s)Model(s)Type of Input VariablesSampling PeriodPerformance
Criteria Used
Conclusion(s)
Addo et al. (2018)
-
Logistic regression (LR)
-
Random forest (RF)
-
Gradient boosting
-
Four architectures of deep neural networks
10 financial variables2016–2017
-
AUC
-
RMSE
-
The class of tree-based algorithms (RF and gradient boosting model) outperforms other applied techniques.
-
The gradient boosting model proved high performance compared to RF.
Hosaka (2019)
-
Convolutional neural network (CNN)
-
Classification and regression trees (CART)
-
Linear discriminant analysis (LDA)
-
Support vector machine (SVM)
-
Multi-layer perceptrons (MLP)
-
AdaBoost
-
Altman’s Z”-score
133 financial items2002–2016
-
Identification rate
-
AUC
-
Deeper network design significantly improves predictive accuracy.
-
CNN based on GoogLeNet outperforms traditional and conventional tools.
Noviantoro and Huang (2021)
-
Decision tree
-
Random forest
-
K-nearest neighbour algorithm
-
Support vector machine
-
Artificial neural network
-
Naïve bayes
-
Logistic regression
-
Rule induction
-
Deep neural network
96 financial indicators1999–2009
-
Accuracy rate
-
F score
-
AUC
-
RF demonstrated high accuracy compared to the other applied models followed by the deep learning algorithm in the second rank.
Shetty et al. (2022)
-
Deep neural network
-
Support vector machine (SVM)
-
Extreme gradient boosted tree method (XGBoost)
Three financial ratios: the return on assets, the current ratio, and the solvency ratio2002–2012
-
Precision (%)
-
Recall (%)
-
F1 score
-
A similar level of prediction accuracy of 82–83% was achieved by using the proposed methods.
Elhoseny et al. (2022)
-
Adaptive whale optimization algorithm with deep learning (AWOA-DL)
-
Logistic regression
-
RBF Network
-
Teaching-learning-based optimization-DL (TLBO-DL)
-
Deep neural network
179 financial attributes2000–2013
-
Accuracy rate
-
F score
-
Kappa measure
-
Better predictions were obtained using AWOA-DL compared to the other models.
Ben Jabeur and Serret (2023)
-
Fuzzy convolutional neural networks (FCNN)
-
Discriminant analysis (DA)
-
Logistic regression (LR)
-
Support vector machine (SVM)
-
Partial least squares discriminant analysis (PLSDA)
-
Multi-layer perceptron (MLP)
17 financial variables2014–2017
-
ACC: Accuracy
-
AUC: Area under the ROC curve
-
GM: Geometric Mean
-
YI: Youden’s Index;
-
MCC: Matthews Correlation Coefficient
-
Sensitivity
-
Specificity
-
F score
-
FCNN outperforms the other adopted techniques
Noh (2023)
-
Long short-term memory (LSTM)
-
Logistic regression (LR)
-
K-Nearest Neighbour (k-NN)
-
Decision tree (DT)
-
Random forest (RF)
13 financial variables2012–2021
-
Accuracy
-
Precision
-
Recall
-
F1 score
-
AUC
-
The proposed method can enhance the prediction accuracy and therefore was selected as the appropriate model for bankruptcy prediction.
Table 2. The series of financial ratios.
Table 2. The series of financial ratios.
R 1 Duration credit to the customer R 14 Permanent capital turnover
R 2 Gross margin rate R 15 Return on permanent capital
R 3 Operating margin rate R 16 Rate of long-term debt
R 4 Ratio of personnel expenses R 17 Ratio of financial independence
R 5 Net margin rate R 18 Total debt ratio
R 6 Asset turnover R 19 Immobilisation coverage by equity capital
R 7 Equity turnover R 20 The long- and medium-term debt capacity
R 8 Economic profitability R 21 Ratio of financial expenses
R 9 rate of return on assets R 22 Financial expenses/total debt
R 10 Operating profitability of total assets R 23 Working capital ratio
R 11 Gross economic profitability R 24 Relative liquidity ratio
R 12 Net economic profitability R 25 Quick ratio
R 13 Rate of return on equity
Table 3. Confusion matrix.
Table 3. Confusion matrix.
Predicted class “0”Predicted class “1”
Real class“0”True positive (T0)False positive (F1)
Real class“1”False negative (F0)True negative (T1)
Table 4. Prediction results and models accuracy.
Table 4. Prediction results and models accuracy.
ModelsAccuracy RateF1-ScoreAUCRank
Linear Discriminant Analysis (LDA)80.9%0.8900.5745
Logistic Regression (LR)85.8%0.9220.6333
Decision Trees (DT)74.3%0.8380.6756
Random Forest (RF)88.2%0.9330.8152
Support Vector Machine (SVM)84.8%0.9100.5634
Deep Neural Network (DNN)93.6%0.9640.8881
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hamdi, M.; Mestiri, S.; Arbi, A. Artificial Intelligence Techniques for Bankruptcy Prediction of Tunisian Companies: An Application of Machine Learning and Deep Learning-Based Models. J. Risk Financial Manag. 2024, 17, 132. https://doi.org/10.3390/jrfm17040132

AMA Style

Hamdi M, Mestiri S, Arbi A. Artificial Intelligence Techniques for Bankruptcy Prediction of Tunisian Companies: An Application of Machine Learning and Deep Learning-Based Models. Journal of Risk and Financial Management. 2024; 17(4):132. https://doi.org/10.3390/jrfm17040132

Chicago/Turabian Style

Hamdi, Manel, Sami Mestiri, and Adnène Arbi. 2024. "Artificial Intelligence Techniques for Bankruptcy Prediction of Tunisian Companies: An Application of Machine Learning and Deep Learning-Based Models" Journal of Risk and Financial Management 17, no. 4: 132. https://doi.org/10.3390/jrfm17040132

Article Metrics

Back to TopTop