Bankruptcy is a legal status of a person or other organization that cannot repay their debts to
creditors. Bankruptcy prediction is the task of predicting bankruptcy and by doing various surveys we can avoid
financial distress of firms. It is a huge area of accounting and finance research. The significance of this area is
an important part of financial specialists and creditors in assessing the probability that a firm may go bankrupt
or not. Estimating the risk of corporate bankruptcies is very important as the effect of bankruptcy is on a global
level. The aim of predicting financial distress is to develop a predictive model that combines various economic
factors which allow foreseeing the financial status of a firm. In this domain, various methods were proposed that
were based on neural networks, Support Vector Machines, Decision Trees, Random Forests, Naïve Bayes,
Balanced Bagging and Logistic Regression. In this paper, we document our observations as we explore and build
a Restricted Boltzmann Machine to Bankruptcy Prediction. We started by carrying out data pre-processing where
we impute the missing data values using Mean Imputation. To solve the data imbalance issue, we apply the
Synthetic Minority Oversampling Technique (SMOTE) to oversample the minority class labels. Finally, we
analyze and evaluate the performance of the model.
Ismail+Reid2016 - Ask The Experts - The Actuary (June 2016)
1) Structured expert judgment (SEJ) is an approach to combine multiple expert opinions in an objective way by weighting each expert based on their performance on seed questions.
2) In an experiment, experts provided judgments on past and future political violence. Their responses to seed questions revealed variability in expertise, with some experts more calibrated and informative.
3) Weighting the experts by their performance on the seed questions and combining their judgments resulted in a consensus estimate that was tighter and potentially more accurate than relying on any single expert or an equal-weighted combination.
A Review on Credit Card Default Modelling using Data Science
In the last few years, credit card issuers have become one of the major consumer lending products in the U.S. as well as several other developed nations of the world, representing roughly 30 of total consumer lending USD 3.6 tn in 2016 . Credit cards issued by banks hold the majority of the market share with approximately 70 of the total outstanding balance. Bank’s credit card charge offs have stabilized after the financial crisis to around 3 of the outstanding total balance. However, there are still differences in the credit card charge off levels between different competitors. Harsh Nautiyal | Ayush Jyala | Dishank Bhandari "A Review on Credit Card Default Modelling using Data Science" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Special Issue | International Conference on Advances in Engineering, Science and Technology - 2021 , May 2021, URL: https://www.ijtsrd.com/papers/ijtsrd42461.pdf Paper URL : https://www.ijtsrd.com/engineering/computer-engineering/42461/a-review-on-credit-card-default-modelling-using-data-science/harsh-nautiyal
This document discusses using a multi-objective evolutionary algorithm (MOEA) for feature selection in bankruptcy prediction models. The goal is to maximize classifier accuracy while minimizing the number of features. A two-objective problem of minimizing features and maximizing accuracy is analyzed using logistic regression and support vector machines classifiers. The methodology is tested on financial data from 1200 French companies and shown to be an efficient feature selection approach, obtaining best results when optimizing both accuracy and classifier parameters simultaneously.
This research aims to identify and analyze the effect of Capital Adequacy Ratio (CAR), Operation Expense (BOPO), Net Interest Margin (NIM), and Non Performing Loan (NPL) of the Loan to Deposit Ratio (LDR) of conventional bank on the Indonesia Stock Exchange period 2012 – 2017, either simultaneously or partially. Independent variables used in this study is CAR, BOPO, NIM and NPL, while LDR as the dependent variable.The population in this research is conventional bank listed on the Indonesia Stock Exchange. The sampling technique in this research is purposive sampling. The number of samples in accordance with the prescribed criteria are as many as 35 samples. Based on the result of the research found that the variable CAR influences negatively insignificantly toward LDR, BOPO and NIM influences positively insignificantly toward LDR, while the variable NPL influences positively significantly toward CAR. But simultaneously CAR, BOPO, NIM, and NPL jointly affect the LDR.
Bankruptcy Prediction is an art of predicting bankruptcy and various measures of financial
distress of public or private firms. In recent past days we are seeing many cases with distress
and bankrupted. It is a huge area of finance and accounting research. The importance of the
world is due partially to the relevance for creditors and investors in evaluating the likelihood
that a firm may go bankrupt. The quantity of research is additionally a function of the supply of
data: for public firms which went bankrupt or not, numerous accounting ratios which
may indicate danger can be calculated, and various other potential explanatory variables also
are available. Consequently, the world is well-suited for testing of increasingly sophisticated,
data-intensive forecasting approaches.
MOVIE SUCCESS PREDICTION AND PERFORMANCE COMPARISON USING VARIOUS STATISTICAL...ijaia
Movies are among the most prominent contributors to the global entertainment industry today, and they
are among the biggest revenue-generating industries from a commercial standpoint. It's vital to divide
films into two categories: successful and unsuccessful. To categorize the movies in this research, a variety
of models were utilized, including regression models such as Simple Linear, Multiple Linear, and Logistic
Regression, clustering techniques such as SVM and K-Means, Time Series Analysis, and an Artificial
Neural Network. The models stated above were compared on a variety of factors, including their accuracy
on the training and validation datasets as well as the testing dataset, the availability of new movie
characteristics, and a variety of other statistical metrics. During the course of this study, it was discovered
that certain characteristics have a greater impact on the likelihood of a film's success than others. For
example, the existence of the genre action may have a significant impact on the forecasts, although another
genre, such as sport, may not. The testing dataset for the models and classifiers has been taken from the
IMDb website for the year 2020. The Artificial Neural Network, with an accuracy of 86 percent, is the best
performing model of all the models discussed.
Ismail+Reid2016 - Ask The Experts - The Actuary (June 2016)Raveem Ismail
1) Structured expert judgment (SEJ) is an approach to combine multiple expert opinions in an objective way by weighting each expert based on their performance on seed questions.
2) In an experiment, experts provided judgments on past and future political violence. Their responses to seed questions revealed variability in expertise, with some experts more calibrated and informative.
3) Weighting the experts by their performance on the seed questions and combining their judgments resulted in a consensus estimate that was tighter and potentially more accurate than relying on any single expert or an equal-weighted combination.
A Review on Credit Card Default Modelling using Data ScienceYogeshIJTSRD
In the last few years, credit card issuers have become one of the major consumer lending products in the U.S. as well as several other developed nations of the world, representing roughly 30 of total consumer lending USD 3.6 tn in 2016 . Credit cards issued by banks hold the majority of the market share with approximately 70 of the total outstanding balance. Bank’s credit card charge offs have stabilized after the financial crisis to around 3 of the outstanding total balance. However, there are still differences in the credit card charge off levels between different competitors. Harsh Nautiyal | Ayush Jyala | Dishank Bhandari "A Review on Credit Card Default Modelling using Data Science" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Special Issue | International Conference on Advances in Engineering, Science and Technology - 2021 , May 2021, URL: https://www.ijtsrd.com/papers/ijtsrd42461.pdf Paper URL : https://www.ijtsrd.com/engineering/computer-engineering/42461/a-review-on-credit-card-default-modelling-using-data-science/harsh-nautiyal
This document discusses using a multi-objective evolutionary algorithm (MOEA) for feature selection in bankruptcy prediction models. The goal is to maximize classifier accuracy while minimizing the number of features. A two-objective problem of minimizing features and maximizing accuracy is analyzed using logistic regression and support vector machines classifiers. The methodology is tested on financial data from 1200 French companies and shown to be an efficient feature selection approach, obtaining best results when optimizing both accuracy and classifier parameters simultaneously.
This document summarizes research that combines statistical and machine learning methods to predict corporate failure using financial data. The researchers empirically compare discriminant analysis, logistic regression, classification trees, rule induction, and Bayesian networks on data from 120 Spanish companies, 60 that went bankrupt and 60 that did not. They also implement voting and Bayesian techniques to combine the individual models, finding improved predictive performance over single models. The key predictor variables are financial ratios gathered from company accounts over the three years before failure or survey date.
Predicting Credit Card Defaults using Machine Learning AlgorithmsSagar Tupkar
This is a project that I worked on as a Capstone for my Masters in Business Analytics program at the University of Cincinnati. In this project, I have performed an end-to-end data mining exercise including data cleaning, distribution analysis, exploratory data analysis, model building etc. to identify and predict Credit Card defaults using Customer's data on past payments and general profile. In the process for building Machine Learning models, I have fit and compared the performance of multiple models and algorithms like Logistic Regreesion, PCA, Classification tree, AdaBoost Classifier, ANN and LDA.
This document discusses using artificial neural networks (ANNs) to enhance stock picking and investment strategies by incorporating earnings forecasts from financial analysts. It aims to compare different ANN models and identify the best model for forecasting stock prices and improving investment profitability. The study uses quarterly data on stock prices, indexes, analyst earnings forecasts and recommendations from 1997-2003 to train and evaluate ANN models. It finds that ANN strategies based on analyst forecasts achieved higher returns than other investment strategies over this period.
IMPROVED TURNOVER PREDICTION OF SHARES USING HYBRID FEATURE SELECTIONIJDKP
Predicting the total turnover of a company in the most unstable stock market and trade conditions has
always proved to be a costly affair causing rise and fall of several trades. Data mining is a well-known
sphere of Computer Science that aims at extracting meaningful information from large databases. However,
despite the existence of many algorithms for the purpose of predicting future trends, their efficiency is
questionable as their predictions suffer from a high error rate. The objective of this paper is to investigate
and rate the performance of classifiers based on the features selected by Hybrid Feature Selection. The
authorized dataset for predicting the turnover was taken fromwww.bsc.com and included the stock market
values of various companies over the past 10 years. The algorithms were investigated using the Weka tool.
The Hybrid feature selection (HFS) algorithm, was run on this dataset to extract the important and
influential features for classification. With these extracted features, the Total Turnover of the company was
predicted using various algorithms like Random Forest, Decision Tree, SVM and Multinomial Regression.
This prediction mechanism was implemented to predict the turnover of a company on an everyday basis
and hence could help navigate through dubious stock markets trades. An accuracy rate of was achieved by
the above prediction process. Moreover, the importance of the stock market attributes through Incremental
Feature Selection (IFS) was established as well.
This document proposes a decision support system to help a production company determine the best country for a new investment abroad. It involves 4 phases: 1) determining criteria for evaluating candidate countries, 2) using AHP to assign weights to the criteria, 3) evaluating 50 candidate countries based on the criteria, selecting the top 10, and 4) further evaluating the top 10 using another AHP model to select the top 5 investment alternatives. The system aims to provide a comprehensive, scientific approach for solving the investment location problem.
The document evaluates the effectiveness of a behavior-based safety initiative (BBSI) card system introduced at a cement manufacturing company in Zimbabwe. The following key points are made:
- Accident and injury rates significantly decreased after the card system was introduced based on a comparison of rates before and after as well as a paired t-test.
- A negative correlation was found between the number of cards issued and the number of accidents/injuries, indicating more cards issued correlated with fewer accidents/injuries.
- Employee attitudes toward safety practices were found to be positively influenced by the card system based on a questionnaire administered to employees.
The card system was effective in reducing accidents and injuries at the company according to
DEEP-LEARNING-BASED HUMAN INTENTION PREDICTION WITH DATA AUGMENTATIONijaia
Data augmentation has been broadly applied in training deep-learning models to increase the diversity of
data. This study ingestigates the effectiveness of different data augmentation methods for deep-learningbased human intention prediction when only limited training data is available. A human participant pitches
a ball to nine potential targets in our experiment. We expect to predict which target the participant pitches
the ball to. Firstly, the effectiveness of 10 data augmentation groups is evaluated on a single-participant
data set using RGB images. Secondly, the best data augmentation method (i.e., random cropping) on the
single-participant data set is further evaluated on a multi-participant data set to assess its generalization
ability. Finally, the effectiveness of random cropping on fusion data of RGB images and optical flow is
evaluated on both single- and multi-participant data sets. Experiment results show that: 1) Data
augmentation methods that crop or deform images can improve the prediction performance; 2) Random
cropping can be generalized to the multi-participant data set (prediction accuracy is improved from 50%
to 57.4%); and 3) Random cropping with fusion data of RGB images and optical flow can further improve
the prediction accuracy from 57.4% to 63.9% on the multi-participant data set.
Tarmo Puolokainen: Public Agencies’ Performance Benchmarking in the Case of D...Eesti Pank
This PhD thesis examines measuring the performance of public agencies in the face of demand uncertainty. It introduces the concept of minimum service level to account for demand uncertainty in efficiency analyses of public agencies. The thesis develops theoretical frameworks and applies stochastic frontier analysis and data envelopment analysis methods to empirically analyze the cost efficiency and potential under-resourcing of Estonian, Finnish, and Swedish fire and rescue services at different administrative levels over multiple time periods. The analyses provide insights into how the different countries' fire and rescue services could potentially improve performance and save costs by better allocating resources across subunits.
REVIEWING PROCESS MINING APPLICATIONS AND TECHNIQUES IN EDUCATIONijaia
Process Mining (PM) emerged from business process management but has recently been applied to
educational data and has been found to facilitate the understanding of the educational process.
Educational Process Mining (EPM) bridges the gap between process analysis and data analysis, based on
the techniques of model discovery, conformance checking and extension of existing process models. We
present a systematic review of the recent and current status of research in the EPM domain, focusing on
application domains, techniques, tools and models, to highlight the use of EPM in comprehending and
improving educational processes.
Ofccp enforcement trends 03 21_13_webinar deckJamie Janvier
This PowerPoint presentation provides information on employment, labor, and immigration law related to OFCCP compliance. It discusses OFCCP's increased focus on pay equity under the Obama administration, including new directives on compensation analysis and increased scrutiny of factors impacting pay. It advises employers to conduct privileged self-analyses of their compensation practices, validate processes impacting pay, and improve documentation of outreach efforts to prepare for more intensive audits.
Managing collaboration within networks and relationships in the serious game ...ijcsit
This research develop the managing within network and relationship mechanism in agribusiness
management through serious game. Agribusiness is represented as sand that work together in the market
(sandpile) to maintain networks and relationships. This research apply agent base model for predicting
activity network based on the parameters that exist in the collaboration. The result indicate that average
selling, average buying and market price (CK = 4) are not approach the value of the open market but
precisely coincide with eachother. Total bought and total sold are tend to be high value. This condition
suggests a very tight competition. The average selling, average buying and market price (CK = 0.01) are
approach the value of the open market. Total bought and total sold are not as high as total bought and total
sold, by using CK = 4, this condition shows the competition is not too tight.
This document describes building models to predict credit card default payments. It retrieves credit card data from a public dataset containing details on customers' personal information, credit limits, payment histories and default statuses. The data is explored through visualizations to identify relationships between variables. Two classification models are built using KNN and decision tree algorithms. The decision tree model achieves a higher accuracy of 80% compared to KNN's 74% accuracy, indicating decision trees are more suitable for predicting default payments from this credit card data.
Decision support systems, Supplier selection, Information systems, Boolean al...ijmpict
For organizations operating with number of products/services and number of suppliers, to select the right
supplier meeting all their requirements will be a challenging job. Such organizations need a good
decision support system to evaluate the suppliers effectively. Several decision support systems have been
reported to deal with complex selection process to decide the right supplier. Many mathematical models
have also been developed. This paper presents a new method, named as Bit Decision Making (BDM)
method, which treats such complex system of decision making as a collection and sequence of reasonable
number of meaningful and manageable sub-systems by identifying and processing the relevant decision
criteria in each sub-system. Help of Boolean logic and Boolean algebra is taken to assign binary digit
values to the selection criteria and generate mathematical equations that correlate the inputs to the output
at each stage of decision making. Each sub-system with its own mathematical model has been treated as a
standardized decision sub-system for that phase of making decision in evaluating suppliers. The sequence
and connectivity of the sub-systems along with their outputs finally lead to selection of the best supplier. A
real-world case of evaluation of information technology (IT) tenders has been dealt with for application of
the proposed method. The paper discusses in detail the theory, methodology, application and features of
the new method.
Impact and Implications of Operations Research in Stock Marketinventionjournals
The motivation of this article is to advocate the administrative routine of settling on choices construct in light of instinct, as well as instinct combined with quantitative investigation. Operations Research (OR) is one of the main administrative choice science instruments utilized by benefit and charitable, for example, stock market. Gauging stock return is an important financial subject that has attracted researchers' consideration for a long time. It includes a supposition that basic data openly accessible in the past has some prescient connections to the future stock returns. This review tries to help the financial specialists in the stock market to choose the better planning for purchasing or offering stocks based on the information extricated from the chronicled costs of such stocks. The choice taken will be founded on choice tree classifier which is one of the Operations Research techniques.
COMPARISON OF BANKRUPTCY PREDICTION MODELS WITH PUBLIC RECORDS AND FIRMOGRAPHICScscpconf
Many business operations and strategies rely on bankruptcy prediction. In this paper, we aim to
study the impacts of public records and firmographics and predict the bankruptcy in a 12-
month-ahead period with using different classification models and adding values to traditionally
used financial ratios. Univariate analysis shows the statistical association and significance of
public records and firmographics indicators with the bankruptcy. Further, seven statistical
models and machine learning methods were developed, including Logistic Regression, Decision
Tree, Random Forest, Gradient Boosting, Support Vector Machine, Bayesian Network, and
Neural Network. The performance of models were evaluated and compared based on
classification accuracy, Type I error, Type II error, and ROC curves on the hold-out dataset.
Moreover, an experiment was set up to show the importance of oversampling for rare event
prediction. The result also shows that Bayesian Network is comparatively more robust than
other models without oversampling.
This document presents a system for predicting corporate bankruptcy using textual disclosures from SEC filings. It discusses how previous studies have used financial ratios and market data to predict bankruptcy, but that textual disclosures also provide important unstructured qualitative information. The proposed system uses natural language processing and machine learning algorithms to extract features from 10-K and 10-Q filings and predict bankruptcy with high accuracy, even before the final bankruptcy occurs. It aims to improve on previous bankruptcy prediction methods by incorporating both financial and textual data sources.
Prediction of Corporate Bankruptcy using Machine Learning Techniques Shantanu Deshpande
Aim is to build a classification model to predict whether company will become bankrupt or not using financial ratios of Polish companies. Applied various machine learning models like Random Forest, KNN, AdaBoost & Decision Tree with pre-processing techniques like SMOTE-ENN (to deal with class imbalance) & feature selection (for identifying ) and trained on Polish Bankruptcy dataset with prediction accuracy of 89%.
Corporate bankruptcy prediction using Deep learning techniquesShantanu Deshpande
This document proposes using deep learning techniques like LSTM neural networks to predict corporate bankruptcy by integrating both financial ratio data and textual disclosures from annual reports. It notes that previous studies have largely relied on statistical models or used only financial data with machine learning. The researcher aims to determine if adding textual data to an LSTM model improves prediction performance over a CNN model using only financial ratios. The document outlines the research question, objectives, and provides an overview of previous bankruptcy prediction studies using statistical, machine learning and deep learning methods.
IRJET- A Comparative Study to Detect Fraud Financial Statement using Data Min...IRJET Journal
This document discusses using data mining and machine learning algorithms to detect fraud in financial statements. It reviews several studies that have applied techniques like neural networks, decision trees, support vector machines, and logistic regression to financial statement fraud detection tasks. Accuracy rates ranging from 45.6% to 85.6% are reported for these techniques. The document also outlines common types of financial statement fraud like manipulation of financial records and intentional omission of information. Overall, the document analyzes the effectiveness of various classification algorithms for detecting financial statement fraud based on past research.
This document discusses the application of meta-learning algorithms in banking sector data mining for fraud detection. It proposes using Classification and Regression Tree (CART), AdaBoost, LogitBoost, Bagging and Dagging algorithms for classification of banking transaction data. The experimental results show that Bagging algorithm has the best performance with the lowest misclassification rate, making it effective for banking fraud detection through data mining. Data mining can help banks detect patterns for applications like credit scoring, payment default prediction, fraud detection and risk management by analyzing customer transaction history and loan details.
A predictive system for detection of bankruptcy using machine learning techni...IJDKP
Bankruptcy is a legal procedure that claims a person or organization as a debtor. It is essential to
ascertain the risk of bankruptcy at initial stages to prevent financial losses. In this perspective, different
soft computing techniques can be employed to ascertain bankruptcy. This study proposes a bankruptcy
prediction system to categorize the companies based on extent of risk. The prediction system acts as a
decision support tool for detection of bankruptcy
INSOLVENCY PREDICTION ANALYSIS OF ITALIAN SMALL FIRMS BY DEEP LEARNINGIJDKP
To improve credit risk management, there is a lot of interest in bankruptcy predictive models. Academic
research has mainly used traditional statistical techniques, but interest in the capability of machine
learning methods is growing. This Italian case study pursues the goal of developing a commercial firms
insolvency prediction model. In compliance with the Basel II Accords, the major objective of the model is
an estimation of the probability of default over a given time horizon, typically one year.
The document describes a new credit risk modeling technique using a Bayesian network with a latent variable. It introduces a discrete Bayesian network model containing a latent variable that represents different classes of probability distributions for credit risk. The model allows evaluating credit risk and clustering loan subscribers. The document then provides details of the Bayesian network model and proposes a customized Expectation Maximization algorithm to learn the model parameters from data. The model and learning approach are applied to a real loan data set to classify loans and analyze credit risk profiles.
This document discusses the relevance and implications of forecasting retail deposits. Forecasting retail deposits involves analyzing macroeconomic data to build models that can accurately predict future deposit levels given economic conditions. Accurately forecasting deposits is important for banks to inform strategic planning and decisions around operations, technology, and infrastructure needs. The implications of deposit forecasting are discussed from social and philosophical perspectives, including how forecasting stems from humans' innate desire to understand and prepare for an uncertain future.
Machine learning algorithms can be used in various areas of banking and central banking. Specifically, this document discusses:
1) Using machine learning for traditional credit risk modeling to forecast probability of default and assess financial stability.
2) Applying machine learning to time series forecasting of macroeconomic variables like inflation for monetary policy purposes.
3) Performing text mining on central bank research documents and news articles to measure economic uncertainty and risk in financial markets.
Predicting Corporate Failure - An Application of Discriminate Analysisscmsnoida5
Corporate failure is a serious problem being
confronted by the corporate world. This issue
has been a subject of intensive research and
discussion by economists, bankers, creditors,
equity shareholders, accountants, marketing
and management experts. The present study
aims at developing a model for prediction
of corporate failure on the basis of financial
ratios. The study is based on the data of
selected firms from chemical industry (with
equal number of failed and non failed firms).
The discriminant analysis has been used to
discriminate between failed and non failed
firms. It is concluded that some of the
financial ratios can significantly differentiate
between failed and non failed firms. The
finding will be useful for the banks and other
financial institutions in designing a suitable
credit appraisal and monitoring system for their
loans. This model could guide the policy makers
to prepare an early warning system to avoid
bankruptcy.
Credit risk assessment with imbalanced data sets using SVMsIRJET Journal
This document discusses using support vector machines (SVMs) to assess credit risk with imbalanced data sets. SVMs have limited performance with imbalanced credit data where unpaid loans are less frequent than paid loans. The author develops an SVM model using two data resampling techniques - random oversampling and SMOTE - to address class imbalance. Performance is evaluated using various criteria like accuracy, sensitivity, specificity, and AUC. The results suggest resampling data can improve SVM performance for accurate credit risk prediction with imbalanced data.
This document summarizes and compares various machine learning models for credit scoring and investment decisions using explainable AI techniques. It finds that ensemble classifiers like random forests and neural networks outperform individual classifiers. LIME and SHAP techniques are used to explain ML credit scoring models. The study also develops new investment models using ML algorithms to maximize profit while minimizing risk. A variety of ML algorithms are tested, including logistic regression, decision trees, LDA, QDA, AdaBoost, random forests, and neural networks. The random forest and AdaBoost models are tuned with hyperparameters. Model performance is evaluated using metrics like accuracy, derived from a confusion matrix.
Here are the key points regarding the financial elder abuse questions:
a. The Bill Pay claims group uses several criteria to identify possible financial elder abuse, including unusual transaction patterns, large withdrawals/transfers, and signs the elder may be coerced or not acting voluntarily.
b. Bankers are trained to report any suspected financial elder abuse to the Elder Financial Abuse team by completing an incident report form with relevant details.
c. Yes, management obtains aggregated data and reports from the Elder Financial Abuse team on incidents reported by other teams like Online Claims. This information is used for ongoing training and to identify any patterns that could help prevent future abuse. Reporting processes are also reviewed and updated based on analysis of past incidents.
Machine learning algorithms can be used in various areas of banking and central banking. Specifically:
1) Traditional credit risk modeling can be enhanced with machine learning to predict probability of credit defaults based on borrower and macroeconomic variables.
2) Central banks can use credit bureau data and machine learning to monitor credit quality in real-time and provide recommendations to commercial banks.
3) Machine learning methods like random forests and neural networks outperform traditional models in time series forecasting of macroeconomic variables like inflation.
4) Unstructured text and narrative data from news, market commentary, and reports can be analyzed with machine learning to measure economic sentiment, risk, uncertainty and consensus.
MACHINE LEARNING CLASSIFIERS TO ANALYZE CREDIT RISKIRJET Journal
This document discusses using machine learning classifiers to analyze credit risk. It examines various machine learning techniques for credit risk analysis, including Bayesian classifiers, naive Bayes, decision trees, k-nearest neighbors, multilayer perceptrons, support vector machines, and ensemble methods like bagging and boosting. Two credit datasets from the UCI machine learning repository were used to test the accuracy of these classifiers. The results showed decision trees had the highest accuracy at 89.9% and 71.25% on the two datasets, while k-nearest neighbors had the lowest. Future work could involve rebuilding the models with more accurate data to improve performance. The objective of credit risk analysis is to help banks and financial institutions balance approving loans to creditworthy borrowers
A Compendium of Various Applications of Machine LearningIRJET Journal
This document provides a review of various applications of machine learning. It begins with an introduction to machine learning and discusses its applications in fields such as energy efficiency, intrusion detection, anomaly detection, quantitative finance, and cancer prediction and prognosis. Specific machine learning algorithms and techniques discussed include decision trees, naive Bayes, k-nearest neighbors, artificial neural networks, support vector machines, and more. The document also provides examples of machine learning applications in each field and references various research papers to support the discussed applications.
BANK LOAN PREDICTION USING MACHINE LEARNINGIRJET Journal
This document discusses using machine learning models to predict whether a bank will approve a loan application. It begins with an abstract that explains the goal is to accurately predict loan approvals to help banks and reduce risk. It then reviews related literature on using algorithms like decision trees, naive Bayes, random forests, and support vector machines (SVM) for loan prediction. The document proposes using models trained on applicant data to determine their likelihood of repayment. It describes collecting data, preparing it, selecting SVM as the best performing model, training and testing the model, and saving it to make predictions on new applicants. Accurately predicting approvals could help banks process applications faster and with less risk.
Similar to An Innovative Approach to Predict Bankruptcy (20)
Understanding the Impact and Challenges of Corona Crisis on Education Sector...vivatechijri
n the second week of March 2020, governments of all states in a country suddenly declared
shutting down of all colleges and schools for a temporary period of time as an immediate measure to stop the
spread of pandemic that is of novel corona virus. As the days pass by almost close to a month with no certainty
when they will again reopen. Due to pandemic like this an alarm bells have started sounding in the field of
education where a huge impact can be seen on teaching and learning process as well as on the entire education
sector in turn. The pandemic disruption like this is actually gave time to educators of today to really think about
the sector. Through the present research article, the author is highlighting on the possible impact of
coronavirus on education sector with the future challenges for education sector with possible suggestions.
LEADERSHIP ONLY CAN LEAD THE ORGANIZATION TOWARDS IMPROVEMENT AND DEVELOPMENT vivatechijri
This document discusses the importance of leadership in leading an organization towards improvement and development. It states that leadership is responsible for providing a clear vision and strategy to successfully achieve that vision. Effective leadership can impact the success of an organization by controlling its direction and motivating employees. Leadership is different from traditional management in that it guides employees towards organizational goals through open communication and motivation, rather than simply directing work. The paper concludes that only leadership can lead an organization to change according to its evolving environment, while management may simply follow old rules. Leadership is key to adapting to new market needs and trends.
The topic of assignment is a critical problem in mathematics and is further explored in the real
physical world. We try to implement a replacement method during this paper to solve assignment problems with
algorithm and solution steps. By using new method and computing by existing two methods, we analyse a
numerical example, also we compare the optimal solutions between this new method and two current methods. A
standardized technique, simple to use to solve assignment problems, may be the proposed method
Structural and Morphological Studies of Nano Composite Polymer Gel Electroly...vivatechijri
The document summarizes research on a nano composite polymer gel electrolyte containing SiO2 nanoparticles. Key points:
1. Polyvinylidene fluoride-co-hexafluoropropylene polymer was used as the base polymer mixed with propylene carbonate, magnesium perchlorate, and SiO2 nanoparticles to synthesize the nano composite polymer gel electrolyte.
2. The electrolyte was characterized using XRD, SEM, and FTIR which confirmed the homogeneous dispersion of SiO2 nanoparticles and increased amorphous nature of the electrolyte, enhancing its ion conductivity.
3. XRD showed decreased crystallinity and disappearance of polymer peaks upon addition of SiO2. SEM revealed
Theoretical study of two dimensional Nano sheet for gas sensing applicationvivatechijri
This study is focus on various two dimensional material for sensing various gases with theoretical
view for new research in gas sensing application. In this paper we review various two dimensional sheet such as
Graphene, Boron Nitride nanosheet, Mxene and their application in sensing various gases present in the
atmosphere.
METHODS FOR DETECTION OF COMMON ADULTERANTS IN FOODvivatechijri
Food is essential forliving. Food adulteration deceives consumers and can endanger their health. The
purpose of this document is to list common food adulterant methods commonly found in India. An adulterant is
a substance found in other substances such as food, cosmetics, pharmaceuticals, fuels, or other chemicals that
compromise the safety or effectiveness of that substance. The addition of adulterants is called adulteration. The
most common reason for adulteration is the use of undeclared materials by manufacturers that are cheaper than
the correct and declared ones. The adulterants can be harmful or reduce the effectiveness of the product, or
they can be harmless.
The novel ideas of being a entrepreneur is a key for everyone to get in the hustle, but developing a
idea from core requires a systematic plan, time management, time investment and most importantly client
attention. The Time required for developing may vary from idea to idea and strength of the team. Leadership to
build a team and manage the same throughout the peak of development is the main quality. Innovations and
Techniques to qualify the huddles is another aspect of Business Development and client Retention.
Innovation for supporting prosperity has for quite some time been a focus on numerous orders, including PC science, brain research, and human-PC connection. In any case, the meaning of prosperity isn't continuously clear and this has suggestions for how we plan for and evaluate advances that intend to cultivate it. Here, we talk about current meanings of prosperity and how it relates with and now and then is a result of self-amazing quality. We at that point center around how innovations can uphold prosperity through encounters of self-amazing quality, finishing with conceivable future bearings.
An Alternative to Hard Drives in the Coming Future:DNA-BASED DATA STORAGEvivatechijri
Demand for data storage is growing exponentially, but the capacity of existing storage media is not keeping up, there emerges a requirement for a storage medium with high capacity, high storage density, and possibility to face up to extreme environmental conditions. According to a research in 2018, every minute Google conducted 3.88 million searches, other people posted 49,000 photos on Instagram, sent 159,362,760 e-mails, tweeted 473,000 times and watched 4.33 million videos on YouTube. In 2020 it estimated a creation of 1.7 megabytes of knowledge per second per person globally, which translates to about 418 zettabytes during a single year. The magnetic or optical data-storage systems that currently hold this volume of 0s and 1s typically cannot last for quite a century. Running data centres takes vast amounts of energy. In short, we are close to have a substantial data-storage problem which will only become more severe over time. Deoxyribonucleic acid (DNA) are often potentially used for these purposes because it isn't much different from the traditional method utilized in a computer. DNA’s information density is notable, 215 petabytes or 215 million gigabytes of data can be stored in just one gram of DNA. First we can encode all data at a molecular level and then store it in a medium that will last for a while and not become out-dated just like floppy disks. Due to the improved techniques for reading and writing DNA, a rapid increase is observed in the amount of possible data storage in DNA.
The usage of chatbots has increased tremendously since past few years. A conversational interface is an interface that the user can interact with by means of a conversation. The conversation can occur by speech but also by text input. When a chatty interface uses text, it is also described as a chatbot or a conversational medium. During this study, the user experience factors of these so called chatbots were investigated. The prime objective is “to spot the state of the art in chatbot usability and applied human-computer interaction methodologies, to research the way to assess chatbots usability". Two sorts of chatbots are formulated, one with and one without personalisation factors. the planning of this research may be a two-by-two factorial design. The independent variables are the two chatbots (unpersonalised versus personalised) and thus the speci?c task or goal the user are ready to do with the chatbot within the ?nancial ?eld (a simple versus a posh task). The results are that there was no noteworthy interaction effect between personalisation and task on the user experience of chatbots. A signi?cant di?erence was found between the two tasks with regard to the user experience of chatbots, however this variation wasn't because of personalisation.
The Smart glasses Technology of wearable computing aims to identify the computing devices into today’s world.(SGT) are wearable Computer glasses that is used to add the information alongside or what the wearer sees. They are also able to change their optical properties at runtime.(SGT) is used to be one of the modern computing devices that amalgamate the humans and machines with the help of information and communication technology. Smart glasses is mainly made up of an optical head-mounted display or embedded wireless glasses with transparent heads- up display or augmented reality (AR) overlay in it. In recent years, it is been used in the medical and gaming applications, and also in the education sector. This report basically focuses on smart glasses, one of the categories of wearable computing which is very popular presently in the media and expected to be a big market in the next coming years. It Evaluate the differences from smart glasses to other smart devices. It introduces many possible different applications from the different companies for the different types of audience and gives an overview of the different smart glasses which are available presently and will be available after the next few years.
Future Applications of Smart Iot Devicesvivatechijri
With the Internet of Things (IoT) bit by bit creating as the resulting time of the headway of the Internet, it gets critical to see the diverse expected zones for the utilization of IoT and the research challenges that are connected with these applications going from splendid savvy urban areas, to medical care administrations, shrewd farming, collaborations and retail. IoT is needed to attack into for all expectations and purposes for all pieces of our day-to-day life. Despite the fact that the current IoT enabling advancements have immensely improved in the continuous years, there are so far different issues that require attention. Since the IoT ideas results from heterogeneous advancements, many examination difficulties will arise. In like manner, IoT is planning for new components of exploration to be finished. This paper presents the progressing headway of IoT advancements and inspects future applications.
Cross Platform Development Using Fluttervivatechijri
Today the development of cross-platform mobile application has under the state of compromise. The developers are not willing to choose an alternative of either building the similar app many times for many operating systems or to accept a lowest common denominator and optimal solution that will going to trade the native speed, accuracy for portability. The Flutter is an open-source SDK for creating high-performance, high fidelity mobile apps for the development of iOS and Android. Few significant features of flutter are - Just-in-time compilation (JIT), Ahead- of-time compilation (AOT compilation) into a native (system-dependent) machine code so that the resulting binary file can execute natively. The Flutter’s hot reload functionality helps us to understand quickly and easily experiment, build UIs, add features, and fix bugs. Hot reload works by injecting updated source code files into the running Dart Virtual Machine (VM). With the help of Flutter, we believe that we would be having a solution that gives us the best of both worlds: hardware accelerated graphics and UI, powered by native ARM code, targeting both popular mobile operating systems.
The Internet, today, has become an important part of our lives. The World Wide Web that was once a small and inaccessible data storage service is now large and valuable. Current activities partially or completely integrated into the physical world can be made to a higher standard. All activities related to our daily life are mapped and linked to another business in the digital world. The world has seen great strides in the Internet and in 3D stereoscopic displays. The time has come to unite the two to bring a new level of experience to the users. 3D Internet is a concept that is yet to be used and requires browsers to be equipped with in-depth visualization and artificial intelligence. When this material is included, the Internet concept of material may become a reality discussed in this paper. In this paper we have discussed the features, possible setting methods, applications, and advantages and disadvantages of using the Internet. With this paper we aim to provide a clear view of 3D Internet and the potential benefits associated with this obviously cost the amount of investment needed to be used.
Recommender System (RS) has emerged as a significant research interest that aims to assist users to seek out items online by providing suggestions that closely match their interests. Recommender system, an information filtering technology employed in many items is presented in internet sites as per the interest of users, and is implemented in applications like movies, music, venue, books, research articles, tourism and social media normally. Recommender systems research is usually supported comparisons of predictive accuracy: the higher the evaluation scores, the higher the recommender. One amongst the leading approaches was the utilization of advice systems to proactively recommend scholarly papers to individual researchers. In today's world, time has more value and therefore the researchers haven't any much time to spend on trying to find the proper articles in line with their research domain. Recommender Systems are designed to suggest users the things that best fit the user needs and preferences. Recommender systems typically produce an inventory of recommendations in one among two ways -through collaborative or content-based filtering. Additionally, both the general public and also the non-public used descriptive metadata are used. The scope of the advice is therefore limited to variety of documents which are either publicly available or which are granted copyright permits. Recommendation systems (RS) support users and developers of varied computer and software systems to beat information overload, perform information discovery tasks and approximate computation, among others.
The study LiFi (Light Fidelity) demonstrates about how can we use this technology as a medium of communication similar to Wifi . This is the latest technology proposed by Harold Haas in 2011. It explains about the process of transmitting data with the help of illumination of an Led bulb and about its speed intensity to transmit data. Basically in this paper, author will discuss about the technology and also explain that how we can replace from WiFi to LiFi . WiFi generally used for wireless coverage within the buildings while LiFi is capable for high intensity wireless data coverage in limited areas with no obstacles .This research paper represents introduction of the Lifi technology,performance,modulation and challenges. This research paper can be used as a reference and knowledge to develop some of LiFitechnology.
Social media platform and Our right to privacyvivatechijri
The advancement of Information Technology has hastened the ability to disseminate information across the globe. In particular, the recent trends in ‘Social Networking’ have led to a spark in personally sensitive information being published on the World Wide Web. While such socially active websites are creative tools for expressing one’s personality it also entails serious privacy concerns. Thus, Social Networking websites could be termed a double edged sword. It is important for the law to keep abreast of these developments in technology. The purpose of this paper is to demonstrate the limits of extending existing laws to battle privacy intrusions in the Internet especially in the context of social networking. It is suggested that privacy specific legislation is the most appropriate means of protecting online privacy. In doing so it is important to maintain a balance between the competing right of expression, the failure of which may hinder the reaping of benefits offered by Internet technology
THE USABILITY METRICS FOR USER EXPERIENCEvivatechijri
THE USABILITY METRICS FOR USER EXPERIENCE was innovatively created by Google engineers and it is ready for production in record time. The success of Google is to attributed the efficient search algorithm, and also to the underlying commodity hardware. As Google run number of application then Google’s goal became to build a vast storage network out of inexpensive commodity hardware. So Google create its own file system, named as THE USABILITY METRICS FOR USER EXPERIENCE that is GFS. THE USABILITY METRICS FOR USER EXPERIENCE is one of the largest file system in operation. Generally THE USABILITY METRICS FOR USER EXPERIENCE is a scalable distributed file system of large distributed data intensive apps. In the design phase of THE USABILITY METRICS FOR USER EXPERIENCE, in which the given stress includes component failures , files are huge and files are mutated by appending data. The entire file system is organized hierarchically in directories and identified by pathnames. The architecture comprises of multiple chunk servers, multiple clients and a single master. Files are divided into chunks, and that is the key design parameter. THE USABILITY METRICS FOR USER EXPERIENCE also uses leases and mutation order in their design to achieve atomicity and consistency. As of there fault tolerance, THE USABILITY METRICS FOR USER EXPERIENCE is highly available, replicas of chunk servers and master exists.
Google File System was innovatively created by Google engineers and it is ready for production in record time. The success of Google is to attributed the efficient search algorithm, and also to the underlying commodity hardware. As Google run number of application then Google’s goal became to build a vast storage network out of inexpensive commodity hardware. So Google create its own file system, named as Google File System that is GFS. Google File system is one of the largest file system in operation. Generally Google File System is a scalable distributed file system of large distributed data intensive apps. In the design phase of Google file system, in which the given stress includes component failures , files are huge and files are mutated by appending data. The entire file system is organized hierarchically in directories and identified by pathnames. The architecture comprises of multiple chunk servers, multiple clients and a single master. Files are divided into chunks, and that is the key design parameter. Google File System also uses leases and mutation order in their design to achieve atomicity and consistency. As of there fault tolerance, Google file system is highly available, replicas of chunk servers and master exists.
A Study of Tokenization of Real Estate Using Blockchain Technologyvivatechijri
Real estate is by far one of the most trusted investments that people have preferred, being a lucrative investment it provides a steady source of income in the form of lease and rents. Although there are numerous advantages, one of the key downsides of real estate investments is lack of liquidity. Thus, even though global real estate investments amount to about twice the size of investments in stock markets, the number of investors in the real estate market is significantly lower. Block chain technology has real potential in addressing the issues of liquidity and transparency, opening the market to even retail investors. Owing to the functionality and flexibility of creating Security Tokens, which are backed by real-world assets, real estate can be made liquid with the help of Special Purpose Vehicles. Tokens of ERC 777 standard, which represent fractional ownership of the real estate can be purchased by an investor and these tokens can also be listed on secondary exchanges. The robustness of Smart Contracts can enable the efficient transfer of tokens and seamless distribution of earnings amongst the investors. This work describes Ethereum blockchainbased solutions to make the existing Real Estate investment system much more efficient.
A vernier caliper is a precision instrument used to measure dimensions with high accuracy. It can measure internal and external dimensions, as well as depths.
Here is a detailed description of its parts and how to use it.
Unblocking The Main Thread - Solving ANRs and Frozen FramesSinan KOZAK
In the realm of Android development, the main thread is our stage, but too often, it becomes a battleground where performance issues arise, leading to ANRS, frozen frames, and sluggish Uls. As we strive for excellence in user experience, understanding and optimizing the main thread becomes essential to prevent these common perforrmance bottlenecks. We have strategies and best practices for keeping the main thread uncluttered. We'll examine the root causes of performance issues and techniques for monitoring and improving main thread health as wel as app performance. In this talk, participants will walk away with practical knowledge on enhancing app performance by mastering the main thread. We'll share proven approaches to eliminate real-life ANRS and frozen frames to build apps that deliver butter smooth experience.
Online music portal management system project report.pdfKamal Acharya
The iMMS is a unique application that is synchronizing both user
experience and copyrights while providing services like online music
management, legal downloads, artists’ management. There are several
other applications available in the market that either provides some
specific services or large scale integrated solutions. Our product differs
from the rest in a way that we give more power to the users remaining
within the copyrights circle.
Social media management system project report.pdfKamal Acharya
The project "Social Media Platform in Object-Oriented Modeling" aims to design
and model a robust and scalable social media platform using object-oriented
modeling principles. In the age of digital communication, social media platforms
have become indispensable for connecting people, sharing content, and fostering
online communities. However, their complex nature requires meticulous planning
and organization.This project addresses the challenge of creating a feature-rich and
user-friendly social media platform by applying key object-oriented modeling
concepts. It entails the identification and definition of essential objects such as
"User," "Post," "Comment," and "Notification," each encapsulating specific
attributes and behaviors. Relationships between these objects, such as friendships,
content interactions, and notifications, are meticulously established.The project
emphasizes encapsulation to maintain data integrity, inheritance for shared behaviors
among objects, and polymorphism for flexible content handling. Use case diagrams
depict user interactions, while sequence diagrams showcase the flow of interactions
during critical scenarios. Class diagrams provide an overarching view of the system's
architecture, including classes, attributes, and methods .By undertaking this project,
we aim to create a modular, maintainable, and user-centric social media platform that
adheres to best practices in object-oriented modeling. Such a platform will offer users
a seamless and secure online social experience while facilitating future enhancements
and adaptability to changing user needs.
Software Engineering and Project Management - Introduction to Project ManagementPrakhyath Rai
Introduction to Project Management: Introduction, Project and Importance of Project Management, Contract Management, Activities Covered by Software Project Management, Plans, Methods and Methodologies, some ways of categorizing Software Projects, Stakeholders, Setting Objectives, Business Case, Project Success and Failure, Management and Management Control, Project Management life cycle, Traditional versus Modern Project Management Practices.
Understanding Cybersecurity Breaches: Causes, Consequences, and PreventionBert Blevins
Cybersecurity breaches are a growing threat in today’s interconnected digital landscape, affecting individuals, businesses, and governments alike. These breaches compromise sensitive information and erode trust in online services and systems. Understanding the causes, consequences, and prevention strategies of cybersecurity breaches is crucial to protect against these pervasive risks.
Cybersecurity breaches refer to unauthorized access, manipulation, or destruction of digital information or systems. They can occur through various means such as malware, phishing attacks, insider threats, and vulnerabilities in software or hardware. Once a breach happens, cybercriminals can exploit the compromised data for financial gain, espionage, or sabotage. Causes of breaches include software and hardware vulnerabilities, phishing attacks, insider threats, weak passwords, and a lack of security awareness.
The consequences of cybersecurity breaches are severe. Financial loss is a significant impact, as organizations face theft of funds, legal fees, and repair costs. Breaches also damage reputations, leading to a loss of trust among customers, partners, and stakeholders. Regulatory penalties are another consequence, with hefty fines imposed for non-compliance with data protection regulations. Intellectual property theft undermines innovation and competitiveness, while disruptions of critical services like healthcare and utilities impact public safety and well-being.
A brief introduction to quadcopter (drone) working. It provides an overview of flight stability, dynamics, general control system block diagram, and the electronic hardware.
In May 2024, globally renowned natural diamond crafting company Shree Ramkrishna Exports Pvt. Ltd. (SRK) became the first company in the world to achieve GNFZ’s final net zero certification for existing buildings, for its two two flagship crafting facilities SRK House and SRK Empire. Initially targeting 2030 to reach net zero, SRK joined forces with the Global Network for Zero (GNFZ) to accelerate its target to 2024 — a trailblazing achievement toward emissions elimination.
1. VIVA-Tech International Journal for Research and Innovation Volume 1, Issue 2 (2019)
ISSN(Online): 2581-7280 Article No. 8
PP 1-6
1
www.viva-technology.org/New/IJRI
An Innovative Approach to Predict Bankruptcy
Mihir H. Panchal1
, Mayur B. Bodar1
, Sunny R. Maurya1
, Tatwadarshi P.
Nagarhalli1
1
(Computer Engineering Department, VIVA Institute of Technology, India)
Abstract: Bankruptcy is a legal status of a person or other organization that cannot repay their debts to
creditors. Bankruptcy prediction is the task of predicting bankruptcy and by doing various surveys we can avoid
financial distress of firms. It is a huge area of accounting and finance research. The significance of this area is
an important part of financial specialists and creditors in assessing the probability that a firm may go bankrupt
or not. Estimating the risk of corporate bankruptcies is very important as the effect of bankruptcy is on a global
level. The aim of predicting financial distress is to develop a predictive model that combines various economic
factors which allow foreseeing the financial status of a firm. In this domain, various methods were proposed that
were based on neural networks, Support Vector Machines, Decision Trees, Random Forests, Naïve Bayes,
Balanced Bagging and Logistic Regression. In this paper, we document our observations as we explore and build
a Restricted Boltzmann Machine to Bankruptcy Prediction. We started by carrying out data pre-processing where
we impute the missing data values using Mean Imputation. To solve the data imbalance issue, we apply the
Synthetic Minority Oversampling Technique (SMOTE) to oversample the minority class labels. Finally, we
analyze and evaluate the performance of the model.
Keywords – Artificial Neural Network, Decision Trees, Logistic Regression, Naïve Bayes, Random Forests,
Restricted Boltzmann machine, Support Vector Machine.
1. INTRODUCTION
In recent years, the problem of corporate bankruptcy has attracted the attention of many stakeholders in
the financial sectors such as business investors, market analyst, banking sectors, lawmakers and shareholders.
Predicting bankruptcy is not only important for decision making in financial institutions but also determines the
country’s financial distress to some extent as wrong decision-making in financial institutions can have a
catastrophic effect on national or sometimes a global scale [1].
Predicting bankruptcy is of great importance in financial decision making. No matter how big or small a
company is when it goes bankrupt it affects everyone on a global level and hence prediction of enterprise
bankruptcies is very necessary. Researchers have been working on this domain for a good amount of time. They
tried to understand the reasons for bankruptcy and eventually trying to avoid bankruptcy. The research on this
domain is also functional to the availability of data.
The aim of the bankruptcy prediction is to predict the financial condition of a company and its future
perspectives within the context of long-term operation on the market. It is a vast area of finance and econometrics
that combines knowledge about the historical data of prosperous and unsuccessful companies. Typically,
enterprises are quantified by numerous indicators that describe their business condition that is further used to
induce a mathematical model using past observations.
There are many related works which focus on predicting the best model for the given financial decision-
making problem. The proposed system not simply examine the probability of liquidation but also to analyze the
best training algorithm. The best training model will be a retreat based on the highest classification rate. In the
proposed method system, we use “Restricted Boltzmann machine”.
The dataset is based on the Polish companies’ bankruptcy data analyzed over 2000-2012. The
informational collection is assembled by Emerging Markets Information Services (EMIS).
2. VIVA-Tech International Journal for Research and Innovation Volume 1, Issue 2 (2019)
ISSN(Online): 2581-7280 Article No. 8
PP 1-6
2
www.viva-technology.org/New/IJRI
2. RELATED WORKS
G. Pranav and K. Govinda [1] have proposed to predict bankruptcy using Artificial Neural Network
(ANN). The specialty of this model is that it is able to take into account past experiences and hence make a more
accurate decision over a period of time. Here, they have used Random Forest as a learning algorithm. The mainly
operate by creating decision trees during the training time and outputting the class that is the mode of the classes
(classification) or mean prediction (regression) of the individual trees. The dataset that G. Pranav and K. Govinda
[1] have used is Polish Companies’ analyzed data which has 65 attributes but to predict bankruptcy they have
considered only 3 such as Solvency, Earnings before Interest and Taxes (EBIT) and Liquidity.
Z. Fatima and S. Achchab [2] analyzed suppliers and customers’ payment delays and based on that they
predicted the failure. In this study, Z. Fatima and S. Achchab [2] have used Multivariate Discriminant Analysis,
Logistic Regression, and Decision Trees. The database used contains annual financial statements data for a sample
of Moroccan firms which has a lot of missing data leading to misleading or improper prediction. Z. Fatima and S.
Achchab [2] used different methods and tried to make work for predicting bankruptcy with the limited dataset.
Y. Zaychenko [3] applied fuzzy neural networks adaptive neuro-fuzzy inference system (ANFIS) and
Takagi-Sugeno-Kang (TSK) and fuzzy Group Method of Data Handling (GMDH) is used. The probably of work
under fuzzy and incomplete data to use expert knowledge in a form of fuzzy methods. For training, they used 115
banks of Europe and the testing sample was of 50 banks
S. Fan et. al. [12] have proposed anomaly detection method to detect bankruptcy with Multivariate
Gaussian distribution, One-Class SVM and Isolation Forest. The experiment dataset is from the UCI machine
learning repository. It describes bankrupt about the financial condition of Polish companies. 1st Year, 2nd Year,
3rd Year, 4th Year and 5th Year cases with respect to the ratio of positive to negative samples are Ratio 1:25,
1:24, 1:20, 1:18 and 1:13 respectively.
C. Cheng and C. Chan [5] have proposed financial distress prediction (FDP) has become increasingly
essential in resolving corporate financial risk. Five classification methods utilized to identify financial distress are
Decision tree C4.5, IBK, SVM, Random Forest, and RBF Network. Data pre-processing the financial database
comes from Taiwan Economic Journal (TEJ) Corporation. The proposed system focuses on the feature selection
and variable selection to predict bankruptcy with special optimizing techniques.
Y. Lu, et. al. [6] is to focus on predicting bankruptcy with the new model of SVM augment by SPSO.
The core merits in it make full of use advantage of Switching PSO Algorithm (SPSO) to search for optimal kernel
parameter of SVM. The Data sets that they have used are from the UCI Machine Learning Repository donated on
9th Feb 2014, which consists of 143 Non-Bankruptcy samples and 107 Bankruptcy sample, the total number is
250 sample data sets. Each attribute contains three parameters.
The dataset that M. Wagle et. al. [7] have contains 240 cases, 112 of which are bankrupted cases and 128
are successful cases. Observations come from two to five years before bankruptcy took place. Among these
companies, 56 went bankrupt two to five years later. The data from the first year of all the 120 companies will be
used as training set to train the classifier and the second-year data of all the 120 companies will be used as the
testing set to test the classifier. M. Wagle et. al. [7] have used five classifiers for the attribute selection process
which were Bayesian network, decision tree, logistic regression, neural networks, and SVM. To increase the
accuracy of the prediction models, they found the two common techniques in WEKA called the boosting technique
and the bagging technique, are applied onto the classifiers to increase their prediction accuracy.
G. Kumar et. al. [8] objective of bankruptcy prediction was to determine whether an organization or
financial firm will go bankrupt or not. The objective of boosting algorithm is to assign more weights to the
misclassified instances so that the learner can focus more on them for the succeeding round to classify it
accurately. The dataset they had obtained from 500 French industrial firms during the year 2002 and 2003. The
proposed system focuses on various boosting algorithms like Logit-Boost, NFS-Boost to predict bankruptcy which
gives better results after comparing with the other boosting algorithms.
A. Aghaie et. al. [9] paper proposes that several significant methodological issues were related to the use
of naive Bayes Bayesian Network (BN) models to predict the bankruptcy. The proposed system focuses on two
methods of Bayes Bayesian Network to predict bankruptcy in which first method, only variables that have
important correlations with the variable of interest and second method they investigated the impact on a naive
Bayes model’s performance of the number of states into which continuous variables were discretized.
D. Kang et. al. [10] paper proposes that the genetic algorithm-based coverage optimization techniques of
the SVM ensemble to solve the multicollinearity problem. Considering this background, D. Kang et.al. [10]
focuses on the genetic algorithm-based coverage optimization techniques of the SVM ensemble (CO-SVM) to
predict bankruptcy which gives accurate results than the SVM & DT.
E. Zibanezhad et. al. [11] have used Clementine software and the method of classification and regression
tree are used for mining financial variables. The data collected from financial statements of firms accepted in
Tehran Stock Exchange (TSE) from 1996 to 2009 which contained a total of 25 required financial ratios. The
3. VIVA-Tech International Journal for Research and Innovation Volume 1, Issue 2 (2019)
ISSN(Online): 2581-7280 Article No. 8
PP 1-6
3
www.viva-technology.org/New/IJRI
structure of the decision tree is a tree structure similar to a flowchart the biggest tie in a tree is the root tie and the
leaf ties show groups or group distribution.
3. PROPOSED SYSTEM
The paper proposes Bankruptcy prediction using Restricted Boltzmann Machine (RBM). Restricted
Boltzmann Machines (RBMs) are neural networks that belong to so called Energy Based Models. This type of
neural networks may be not that familiar, yet this kind of neural networks gained big popularity in recent years in
the context of the Netflix Prize where RBMs achieved state of the art performance in collaborative filtering and
have beaten most of the competition. A restricted Boltzmann machine (RBM) is generative stochastic artificial
neural network that can learn a probability distribution over its set of inputs.
Polish dataset has been used for proposed system. It is hosted by the University of California Irvine (UCI)
Machin Learning Repository which is a huge repository of freely accessible datasets for research and learning
purposes for the Machine Learning/Data Science community. This information was collected from the Emerging
Markets Information Service [6] (EMIS), which is a database containing data on developing markets far and wide.
The bankrupt organizations were examined in the period 2000-2012, while the as yet working organizations were
assessed from 2007 to 2013.
Table 1: Summary of the Polish bankruptcy dataset
Dataset characteristic Multivariate
Number of Features 64
Number of Instances
Data
Total Instances Bankrupt
instances
Number of
Instances
1st
year 7027 271 6756
2nd
year 10173 400 9773
3rd
year 10503 495 10008
4th
year 9792 515 9227
5th
year 5910 410 5500
Feature characteristics Real values
Has missing data? Yes
Associated tasks Classification
Date donated 04-11-2016
In table 2, 2nd
column display, the total number of instances in all dataset and 3rd
column display the
number of instances or rows with missing values for at least one of the features 4th
Column display the number of
instances that would remain in each dataset if all rows with missing values were dropped. 5th
Column displays the
percent of data loss if all the rows with missing data values were dropped. As the data loss rate in most of the
datasets is more than 50%, it is now clear that we cannot simply drop the rows with missing values, as it leads to
a loss in the representativeness of data.
Table 2: Missing Data
Data
Set
#Total
Instances
# Instances
with
missing
values
# Instances that would
remain if all rows with
missing values were
dropped
% Data loss if rows
with missing values
were dropped
Year 1 7027 3833 3194 54.54 %
Year 2 10173 6085 4088 59.81 %
Year 3 10503 5618 4885 53.48 %
Year 4 9792 5023 4769 51.29 %
Year 5 5910 2879 3031 48.71 %
There are mainly three issues comes in missing data
1. Missing data information can introduce a major amount of bias.
2. Handling and analysis are of the data more difficult.
3. Create more reductions in efficiency.
4. VIVA-Tech International Journal for Research and Innovation Volume 1, Issue 2 (2019)
ISSN(Online): 2581-7280 Article No. 8
PP 1-6
4
www.viva-technology.org/New/IJRI
Table 3 display below summarizes the populations of class labels. 2nd
Column displays the total instances,
while the 3rd
Column and 4th
Column display the total number of instances with the class as Bankrupt and Non-
Bankrupt respectively. While 5th
Column shows the population percentage of the minority class, i.e., the
Bankruptcy class label, among the total population of the dataset. These numbers in column 5 display that there
is a huge data imbalance.
Table 3: Data Imbalance
Data
Set
# Total
Instances
# Bankrupt
instances in
this forecasting
period
# Non-
Bankrupt
instances in
this forecasting
period
Percentage of
minority class
samples
Year 1 7027 271 6756 3.85 %
Year 2 10173 400 9773 3.93 %
Year 3 10503 495 10008 4.71 %
Year 4 9792 515 9277 5.25 %
Year 5 5910 410 5500 6.93 %
To deal with missing data in the proposed system mean imputation is used. It is the process in which all
the missing value replace with their mean of that context variable. Once completed dataset will not contain any
missing values. Data imbalance will be dealt by oversampling the dataset using Synthetic Minority Oversampling
Technique (SMOTE). It is a majorly used oversampling technique.
Figure 1 shows the System flow of proposed system. In this above-mentioned dataset will be taken and
pre-processing will be done on that to create balanced data. This balanced data will not have any missing data and
will be balanced for both the instances. This balanced data now will be used to train Restricted Boltzmann
Machine.
Figure 1: Proposed System
As the dataset used contains 64 attributes. Users will have to upload provided .csv file for prediction
which contains the 64 attributes with their descriptions.
4. RESULTS AND ANALYSIS
Proposed system outperformed all the previously implemented models with giving much higher accuracy of
70.86%. While previously implemented best model i.e. Support Vector Machine (SVM) struggled to give
accuracy higher than 60%. Hence our proposed model is more efficient to predict bankruptcy and gives much
5. VIVA-Tech International Journal for Research and Innovation Volume 1, Issue 2 (2019)
ISSN(Online): 2581-7280 Article No. 8
PP 1-6
5
www.viva-technology.org/New/IJRI
better performance and result. Table 4 shows the comparison of accuracy of different models and proposed
system. In this we can clearly see that proposed system has higher accuracy than other models. Based on this
predictions user can take required actions and by using proposed system for prediction user can foresee the
financial status of company.
Table 4: Comparative Analysis
Model Dataset Accuracy
Multivariate Gaussian
distribution (MG)
Polish Dataset 0.89
Support Vector Machine
(SVM)
Polish Dataset 0.8975
RBM
(Proposed)
Polish Dataset 0.9614
4. CONCLUSION
In global level bankruptcy is the major issue. Bankruptcy prediction is very important for corporate world
and it must be done to reduce financial distress. Models that have been implemented before were only able to give
results around 60% which is less for such a big problem. The dataset that has been used in proposed system had a
lot of missing data and data that was in that was also unbalanced that is the companies that went bankrupt were
very less as compared to compares to the companies which didn’t go bankrupt. To deal with missing values Mean
imputation technique was used and to deal with unbalanced data issue Synthetic Minority Oversampling
Technique (SMOTE) was used. After pre-processing was done model was trained and it gave 70.86% which is
higher than all the other models that were created before as they were only able to give results around 60%. Most
was given by Support Vector Machine (SVM) at 63.8%. Having a better model and being able to foresee the
financial status of a company or organization can help in avoiding global financial problem.
REFERENCES
[1] G. Pranav, K. Govinda, “Bankruptcy Prediction Using Neural Network”, International Conference on Inventive Systems and Control
(ICISC), 2018, pp. 248-251.
[2] Z. Fatima, S. Achchab, “The impact of payment delays on bankruptcy prediction”, 3rd International Conference of Cloud Computing
Technologies and Applications, 2017.
[3] Y. Zaychenko, “Banks bankruptcy risk forecasting with application of FNN”, 11th International Scientific and Technical Conference
Computer Sciences and Information Technologies, 2016, pp.196-199.
[4] S. Karlos, S. Kotsiantis, N. Fazakis, “Effectiveness of semi-supervised learning in bankruptcy prediction”, 7th International
Conference on Information”, Intelligence, Systems & Applications (IISA), 2016.
[5] C. Cheng, C. Chan, “An attribute selection-based classifier to predict financial distress”, 12th International Conference on Natural
Computation, Fuzzy Systems and Knowledge Discovery, 2016, pp. 1119-1122.
[6] Y. Lu, J. Zhu, N. Zhang, “A hybrid switching PSO algorithm and support vector machines for bankruptcy prediction”, International
Conference on Mechatronics and Control (ICMC), 2014, pp. 1329-1333.
[7] M. Wagle, Z. Yang, Y. Benslimane, “Bankruptcy prediction using data mining techniques”, 8th International Conference of
Information and Communication Technology for Embedded Systems (IC-ICTES), 2017.
[8] G. Kumar, S. Roy, “Development of Hybrid Boosting Technique for Bankruptcy Prediction”, International Conference on Information
Technology (ICIT), 2016, pp. 248-P253.
[9] A. Aghaie, A. Saeedi, “Using Bayesian Networks for Bankruptcy Prediction: Empirical Evidence from Iranian Companies”,
International Conference on Information Management and Engineering, 2009, pp. 450-455.
[10] D. Kang, M. Kim, “Performance enhancement of SVM ensembles using genetic algorithm in bankruptcy prediction”, 3rd International
Conference on Advanced Computer Theory Engineering and (ICACTE), 2010, pp- V2-154 – V2-158.
6. VIVA-Tech International Journal for Research and Innovation Volume 1, Issue 2 (2019)
ISSN(Online): 2581-7280 Article No. 8
PP 1-6
6
www.viva-technology.org/New/IJRI
[11] E. zibanezhad, D. Foroghi, A. Monadjemi, “Applying decision tree to predict bankruptcy”, IEEE International Conference on
Computer Science and Automation Engineering, 2011, pp. 165-169.
[12] S. Fan, G. Liu, Z. Chen, “Anomaly detection method for bankruptcy prediction”, 4th International Conference on Systems and
`Informatics (ICSAI), 2011, pp. 1456-1460.
[13] https://en.wikipedia.org/wiki/Machine_learning , Last Accessed on 15th
Sept. 2018.
[14] https://en.wikipedia.org/wiki/Linear_regression , Last Accessed on 16th
Sept. 2018.
[15] https://en.wikipedia.org/wiki/Random_forest , Last Accessed on 16th
Sept. 2018.