• norsk
    • English
  • English 
    • norsk
    • English
  • Login
View Item 
  •   Home
  • Norges Handelshøyskole
  • Thesis
  • Master Thesis
  • View Item
  •   Home
  • Norges Handelshøyskole
  • Thesis
  • Master Thesis
  • View Item
JavaScript is disabled for your browser. Some features of this site may not work without it.

Explaining Individual Predictions on Financially Distressed Companies Using Shapley Values

Dokset, Henrik Rodahl; Vindenes, Eirik
Master thesis
Thumbnail
View/Open
masterthesis.pdf (843.9Kb)
URI
https://hdl.handle.net/11250/2774871
Date
2021
Metadata
Show full item record
Collections
  • Master Thesis [4207]
Abstract
Prediction results from complex machine learning models can be challenging to interpret. Understanding these models is essential when trusting results in decision-making. In this master thesis, we will utilize Shapley values to explain individual predictions from a complex machine learning algorithm. Our aim is to explain why prediction models obtain their results, so people can interpret them better.

The chosen case is based on a thesis called “Predicting Financial Distress in Norway” by Zhang and Ye (2019) where they used logistic regression and random forest models. Their thesis predicts whether a company enters financial distress within the next two years or not. In this thesis, we will take advantage of the powerful algorithm in xgboost (extreme gradient boosting). To illustrate the benefits of using a complex model versus a simple model, we will also present a decision tree as our baseline.

Our explanation analysis shows that predictions made by xgboost can be explained with the Shapley value framework to obtain clear and intuitive explanations. Calculating Shapley values for a larger group of predictions enables proper understanding of the model by investigating which feature values lead to what probability increase or decrease of distress. The explanation framework enables detection of possible model bias which sometimes can lead to discrimination. We conclude that using Shapley values as an explanatory framework enables decision-makers to continue using complex machine learning models. This is important, as we find the tool satisfying relevant regulations for decisions made by automatic systems to be explained upon request.

Contact Us | Send Feedback

Privacy policy
DSpace software copyright © 2002-2019  DuraSpace

Service from  Unit
 

 

Browse

ArchiveCommunities & CollectionsBy Issue DateAuthorsTitlesSubjectsDocument TypesJournalsThis CollectionBy Issue DateAuthorsTitlesSubjectsDocument TypesJournals

My Account

Login

Statistics

View Usage Statistics

Contact Us | Send Feedback

Privacy policy
DSpace software copyright © 2002-2019  DuraSpace

Service from  Unit