Abstract
In this paper, we are focusing on the problem of interpreting Neural Networks on the instance level. The proposed approach uses the Feature Contributions, numerical values that domain experts further interpret to reveal some phenomena about a particular instance or model behaviour. In our method, Feature Contributions are calculated from the Random Forest model trained to mimic the Artificial Neural Network’s classification as close as possible. We assume that we can trust the Feature Contributions results when both predictions are the same, i.e., Neural Network and Feature Contributions give the same results. The results show that this highly depends on the level the Neural Network is trained because the error is then propagated to the Random Forest model. For good trained ANNs, we can trust in interpretation based on Feature Contributions on average in 80%.
More Information
Identification Number: | https://doi.org/10.1007/978-3-030-77964-1_12 |
---|---|
Status: | Published |
Refereed: | Yes |
Publisher: | Springer |
Depositing User (symplectic) | Deposited by Palczewska, Anna |
Date Deposited: | 02 Sep 2019 14:29 |
Last Modified: | 11 Jul 2024 08:02 |
Item Type: | Article |
Download
Note: this is the author's final manuscript and may differ from the published version which should be used for citation purposes.
| Preview
Export Citation
Explore Further
Read more research from the author(s):