skip to main content
10.1145/3219819.3220063acmotherconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article

Exact and Consistent Interpretation for Piecewise Linear Neural Networks: A Closed Form Solution

Published: 19 July 2018 Publication History

Abstract

Strong intelligent machines powered by deep neural networks are increasingly deployed as black boxes to make decisions in risk-sensitive domains, such as finance and medical. To reduce potential risk and build trust with users, it is critical to interpret how such machines make their decisions. Existing works interpret a pre-trained neural network by analyzing hidden neurons, mimicking pre-trained models or approximating local predictions. However, these methods do not provide a guarantee on the exactness and consistency of their interpretations. In this paper, we propose an elegant closed form solution named $OpenBox$ to compute exact and consistent interpretations for the family of Piecewise Linear Neural Networks (PLNN). The major idea is to first transform a PLNN into a mathematically equivalent set of linear classifiers, then interpret each linear classifier by the features that dominate its prediction. We further apply $OpenBox$ to demonstrate the effectiveness of non-negative and sparse constraints on improving the interpretability of PLNNs. The extensive experiments on both synthetic and real world data sets clearly demonstrate the exactness and consistency of our interpretation.

Supplementary Material

MP4 File (chu_neural_networks.mp4)

References

[1]
Aishwarya Agrawal, Dhruv Batra, and Devi Parikh . 2016. Analyzing the behavior of visual question answering models. arXiv:1606.07356 (2016).
[2]
Jimmy Ba and Rich Caruana . 2014. Do deep nets really need to be deep?. In NIPS. 2654--2662.
[3]
Osbert Bastani, Carolyn Kim, and Hamsa Bastani . 2017. Interpreting Blackbox Models via Model Extraction. arXiv:1705.08504 (2017).
[4]
C Bishop . 2007. Pattern Recognition and Machine Learning (Information Science and Statistics). Springer, New York (2007).
[5]
C. Cao, X. Liu, Y Yang, Y. Yu, J. Wang, Z. Wang, Y. Huang, L. Wang, C. Huang, et almbox. . 2015. Look and think twice: Capturing top-down visual attention with feedback convolutional neural networks. In ICCV. 2956--2964.
[6]
RJ Caron, JF McDonald, and CM Ponic . 1989. A degenerate extreme point strategy for the classification of linear constraints as redundant or necessary. JOTA Vol. 62, 2 (1989), 225--237.
[7]
Z. Che, S. Purushotham, R. Khemani, and Y. Liu . 2015. Distilling knowledge from deep networks with applications to healthcare domain. arXiv:1512.03542 (2015).
[8]
Jan Chorowski and Jacek M Zurada . 2015. Learning understandable neural networks with nonnegative weight constraints. TNNLS Vol. 26, 1 (2015), 62--69.
[9]
Alexey Dosovitskiy and Thomas Brox . 2016. Inverting visual representations with convolutional networks CVPR. 4829--4837.
[10]
D. Erhan, Yoshua Bengio, A. Courville, and P. Vincent . 2009. Visualizing higher-layer features of a deep network. University of Montreal Vol. 1341 (2009), 3.
[11]
Ruth Fong and Andrea Vedaldi . 2017. Interpretable Explanations of Black Boxes by Meaningful Perturbation. arXiv:1704.03296 (2017).
[12]
Nicholas Frosst and Geoffrey Hinton . 2017. Distilling a Neural Network Into a Soft Decision Tree. arXiv:1711.09784 (2017).
[13]
Amirata Ghorbani, Abubakar Abid, and James Zou . 2017. Interpretation of Neural Networks is Fragile. arXiv:1710.10547 (2017).
[14]
Xavier Glorot, Antoine Bordes, and Yoshua Bengio . 2011. Deep sparse rectifier neural networks. In ICAIS. 315--323.
[15]
Ian Goodfellow, Yoshua Bengio, and Aaron Courville . 2016. Deep Learning. MIT Press. http://www.deeplearningbook.org
[16]
Ian J Goodfellow, David Warde-Farley, Mehdi Mirza, Aaron Courville, and Yoshua Bengio . 2013. Maxout networks. arXiv:1302.4389 (2013).
[17]
B. Goodman and S. Flaxman . 2016. European Union regulations on algorithmic decision-making and a" right to explanation". arXiv:1606.08813 (2016).
[18]
Nick Harvey, Chris Liaw, and Abbas Mehrabian . 2017. Nearly-tight VC-dimension bounds for piecewise linear neural networks. arXiv:1703.02930 (2017).
[19]
K. He, X. Zhang, S. Ren, and J. Sun . 2015. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In ICCV. 1026--1034.
[20]
Geoffrey Hinton, Oriol Vinyals, and Jeff Dean . 2015. Distilling the knowledge in a neural network. arXiv:1503.02531 (2015).
[21]
Patrik O Hoyer . 2002. Non-negative sparse coding. In WNNSP. 557--565.
[22]
Pieter-Jan Kindermans, Sara Hooker, Julius Adebayo, Maximilian Alber, Kristof T Schütt, Sven D"ahne, Dumitru Erhan, and Been Kim . 2017. The (Un) reliability of saliency methods. arXiv:1711.00867 (2017).
[23]
Pang Wei Koh and Percy Liang . 2017. Understanding black-box predictions via influence functions. arXiv:1703.04730 (2017).
[24]
Pascal Koiran and Eduardo D Sontag . 1996. Neural networks with quadratic VC dimension. In NIPS. 197--203.
[25]
Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton . 2012. Imagenet classification with deep convolutional neural networks NIPS. 1097--1105.
[26]
Yann LeCun, Yoshua Bengio, and Geoffrey Hinton . 2015. Deep learning. nature Vol. 521, 7553 (2015), 436.
[27]
Honglak Lee, Alexis Battle, Rajat Raina, and Andrew Y Ng . 2007. Efficient sparse coding algorithms. In NIPS. 801--808.
[28]
Jiwei Li, Xinlei Chen, Eduard Hovy, and Dan Jurafsky . 2015. Visualizing and understanding neural models in NLP. arXiv:1506.01066 (2015).
[29]
Aravindh Mahendran and Andrea Vedaldi . 2015. Understanding deep image representations by inverting them CVPR. 5188--5196.
[30]
Guido F Montufar, Razvan Pascanu, Kyunghyun Cho, and Yoshua Bengio . 2014. On the number of linear regions of deep neural networks NIPS. 2924--2932.
[31]
Vinod Nair and Geoffrey E Hinton . 2010. Rectified linear units improve restricted boltzmann machines ICML. 807--814.
[32]
R. B. Palm . 2012. Prediction as a candidate for learning deep hierarchical models of data. (2012).
[33]
Razvan Pascanu, Guido Montufar, and Yoshua Bengio . 2013. On the number of response regions of deep feed forward networks with piece-wise linear activations. arXiv:1312.6098 (2013).
[34]
Nadeem N Rather, Chintan O Patel, and Sharib A Khan . 2017. Using Deep Learning Towards Biomedical Knowledge Discovery. IJMSC Vol. 3, 2 (2017), 1.
[35]
Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin . 2016. Why should i trust you?: Explaining the predictions of any classifier KDD. ACM, 1135--1144.
[36]
R. R Selvaraju, A. Das, R. Vedantam, M. Cogswell, D. Parikh, and D. Batra . 2016. Grad-cam: Why did you say that? visual explanations from deep networks via gradient-based localization. arXiv:1610.02391 (2016).
[37]
A. Shrikumar, P. Greenside, and A. Kundaje . 2017. Learning important features through propagating activation differences. arXiv:1704.02685 (2017).
[38]
Karen Simonyan, Andrea Vedaldi, and Andrew Zisserman . 2013. Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv:1312.6034 (2013).
[39]
D. Smilkov, N. Thorat, B. Kim, F. Viégas, and M. Wattenberg . 2017. SmoothGrad: removing noise by adding noise. arXiv:1706.03825 (2017).
[40]
Eduardo D Sontag . 1998. VC dimension of neural networks. NATO ASI Series F Computer and Systems Sciences Vol. 168 (1998), 69--96.
[41]
Mukund Sundararajan, Ankur Taly, and Qiqi Yan . 2017. Axiomatic Attribution for Deep Networks. arXiv:1703.01365 (2017).
[42]
M. Wu, M. C Hughes, S. Parbhoo, M. Zazzi, V. Roth, and F. Doshi-Velez . 2018. Beyond Sparsity: Tree Regularization of Deep Models for Interpretability. AAAI (2018).
[43]
Han Xiao, Kashif Rasul, and Roland Vollgraf . 2017. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms. (2017). showeprint{arXiv}cs.LG/cs.LG/1708.07747
[44]
J. Yosinski, J. Clune, A. Nguyen, T. Fuchs, and H. Lipson . 2015. Understanding neural networks through deep visualization. arXiv:1506.06579 (2015).
[45]
Rich Zemel, Yu Wu, Kevin Swersky, Toni Pitassi, and Cynthia Dwork . 2013. Learning fair representations. In ICML. 325--333.
[46]
Bolei Zhou, David Bau, Aude Oliva, and Antonio Torralba . 2017. Interpreting Deep Visual Representations via Network Dissection. arXiv:1711.05611 (2017).
[47]
Bolei Zhou, Aditya Khosla, Agata Lapedriza, Aude Oliva, and Antonio Torralba . 2016. Learning deep features for discriminative localization CVPR. 2921--2929.
[48]
J. Zhu, Y. Shan, JC Mao, D. Yu, H. Rahmanian, and Y. Zhang . 2017. Deep embedding forest: Forest-based serving with deep embedding features KDD. 1703--1711.

Cited By

View all

Index Terms

  1. Exact and Consistent Interpretation for Piecewise Linear Neural Networks: A Closed Form Solution

        Recommendations

        Comments

        Information & Contributors

        Information

        Published In

        cover image ACM Other conferences
        KDD '18: Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining
        July 2018
        2925 pages
        ISBN:9781450355520
        DOI:10.1145/3219819
        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Sponsors

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        Published: 19 July 2018

        Permissions

        Request permissions for this article.

        Check for updates

        Author Tags

        1. closed form
        2. deep neural network
        3. exact and consistent interpretation

        Qualifiers

        • Research-article

        Funding Sources

        • the NSERC Strategic Grant program
        • the NSERC Discovery Grant program
        • the Canada Research Chair program

        Conference

        KDD '18
        Sponsor:

        Acceptance Rates

        KDD '18 Paper Acceptance Rate 107 of 983 submissions, 11%;
        Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

        Contributors

        Other Metrics

        Bibliometrics & Citations

        Bibliometrics

        Article Metrics

        • Downloads (Last 12 months)88
        • Downloads (Last 6 weeks)7
        Reflects downloads up to 14 Sep 2024

        Other Metrics

        Citations

        Cited By

        View all

        View Options

        Get Access

        Login options

        View options

        PDF

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader

        Media

        Figures

        Other

        Tables

        Share

        Share

        Share this Publication link

        Share on social media