skip to main content
10.1145/3594300.3594311acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicmaiConference Proceedingsconference-collections
research-article

Building Interpretable Deep CNN Based on Binary Tree Structure Convolution Layers

Published: 13 September 2023 Publication History

Abstract

In recent years, Deep CNN (DCNN) models have achieved great success in the field of computer vision. However, such models are still considered to lack interpretability. One of fundamental issues underlying this problem can be noted as follows: The decision-making of a DCNN model is considered as a “black-box” operation. In this study, we propose to use binary tree structure convolution layers (TSCL) to interpret the decision-making mechanism of a DCNN model in the image recognition task. First, we design a TSCL module, in which each parent layer generates two child layers, and then integrate them into a normal DCNN. Second, we design an information coding objective to guide each two child nodes of one parent node to learn the particular information coding that we expected. Through the experiments, we can verify that: 1) the logical process of decision-making made by ResNet models can be explained well based on the "decision information flow path" formed in the TSCL module; 2) the decision-path can reasonably interpret the decision reversal mechanism (Robustness mechanism) of the DCNN model; 3) the credibility of decision-making can be measured by the matching degree between the actual and expected decision-path.

References

[1]
Georgios Arvanitidis, Lars Kai Hansen, and Søren Hauberg. 2017. Latent space oddity: on the curvature of deep generative models. arXiv preprint arXiv:1710.11379 (2017).
[2]
Ken Chatfield, Karen Simonyan, Andrea Vedaldi, and Andrew Zisserman. 2014. Return of the devil in the details: Delving deep into convolutional nets. arXiv preprint arXiv:1405.3531 (2014).
[3]
Zhi Chen, Yijie Bei, and Cynthia Rudin. 2020. Concept whitening for interpretable image recognition. Nature Machine Intelligence 2, 12 (2020), 772–782.
[4]
Dawei Dai, Chengfu Tang, Guoyin Wang, and Shuyin Xia. 2021. Building partially understandable convolutional neural networks by differentiating class-related neural nodes. Neurocomputing 452 (2021), 169–181.
[5]
Dawei Dai, Liping Yu, and Hui Wei. 2020. Parameters sharing in residual neural networks. Neural Processing Letters 51, 2 (2020), 1393–1410.
[6]
Ming Ding, Chang Zhou, Qibin Chen, Hongxia Yang, and Jie Tang. 2019. Cognitive graph for multi-hop reading comprehension at scale. arXiv preprint arXiv:1905.05460 (2019).
[7]
Nicholas Frosst and Geoffrey Hinton. 2017. Distilling a neural network into a soft decision tree. arXiv preprint arXiv:1711.09784 (2017).
[8]
Amirata Ghorbani, James Wexler, James Zou, and Been Kim. 2019. Towards automatic concept-based explanations. arXiv preprint arXiv:1902.03129 (2019).
[9]
Xianfeng Gu, Feng Luo, Jian Sun, and S-T Yau. 2013. Variational principles for Minkowski type problems, discrete optimal transport, and discrete Monge-Ampere equations. arXiv preprint arXiv:1302.5472 (2013).
[10]
Shanyan Guan, Ying Tai, Bingbing Ni, Feida Zhu, Feiyue Huang, and Xiaokang Yang. 2020. Collaborative learning for faster stylegan embedding. arXiv preprint arXiv:2007.01758 (2020).
[11]
Matthew Hausknecht and Peter Stone. 2015. Deep recurrent q-learning for partially observable mdps. In 2015 aaai fall symposium series.
[12]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition. 770–778.
[13]
Zhuxi Jiang, Yin Zheng, Huachun Tan, Bangsheng Tang, and Hanning Zhou. 2016. Variational deep embedding: An unsupervised and generative approach to clustering. arXiv preprint arXiv:1611.05148 (2016).
[14]
Been Kim, Martin Wattenberg, Justin Gilmer, Carrie Cai, James Wexler, Fernanda Viegas, 2018. Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (tcav). In International conference on machine learning. PMLR, 2668–2677.
[15]
Hao Li, Zheng Xu, Gavin Taylor, Christoph Studer, and Tom Goldstein. 2017. Visualizing the loss landscape of neural nets. arXiv preprint arXiv:1712.09913 (2017).
[16]
Matiur Rahman Minar and Jibon Naher. 2018. Recent advances in deep learning: An overview. arXiv preprint arXiv:1807.08169 (2018).
[17]
P Rastegari, M Majidi, and M Khalilian. 2013. Analysis of WiMAX Performance Improvement Using Serial and Parallel Concatenated Convolutional Codes. International Journal of Computer Theory and Engineering 5, 2 (2013), 326.
[18]
Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. 2016. " Why should i trust you?" Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining. 1135–1144.
[19]
Ramprasaath R Selvaraju, Abhishek Das, Ramakrishna Vedantam, Michael Cogswell, Devi Parikh, and Dhruv Batra. 2016. Grad-CAM: Why did you say that?arXiv preprint arXiv:1611.07450 (2016).
[20]
TTP Thanh, Caleb Vununu, Sukhrob Atoev, Suk-Hwan Lee, and Ki-Ryong Kwon. 2018. Leukemia blood cell image classification using convolutional neural network. International Journal of Computer Theory and Engineering 10, 2 (2018), 54–58.
[21]
Paul Upchurch, Jacob Gardner, Geoff Pleiss, Robert Pless, Noah Snavely, Kavita Bala, and Kilian Weinberger. 2017. Deep feature interpolation for image content changes. In Proceedings of the IEEE conference on computer vision and pattern recognition. 7064–7073.
[22]
Hui Wei, Dawei Dai, and Yijie Bu. 2017. A plausible neural circuit for decision making and its formation based on reinforcement learning. Cognitive neurodynamics 11, 3 (2017), 259.
[23]
Jason Yosinski, Jeff Clune, Anh Nguyen, Thomas Fuchs, and Hod Lipson. 2015. Understanding neural networks through deep visualization. arXiv preprint arXiv:1506.06579 (2015).
[24]
Aiguo Zhou, Zhenyu Li, and Yong Shen. 2019. Anomaly detection of CAN bus messages using a deep neural network for autonomous vehicles. Applied Sciences 9, 15 (2019), 3174.
[25]
Bolei Zhou, Yiyou Sun, David Bau, and Antonio Torralba. 2018. Interpretable basis decomposition for visual explanation. In Proceedings of the European Conference on Computer Vision (ECCV). 119–134.
[26]
Zhi-Hua Zhou and Ji Feng. 2017. Deep forest. arXiv preprint arXiv:1702.08835 (2017).

Index Terms

  1. Building Interpretable Deep CNN Based on Binary Tree Structure Convolution Layers

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    ICMAI '23: Proceedings of the 2023 8th International Conference on Mathematics and Artificial Intelligence
    April 2023
    106 pages
    ISBN:9781450399982
    DOI:10.1145/3594300
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 13 September 2023

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Binary tree
    2. Decision-making
    3. Information coding
    4. Interpretable neural network

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    ICMAI 2023

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 27
      Total Downloads
    • Downloads (Last 12 months)13
    • Downloads (Last 6 weeks)3
    Reflects downloads up to 04 Feb 2025

    Other Metrics

    Citations

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media