skip to main content
10.1145/3400302.3415731acmconferencesArticle/Chapter ViewAbstractPublication PagesiccadConference Proceedingsconference-collections
research-article

NASCaps: a framework for neural architecture search to optimize the accuracy and hardware efficiency of convolutional capsule networks

Published: 17 December 2020 Publication History

Abstract

Deep Neural Networks (DNNs) have made significant improvements to reach the desired accuracy to be employed in a wide variety of Machine Learning (ML) applications. Recently the Google Brain's team demonstrated the ability of Capsule Networks (CapsNets) to encode and learn spatial correlations between different input features, thereby obtaining superior learning capabilities compared to traditional (i.e., non-capsule based) DNNs. However, designing CapsNets using conventional methods is a tedious job and incurs significant training effort. Recent studies have shown that powerful methods to automatically select the best/optimal DNN model configuration for a given set of applications and a training dataset are based on the Neural Architecture Search (NAS) algorithms. Moreover, due to their extreme computational and memory requirements, DNNs are employed using the specialized hardware accelerators in IoT-Edge/CPS devices.
In this paper, we propose NASCaps, an automated framework for the hardware-aware NAS of different types of DNNs, covering both traditional convolutional DNNs and CapsNets. We study the efficacy of deploying a multi-objective Genetic Algorithm (e.g., based on the NSGA-II algorithm). The proposed framework can jointly optimize the network accuracy and the corresponding hardware efficiency, expressed in terms of energy, memory, and latency of a given hardware accelerator executing the DNN inference. Besides supporting the traditional DNN layers (such as, convolutional and fully-connected), our framework is the first to model and supports the specialized capsule layers and dynamic routing in the NAS-flow. We evaluate our framework on different datasets, generating different network configurations, and demonstrate the tradeoffs between the different output metrics. We will open-source the complete framework and configurations of the Pareto-optimal architectures at https://github.com/ehw-fit/nascaps.

References

[1]
M. Abadi et al. Tensorflow: A system for large-scale machine learning. In OSDI, 2016.
[2]
M. Capra et al. An updated survey of efficient hardware architectures for accelerating deep convolutional neural networks. Future Internet, 2020.
[3]
K. Deb, A. Pratap, S. Agarwal, and T. Meyarivan. A fast and elitist multiobjective genetic algorithm: Nsga-ii. IEEE transactions on evolutionary computation, 2002.
[4]
S. Grigorescu, B. Trasnea, T. Cocias, and G. Macesanu. A survey of deep learning techniques for autonomous driving. Journal of Field Robotics, 2019.
[5]
A. Harlap et al. Pipedream: Fast and efficient pipeline parallel dnn training. ArXiv, 2018.
[6]
K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. CVPR, 2015.
[7]
K. He, X. Zhang, S. Ren, and J. Sun. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In ICCV, 2015.
[8]
G. E. Hinton, A. Krizhevsky, and S. D. Wang. Transforming autoencoders. In ICANN, 2011.
[9]
W. Jiang et al. Accuracy vs. efficiency: Achieving both through fpga-implementation aware neural architecture search. In DAC, 2019.
[10]
W. Jiang et al. Device-circuit-architecture co-exploration for computing-in-memory neural accelerators. IEEE Transactions on Computers, 2020.
[11]
A. Krizhevsky. Learning multiple layers of features from tiny images. Technical report, Department of Computer Science, University of Toronto, 2009.
[12]
A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional neural networks. Commun. ACM, 2017.
[13]
Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 1998.
[14]
Z. Lu, K. Deb, and V. N. Boddeti. Muxconv: Information multiplexing in convolutional neural networks. ArXiv, 2020.
[15]
A. Marchisio, M. A. Hanif, and M. Shafique. Capsacc: An efficient hardware accelerator for capsulenets with data reuse. In DATE, 2019.
[16]
A. Marchisio, M. A. Hanif, M. T. Teimoori, and M. Shafique. Capstore: Energy-efficient design and management of the on-chip memory for capsulenet inference accelerators. CoRR, abs/1902.01151, 2019.
[17]
A. Marchisio, V. Mrazek, M. A. Hanif, and M. Shafique. Red-cane: A systematic methodology for resilience analysis and design of capsule networks under approximations. In DATE, 2020.
[18]
A. Marchisio et al. Deep learning for edge computing: Current trends, cross-layer optimizations, and open research challenges. In ISVLSI, 2019.
[19]
A. Marchisio et al. Fastrcaps: An integrated framework for fast yet accurate training of capsule networks. In IJCNN, 2020.
[20]
A. Marchisio et al. Q-capsnets: A specialized framework for quantizing capsule networks. In DAC, 2020.
[21]
R. Miikkulainen et al. Evolving deep neural networks. arXiv, 2017.
[22]
V. Mrazek et al. Alwann: Automatic layer-wise approximation of deep neural network accelerators without retraining. In ICCAD, 2019.
[23]
Y. Netzer et al. Reading digits in natural images with unsupervised feature learning. NIPS, 2011.
[24]
K. Pearson. Note on regression and inheritance in the case of two parents. Proceedings of the Royal Society of London, 1895.
[25]
J. Rajasegaran et al. Deepcaps: Going deeper with capsule networks. In CVPR, 2019.
[26]
J. Redmon, S. K. Divvala, R. B. Girshick, and A. Farhadi. You only look once: Unified, real-time object detection. CVPR, 2016.
[27]
S. Sabour, N. Frosst, and G. E. Hinton. Dynamic routing between capsules. In NIPS, 2017.
[28]
M. Shafique et al. An overview of next-generation architectures for machine learning: Roadmap, opportunities and challenges in the iot era. In DATE, 2018.
[29]
K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. In ICLR, 2015.
[30]
D. Stamoulis et al. Single-path nas: Designing hardware-efficient convnets in less than 4 hours. In ECML/PKDD, 2019.
[31]
Y. Sun, B. Xue, M. Zhang, and G. G. Yen. Automatically designing cnn architectures using genetic algorithm for image classification. IEEE transactions on cybernetics, 2020.
[32]
V. Sze, Y.-H. Chen, T.-J. Yang, and J. S. Emer. Efficient processing of deep neural networks: A tutorial and survey. Proceedings of the IEEE, 2017.
[33]
C. Szegedy, S. Ioffe, V. Vanhoucke, and A. A. Alemi. Inception-v4, inception-resnet and the impact of residual connections on learning. In AAAI, 2017.
[34]
C. Szegedy et al. Going deeper with convolutions. In CVPR, 2015.
[35]
H. Xiao, K. Rasul, and R. Vollgraf. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. ArXiv, 2017.
[36]
L. Xie and A. L. Yuille. Genetic cnn. ICCV, 2017.
[37]
B. Zoph and Q. V. Le. Neural architecture search with reinforcement learning. In ICLR, 2017.
[38]
B. Zoph, V. Vasudevan, J. Shlens, and Q. V. Le. Learning transferable architectures for scalable image recognition. In CVPR, 2018.

Cited By

View all
  1. NASCaps: a framework for neural architecture search to optimize the accuracy and hardware efficiency of convolutional capsule networks

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    ICCAD '20: Proceedings of the 39th International Conference on Computer-Aided Design
    November 2020
    1396 pages
    ISBN:9781450380263
    DOI:10.1145/3400302
    • General Chair:
    • Yuan Xie
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    In-Cooperation

    • IEEE CAS
    • IEEE CEDA
    • IEEE CS

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 17 December 2020

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. DNNs
    2. accuracy
    3. capsule networks
    4. deep neural networks
    5. design space
    6. energy efficiency
    7. evolutionary algorithms
    8. genetic algorithms
    9. hardware accelerators
    10. latency
    11. memory
    12. multi-objective
    13. neural architecture search
    14. optimization

    Qualifiers

    • Research-article

    Conference

    ICCAD '20
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 457 of 1,762 submissions, 26%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)30
    • Downloads (Last 6 weeks)7
    Reflects downloads up to 09 Jan 2025

    Other Metrics

    Citations

    Cited By

    View all

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media