skip to main content
10.1145/3449639.3459308acmconferencesArticle/Chapter ViewAbstractPublication PagesgeccoConference Proceedingsconference-collections
research-article

Using novelty search to explicitly create diversity in ensembles of classifiers

Published: 26 June 2021 Publication History

Abstract

The diversity between individual learners in an ensemble is known to influence its performance. However, there is no standard agreement on how diversity should be defined, and thus how to exploit it to construct a high-performing classifier. We propose two new behavioural diversity metrics based on the divergence of errors between models. Following a neuroevolution approach, these metrics are then used to guide a novelty search algorithm to search a space of neural architectures and discover behaviourally diverse classifiers, iteratively adding the models with high diversity score to an ensemble. The parameters of each ANN are tuned individually with a standard gradient descent procedure. We test our approach on three benchmark datasets from Computer Vision --- CIFAR-10, CIFAR-100, and SVHN --- and find that the ensembles generated significantly outperform ensembles created without explicitly searching for diversity and that the error diversity metrics we propose lead to better results than others in the literature. We conclude that our empirical results signpost an improved approach to promoting diversity in ensemble learning, identifying what sort of diversity is most relevant and proposing an algorithm that explicitly searches for it without selecting for accuracy.

References

[1]
Alejandro Baldominos, Yago Saez, and Pedro Isasi. 2019. Hybridizing evolutionary computation and deep neural networks: an approach to handwriting recognition using committees and transfer learning. Complexity 2019 (2019).
[2]
Urvesh Bhowan, Mark Johnston, Mengjie Zhang, and Xin Yao. 2012. Evolving diverse ensembles using genetic programming for classification with unbalanced data. IEEE Transactions on Evolutionary Computation 17, 3 (2012), 368--386.
[3]
Urvesh Bhowan, Mark Johnston, Mengjie Zhang, and Xin Yao. 2013. Reusing genetic programming for ensemble selection in classification of unbalanced data. IEEE Transactions on Evolutionary Computation 18, 6 (2013), 893--908.
[4]
Yijun Bian and Huanhuan Chen. 2019. When does Diversity Help Generalization in Classification Ensembles? (2019). arXiv:1910.13631
[5]
Gavin Brown, Jeremy L. Wyatt, and Peter Tiňo. 2005. Managing diversity in regression ensembles. (2005).
[6]
Rui P. Cardoso, Emma Hart, David Burth Kurka, and Jeremy Pitt. 2021. WILDA: Wide Learning of Diverse Architectures for Classification of Large Datasets. In Applications of Evolutionary Computation - 24th International Conference, EvoApplications 2021, Held as Part of EvoStar 2021, Virtual Event, April 7-9, 2021, Proceedings (Lecture Notes in Computer Science), Pedro A. Castillo and Juan Luis Jiménez Laredo (Eds.), Vol. 12694. Springer, 649--664.
[7]
Rui P Cardoso, Emma Hart, and Jeremy V Pitt. 2020. Diversity-Driven Wide Learning for Training Distributed Classification Models. In Proceedings of the 2020 Genetic and Evolutionary Computation Conference Companion (GECCO '20). Association for Computing Machinery, New York, NY, USA, 119--120.
[8]
Tsung-Han Chan, Kui Jia, Shenghua Gao, Jiwen Lu, Zinan Zeng, and Yi Ma. 2015. PCANet: A Simple Deep Learning Baseline for Image Classification? IEEE Transactions on Image Processing 24, 12 (Dec 2015), 5017--5032.
[9]
Ekin D. Cubuk, Barret Zoph, Dandelion Mane, Vijay Vasudevan, and Quoc V. Le. 2019. AutoAugment: Learning Augmentation Policies from Data. (2019). arXiv:cs.CV/1805.09501
[10]
Terrance DeVries and Graham W. Taylor. 2017. Improved Regularization of Convolutional Neural Networks with Cutout. (2017). arXiv:cs.CV/1708.04552
[11]
Thomas G Dietterich. 2000. Ensemble methods in machine learning. In International workshop on multiple classifier systems. Springer, 1--15.
[12]
Dario Floreano, Peter Dürr, and Claudio Mattiussi. 2008. Neuroevolution: From architectures to learning. (2008).
[13]
Nicolás García-Pedrajas, César Hervás-Martínez, and Domingo Ortiz-Boyer. 2005. Cooperative coevolution of artificial neural network ensembles for pattern classification. IEEE transactions on evolutionary computation 9, 3 (2005), 271--302.
[14]
Jorge Gomes, Pedro Mariano, and Anders Lyhne Christensen. 2015. Devising effective novelty search algorithms: A comprehensive empirical study. In GECCO 2015 - Proceedings of the 2015 Genetic and Evolutionary Computation Conference.
[15]
S Gu, R Cheng, and Y Jin. 2015. Multi-objective ensemble generation. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery 5, 5 (2015), 234--245.
[16]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition. 770--778.
[17]
Yanping Huang, Youlong Cheng, Ankur Bapna, Orhan Firat, Mia Xu Chen, Dehao Chen, HyoukJoong Lee, Jiquan Ngiam, Quoc V. Le, Yonghui Wu, and Zhifeng Chen. 2019. GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism. (2019). arXiv:cs.CV/1811.06965
[18]
Yangqing Jia, Chang Huang, and Trevor Darrell. 2012. Beyond spatial pyramids: Receptive field learning for pooled image features. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition.
[19]
Alexander Kolesnikov, Lucas Beyer, Xiaohua Zhai, Joan Puigcerver, Jessica Yung, Sylvain Gelly, and Neil Houlsby. 2020. Big Transfer (BiT): General Visual Representation Learning. (2020). arXiv:cs.CV/1912.11370
[20]
Alex Krizhevsky. 2009. Learning Multiple Layers of Features from Tiny Images. ... Science Department, University of Toronto, Tech.... (2009). https://doi.org/10.1.1.222.9220 arXiv:arXiv:1011.1669v3
[21]
Ludmila I. Kuncheva and Christopher J. Whitaker. 2003. Measures of diversity in classifier ensembles and their relationship with the ensemble accuracy. Machine Learning (2003).
[22]
Joel Lehman and Kenneth O. Stanley. 2011. Abandoning objectives: Evolution through the search for novelty alone. Evolutionary Computation (2011).
[23]
Joel Lehman and Kenneth O Stanley. 2011. Evolving a diversity of virtual creatures through novelty search and local competition. In Proceedings of the 13th annual conference on Genetic and evolutionary computation. ACM, 211--218.
[24]
Senwei Liang, Yuehaw Khoo, and Haizhao Yang. 2020. Drop-Activation: Implicit Parameter Reduction and Harmonic Regularization. (2020). arXiv:cs.LG/1811.05850
[25]
Chenxi Liu, Liang-Chieh Chen, Florian Schroff, Hartwig Adam, Wei Hua, Alan Yuille, and Li Fei-Fei. 2019. Auto-DeepLab: Hierarchical Neural Architecture Search for Semantic Image Segmentation. (2019). arXiv:cs.CV/1901.02985
[26]
Julien Mairal, Piotr Koniusz, Zaid Harchaoui, and Cordelia Schmid. 2014. Convolutional Kernel Networks. (2014). arXiv:cs.CV/1406.3332
[27]
Mateusz Malinowski and Mario Fritz. 2014. Learning Smooth Pooling Regions for Visual Recognition.
[28]
Mark D. McDonnell and Tony Vladusich. 2015. Enhanced Image Classification With a Fast-Learning Shallow Convolutional Neural Network. (2015). arXiv:cs.NE/1503.04596
[29]
Jean-Baptiste Mouret and Jeff Clune. 2015. Illuminating search spaces by mapping elites. (2015). arXiv:cs.AI/1504.04909
[30]
Kaustuv Nag and Nikhil R Pal. 2015. A multiobjective genetic programming-based ensemble for simultaneous feature selection and classification. IEEE transactions on cybernetics 46, 2 (2015), 499--510.
[31]
Vinod Nair and Geoffrey E. Hinton. 2010. Rectified linear units improve Restricted Boltzmann machines. In ICML 2010 - Proceedings, 27th International Conference on Machine Learning.
[32]
Yuval Netzer, Tao Wang, Adam Coates, Alessandro Bissacco, Bo Wu, and Andrew Y Ng. 2011. Reading Digits in Natural Images with Unsupervised Feature Learning. (2011).
[33]
R Pasti, L N De Castro, G P Coelho, and F J Von Zuben. 2010. Neural network ensembles: Immune-inspired approaches to the diversity of components. Natural Computing 9, 3 (2010), 625--653.
[34]
Adam Paszke, Sam Gross, Soumith Chintala, Gregory Chanan, Edward Yang, Zachary Devito Facebook, A I Research, Zeming Lin, Alban Desmaison, Luca Antiga, Orobix Srl, and Adam Lerer. 2019. Automatic differentiation in PyTorch. In Advances in Neural Information Processing Systems 32.
[35]
Esteban Real, Alok Aggarwal, Yanping Huang, and Quoc V. Le. 2019. Regularized Evolution for Image Classifier Architecture Search. Proceedings of the AAAI Conference on Artificial Intelligence 33, 01 (Jul. 2019), 4780--4789.
[36]
Pierre Sermanet, Soumith Chintala, and Yann LeCun. 2012. Convolutional Neural Networks Applied to House Numbers Digit Classification. (2012). arXiv:cs.CV/1204.3968
[37]
Julien Siems, Lucas Zimmer, Arber Zela, Jovita Lukasik, Margret Keuper, and Frank Hutter. 2020. NAS-Bench-301 and the case for surrogate benchmarks for neural architecture search. (2020). arXiv:2008.09777
[38]
Kenneth O. Stanley, Jeff Clune, Joel Lehman, and Risto Miikkulainen. 2019. Designing neural networks through neuroevolution. (2019).
[39]
Kenneth O. Stanley and Risto Miikkulainen. 2002. Evolving Neural Networks Through Augmenting Topologies. Evolutionary Computation 10, 2 (2002), 99--127. http://nn.cs.utexas.edu/?stanley:ec02
[40]
Y. Sun, B. Xue, M. Zhang, G. G. Yen, and J. Lv. 2020. Automatically Designing CNN Architectures Using the Genetic Algorithm for Image Classification. IEEE Transactions on Cybernetics 50, 9 (2020), 3840--3854.
[41]
Paul A. Szerlip, Gregory Morse, Justin K. Pugh, and Kenneth O. Stanley. 2014. Unsupervised Feature Learning through Divergent Discriminative Feature Accumulation. (2014). arXiv:cs.NE/1406.1833
[42]
Mingxing Tan and Quoc V. Le. 2020. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. (2020). arXiv:cs.LG/1905.11946
[43]
Rick Van Krevelen. 2005. Error Diversity in Classification Ensembles. Ph.D. Dissertation.
[44]
David H. Wolpert. 1992. Stacked generalization. Neural Networks (1992).
[45]
Sergey Zagoruyko and Nikos Komodakis. 2016. Wide Residual Networks. (2016). arXiv:cs.CV/1605.07146
[46]
Yifeng Zhang and Siddhartha Bhattacharyya. 2004. Genetic programming in classifying large-scale data: an ensemble method. Information Sciences 163, 1-3 (2004), 85--101.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
GECCO '21: Proceedings of the Genetic and Evolutionary Computation Conference
June 2021
1219 pages
ISBN:9781450383509
DOI:10.1145/3449639
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 26 June 2021

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. diversity
  2. ensemble
  3. machine learning
  4. novelty search

Qualifiers

  • Research-article

Conference

GECCO '21
Sponsor:

Acceptance Rates

Overall Acceptance Rate 1,669 of 4,410 submissions, 38%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)5
  • Downloads (Last 6 weeks)0
Reflects downloads up to 06 Jan 2025

Other Metrics

Citations

Cited By

View all

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media