skip to main content
10.1145/3463944.3469098acmconferencesArticle/Chapter ViewAbstractPublication PagesicdarConference Proceedingsconference-collections
research-article

Scattering Transform Based Image Clustering using Projection onto Orthogonal Complement

Published: 21 August 2021 Publication History

Abstract

In the last few years, large improvements in image clustering have been driven by the recent advances in deep learning. However, due to the architectural complexity of deep neural networks, there is no mathematical theory that explains the success of deep clustering techniques. In this work we introduce Projected-Scattering Spectral Clustering (PSSC), a state-of-the-art, stable, and fast algorithm for image clustering, which is also mathematically interpretable. PSSC includes a novel method to exploit the geometric structure of the scattering transform of small images. This method is inspired by the observation that, in the scattering transform domain, the subspaces formed by the eigenvectors corresponding to the few largest eigenvalues of the data matrices of individual classes are nearly shared among different classes. Therefore, projecting out those shared subspaces reduces the intra-class variability, substantially increasing the clustering performance. We call this method 'Projection onto Orthogonal Complement' (POC). Our experiments demonstrate that PSSC obtains the best results among all shallow clustering algorithms. Moreover, it achieves comparable clustering performance to that of recent state-of-the-art clustering techniques, while reducing the execution time by more than one order of magnitude.

Supplementary Material

MP4 File (ICDAR21-icdar08.mp4)
We present PSSC, a novel image clustering method that groups similar images based on their processed scattering representations. PSSC outperforms all other shallow clustering algorithms, and it achieves comparable clustering performance to that of recent state-of-the-art clustering techniques, while reducing the execution time by more than one order of magnitude.

References

[1]
Joakim Andén and Stéphane Mallat. 2014. Deep scattering spectrum. IEEE Transactions on Signal Processing, Vol. 62, 16 (2014), 4114--4128.
[2]
M. Andreux, T. Angles, G. Exarchakis, R. Leonarduzzi, G. Rochette, L. Thiry, J. Zarka, S. Mallat, J. Andén, E. Belilovsky, J. Bruna, V. Lostanlen, M. J. Hirn, E. Oyallon, S. Zhang, C. Cella, and M. Eickenberg. 2018. Kymatio: Scattering transforms in python. CoRR, Vol. 1812.11214 (2018).
[3]
Ake Bjorck and Gene H Golub. 1973. Numerical methods for computing angles between linear subspaces. Mathematics of computation, Vol. 27, 123 (1973), 579--594.
[4]
Joan Bruna and Stephane Mallat. 2012. Invariant Scattering Convolutional Networks. (2012).
[5]
Yixin Chen, James Z Wang, and Robert Krovetz. 2003. Content-based image retrieval by clustering. In Proceedings of the 5th ACM SIGMM international workshop on Multimedia information retrieval. 193--200.
[6]
Dorin Comaniciu and Peter Meer. 2002. Mean shift: A robust approach toward feature space analysis. IEEE Transactions on pattern analysis and machine intelligence, Vol. 24, 5 (2002), 603--619.
[7]
Nameirakpam Dhanachandra, Khumanthem Manglem, and Yambem Jina Chanu. 2015. Image segmentation using K-means clustering algorithm and subtractive clustering algorithm. Procedia Computer Science, Vol. 54 (2015), 764--771.
[8]
Jeff Donahue, Philipp Kr"ahenbühl, and Trevor Darrell. 2016. Adversarial feature learning. arXiv preprint arXiv:1605.09782 (2016).
[9]
Martin Ester, Hans-Peter Kriegel, Jörg Sander, Xiaowei Xu, et al. 1996. A density-based algorithm for discovering clusters in large spatial databases with noise. In Kdd, Vol. 96. 226--231.
[10]
Kamran Ghasedi Dizaji, Amirhossein Herandi, Cheng Deng, Weidong Cai, and Heng Huang. 2017. Deep clustering via joint convolutional autoencoder embedding and relative entropy minimization. In Proceedings of the IEEE international conference on computer vision. 5736--5745.
[11]
Xifeng Guo, Long Gao, Xinwang Liu, and Jianping Yin. 2017a. Improved deep embedded clustering with local structure preservation. In IJCAI. 1753--1759.
[12]
Xifeng Guo, Xinwang Liu, En Zhu, and Jianping Yin. 2017b. Deep clustering with convolutional autoencoders. In International conference on neural information processing. Springer, 373--382.
[13]
Charles R Harris, K Jarrod Millman, Stéfan J van der Walt, Ralf Gommers, Pauli Virtanen, David Cournapeau, Eric Wieser, Julian Taylor, Sebastian Berg, Nathaniel J Smith, et al. 2020. Array programming with NumPy. Nature, Vol. 585, 7825 (2020), 357--362.
[14]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition. 770--778.
[15]
Li He, Nilanjan Ray, Yisheng Guan, and Hong Zhang. 2018. Fast large-scale spectral clustering via explicit feature mapping. IEEE transactions on cybernetics, Vol. 49, 3 (2018), 1058--1071.
[16]
Dong Huang, Chang-Dong Wang, Jian-Sheng Wu, Jian-Huang Lai, and Chee-Keong Kwoh. 2019. Ultra-scalable spectral clustering and ensemble clustering. IEEE Transactions on Knowledge and Data Engineering, Vol. 32, 6 (2019), 1212--1226.
[17]
Jiaji Huang, Qiang Qiu, and Robert Calderbank. 2015. The role of principal angles in subspace classification. IEEE Transactions on Signal Processing, Vol. 64, 8 (2015), 1933--1945.
[18]
Jonathan J. Hull. 1994. A database for handwritten text recognition research. IEEE Transactions on pattern analysis and machine intelligence, Vol. 16, 5 (1994), 550--554.
[19]
Stephen C Johnson. 1967. Hierarchical clustering schemes. Psychometrika, Vol. 32, 3 (1967), 241--254.
[20]
Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. 2012. Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems. 1097--1105.
[21]
Yann LeCun, Corinna Cortes, and Christopher JC Burges. 1998. The MNIST database of handwritten digits, 1998. URL http://yann. lecun. com/exdb/mnist, Vol. 10 (1998), 34.
[22]
Yann LeCun, Koray Kavukcuoglu, and Clément Farabet. 2010. Convolutional networks and applications in vision. In Proceedings of 2010 IEEE international symposium on circuits and systems. IEEE, 253--256.
[23]
Zhenguo Li, Xiao-Ming Wu, and Shih-Fu Chang. 2012. Segmentation using superpixels: A bipartite graph partitioning approach. In 2012 IEEE Conference on Computer Vision and Pattern Recognition. IEEE, 789--796.
[24]
Ulrike Von Luxburg. 2007. A Tutorial on Spectral Clustering.
[25]
James MacQueen et al. 1967. Some methods for classification and analysis of multivariate observations. In Proceedings of the fifth Berkeley symposium on mathematical statistics and probability, Vol. 1. Oakland, CA, USA, 281--297.
[26]
Yury A Malkov and Dmitry A Yashunin. 2018. Efficient and robust approximate nearest neighbor search using hierarchical navigable small world graphs. IEEE transactions on pattern analysis and machine intelligence (2018).
[27]
Stéphane Mallat. 2012. Group invariant scattering. Communications on Pure and Applied Mathematics, Vol. 65, 10 (2012), 1331--1398.
[28]
Stéphane Mallat. 2016. Understanding deep convolutional networks. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences, Vol. 374, 2065 (2016), 20150203.
[29]
Ryan McConville, Raul Santos-Rodriguez, Robert J Piechocki, and Ian Craddock. 2019. N2d:(not too) deep clustering via clustering the local manifold of an autoencoded embedding. arXiv preprint arXiv:1908.05968 (2019).
[30]
Nairouz Mrabah, Mohamed Bouguessa, and Riadh Ksantini. 2020. Adversarial deep embedded clustering: on a better trade-off between feature randomness and feature drift. IEEE Transactions on Knowledge and Data Engineering (2020).
[31]
VSVS Murthy, E Vamsidhar, JNVR Swarup Kumar, P Sankara Rao, et al. 2010. Content based image retrieval using Hierarchical and K-means clustering techniques. International Journal of Engineering Science and Technology, Vol. 2, 3 (2010), 209--212.
[32]
Andrew Y Ng, Michael I Jordan, and Yair Weiss. 2002. On spectral clustering: Analysis and an algorithm. In Advances in neural information processing systems. 849--856.
[33]
Edouard Oyallon, Sergey Zagoruyko, Gabriel Huang, Nikos Komodakis, Simon Lacoste-Julien, Matthew Blaschko, and Eugene Belilovsky. 2018. Scattering networks for hybrid representation learning. IEEE transactions on pattern analysis and machine intelligence, Vol. 41, 9 (2018), 2208--2221.
[34]
Fabian Pedregosa, Gaël Varoquaux, Alexandre Gramfort, Vincent Michel, Bertrand Thirion, Olivier Grisel, Mathieu Blondel, Peter Prettenhofer, Ron Weiss, Vincent Dubourg, et al. 2011. Scikit-learn: Machine learning in Python. the Journal of machine Learning research, Vol. 12 (2011), 2825--2830.
[35]
Yazhou Ren, Ni Wang, Mingxia Li, and Zenglin Xu. 2020. Deep density-based image clustering. Knowledge-Based Systems (2020), 105841.
[36]
Douglas A Reynolds. 2009. Gaussian Mixture Models. Encyclopedia of biometrics, Vol. 741 (2009).
[37]
Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus. 2013. Intriguing properties of neural networks. arXiv preprint arXiv:1312.6199 (2013).
[38]
Manish Verma, Mauly Srivastava, Neha Chack, Atul Kumar Diswar, and Nidhi Gupta. 2012. A comparative study of various clustering algorithms in data mining. International Journal of Engineering Research and Applications (IJERA), Vol. 2, 3 (2012), 1379--1384.
[39]
Nguyen Xuan Vinh, Julien Epps, and James Bailey. 2010. Information theoretic measures for clusterings comparison: Variants, properties, normalization and correction for chance. The Journal of Machine Learning Research, Vol. 11 (2010), 2837--2854.
[40]
Han Xiao, Kashif Rasul, and Roland Vollgraf. 2017. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747 (2017).
[41]
Junyuan Xie, Ross Girshick, and Ali Farhadi. 2016. Unsupervised deep embedding for clustering analysis. In International conference on machine learning. 478--487.
[42]
Bo Yang, Xiao Fu, Nicholas D Sidiropoulos, and Mingyi Hong. 2017. Towards k-means-friendly spaces: Simultaneous deep learning and clustering. In international conference on machine learning. PMLR, 3861--3870.
[43]
Jianwei Yang, Devi Parikh, and Dhruv Batra. 2016. Joint unsupervised learning of deep representations and image clusters. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 5147--5156.
[44]
John Zarka, Louis Thiry, Tomás Angles, and Stéphane Mallat. 2019. Deep network classification by scattering and homotopy dictionary learning. arXiv preprint arXiv:1910.03561 (2019).

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
ICDAR '21: Proceedings of the 2021 ACM Workshop on Intelligent Cross-Data Analysis and Retrieval
August 2021
72 pages
ISBN:9781450385299
DOI:10.1145/3463944
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 21 August 2021

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. image clustering
  2. orthogonal projection
  3. representation learning
  4. scattering transform
  5. spectral clustering

Qualifiers

  • Research-article

Conference

ICMR '21
Sponsor:

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)9
  • Downloads (Last 6 weeks)4
Reflects downloads up to 09 Jan 2025

Other Metrics

Citations

Cited By

View all

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media