skip to main content
10.1145/1835804.1835947acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article

Semi-supervised sparse metric learning using alternating linearization optimization

Published: 25 July 2010 Publication History

Abstract

In plenty of scenarios, data can be represented as vectors and then mathematically abstracted as points in a Euclidean space. Because a great number of machine learning and data mining applications need proximity measures over data, a simple and universal distance metric is desirable, and metric learning methods have been explored to produce sensible distance measures consistent with data relationship. However, most existing methods suffer from limited labeled data and expensive training. In this paper, we address these two issues through employing abundant unlabeled data and pursuing sparsity of metrics, resulting in a novel metric learning approach called semi-supervised sparse metric learning. Two important contributions of our approach are: 1) it propagates scarce prior affinities between data to the global scope and incorporates the full affinities into the metric learning; and 2) it uses an efficient alternating linearization method to directly optimize the sparse metric. Compared with conventional methods, ours can effectively take advantage of semi-supervision and automatically discover the sparse metric structure underlying input data patterns. We demonstrate the efficacy of the proposed approach with extensive experiments carried out on six datasets, obtaining clear performance gains over the state-of-the-arts.

Supplementary Material

JPG File (kdd2010_liu_sss_01.jpg)
MOV File (kdd2010_liu_sss_01.mov)

References

[1]
O. Banerjee, L. E. Ghaoui, and A. d'Aspremont. Model selection through sparse maximum likelihood estimation for multivariate gaussian or binary data. Journal of Machine Learning Research, 9:485--516, 2008.
[2]
A. Bar-Hillel, T. Hertz, N. Shental, and D. Weinshall. Learning a mahalanobis metric from equivalence constraints. JMLR, 6:937--965, 2005.
[3]
S. Basu, M. Bilenko, and R. Mooney. A probabilistic framework for semi-supervised clustering. In Proc. KDD, 2004.
[4]
M. Bilenko, S. Basu, and R. Mooney. Integrating constraints and metric learning in semi-supervised clustering. In Proc. ICML, 2004.
[5]
S. Boyd and L. Vandenberge. Convex Optimization. Cambridge University Press, Cambridge, UK, 2003.
[6]
T.-S. Chua, J. Tang, R. Hong, H. Li, Z. Luo, and Y. Zheng. Nus-wide: A real-world web image database from national university of singapore. In Proc. CIVR, 2009.
[7]
J. V. Davis and I. S. Dhillon. Structured metric learning for high dimensional problems. In Proc. KDD, 2008.
[8]
J. V. Davis, B. Kulis, P. Jain, S. Sra, and I. S. Dhillon. Information-theoretic metric learning. In Proc. ICML, 2007.
[9]
J. Friedman, T. Hastie, and R. Tibshirani. Sparse inverse covariance estimation with the graphical lasso. Biostatistics, 8(1):1--10, 2007.
[10]
A. Globerson and S. Roweis. Metric learning by collapsing classes. In NIPS 18, 2006.
[11]
J. Goldberger, S. Roweis, and R. Salakhutdinov. Neighbourhood components analysis. In NIPS 17, 2005.
[12]
D. Goldfarb and S. Ma. Fast alternating linearization methods for minimizing the sum of two convex functions. Technical report, Department of IEOR, Columbia University, 2009. Preprint available at http://arxiv.org/abs/0912.4571.
[13]
J.-B. Hiriart-Urruty and C. Lemaréchal. Convex Analysis and Minimization Algorithms II: Advanced Theory and Bundle Methods. Springer-Verlag, New York, 1993.
[14]
S. C. Hoi, W. Liu, and S.-F. Chang. Semi-supervised distance metric learning for collaborative image retrieval. In Proc. CVPR, 2008.
[15]
B. Kulis, S. Basu, I. Dhillon, and R. Mooney. Semi-supervised graph clustering: A kernel approach. Machine Learning, 74(1):1--22, 2009.
[16]
W. Liu, B. Qian, J. Cui, and J. Liu. Spectral kernel learning for semi-supervised classification. In Proc. IJCAI, 2009.
[17]
W. Liu, D. Tao, and J. Liu. Transductive component analysis. In Proc. IEEE ICDM, 2008.
[18]
W. Liu, X. Tian, D. Tao, and J. Liu. Constrained metric learning via distance gap maximization. In Proc. AAAI, 2010.
[19]
Z. Lu. Smooth optimization approach for sparse covariance selection. SIAM J. Optim., 19(4):1807--1827, 2009.
[20]
Y. E. Nesterov. Smooth minimization for non-smooth functions. Math. Program. Ser. A, 103:127--152, 2005.
[21]
G.-J. Qi, J. Tang, Z.-J. Zha, T.-S. Chua, and H.-J. Zhang. An efficient sparse metric learning in high-dimensional space via l1-penalized log-determinant regularization. In Proc. ICML, 2009.
[22]
R. Rosales and G. Fung. Learning sparse metrics via linear programming. In Proc. KDD, 2006.
[23]
Y. Rui, T. S. Huang, M. Ortega, and S. Mehrotra. Relevance feedback: A powerful tool in interactive content-based image retrieval. IEEE Trans. on Circuits and Systems for Video Technology, 8(5):644--655, 1998.
[24]
A. W. M. Smeulders, M. Worring, S. Santini, A. Gupta, and R. Jain. Content-based image retrieval at the end of the early years. IEEE Trans. PAMI, 22(12):1349--1380, 2000.
[25]
W. Tang, H. Xiong, S. Zhong, and J. Wu. Enhancing semi-supervised clustering: A feature projection perspective. In Proc. KDD, 2007.
[26]
K. Wagstaff, C. Cardie, S. Rogers, and S. Schroedl. Constrained k-means clustering with background knowledge. In Proc. ICML, 2001.
[27]
M. Wang, L. Yang, and X.-S. Hua. Msra-mm: Bridging research and industrial societies for multimedia information retrieval. Technical report, Miscosoft Research Asia, 2009. No. MSR-TR-2009-30.
[28]
K. Q. Weinberger and L. K. Saul. Distance metric learning for large margin nearest neighbor classification. Journal of Machine Learning Research, 10:207--244, 2009.
[29]
E. P. Xing, A. Y. Ng, M. I. Jordan, and S. Russell. Distance metric learning with application to clustering with side-information. In NIPS 15, 2003.
[30]
Y. Ying, K. Huang, and C. Campbell. Sparse metric learning via smooth optimization. In NIPS 22, 2010.
[31]
X. Zhu. Semi-supervied learning literature survey. Technical report, University of Wisconsin Madison, 2008.

Cited By

View all

Index Terms

  1. Semi-supervised sparse metric learning using alternating linearization optimization

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      KDD '10: Proceedings of the 16th ACM SIGKDD international conference on Knowledge discovery and data mining
      July 2010
      1240 pages
      ISBN:9781450300551
      DOI:10.1145/1835804
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 25 July 2010

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. alternating linearization
      2. metric learning
      3. semi-supervised sparse metric learning
      4. sparse inverse covariance estimation

      Qualifiers

      • Research-article

      Conference

      KDD '10
      Sponsor:

      Acceptance Rates

      Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

      Upcoming Conference

      KDD '25

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)26
      • Downloads (Last 6 weeks)2
      Reflects downloads up to 29 Jan 2025

      Other Metrics

      Citations

      Cited By

      View all

      View Options

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Figures

      Tables

      Media

      Share

      Share

      Share this Publication link

      Share on social media