Export Citations
Save this search
Please login to be able to save your searches and receive alerts for new content matching your search criteria.
- research-articleJuly 2008
Efficient multiclass maximum margin clustering
ICML '08: Proceedings of the 25th international conference on Machine learningPages 1248–1255https://doi.org/10.1145/1390156.1390313This paper presents a cutting plane algorithm for multiclass maximum margin clustering (MMC). The proposed algorithm constructs a nested sequence of successively tighter relaxations of the original MMC problem, and each optimization problem in this ...
- research-articleJuly 2008
Estimating local optimums in EM algorithm over Gaussian mixture model
ICML '08: Proceedings of the 25th international conference on Machine learningPages 1240–1247https://doi.org/10.1145/1390156.1390312EM algorithm is a very popular iteration-based method to estimate the parameters of Gaussian Mixture Model from a large observation set. However, in most cases, EM algorithm is not guaranteed to converge to the global optimum. Instead, it stops at some ...
- research-articleJuly 2008
Improved Nyström low-rank approximation and error analysis
ICML '08: Proceedings of the 25th international conference on Machine learningPages 1232–1239https://doi.org/10.1145/1390156.1390311Low-rank matrix approximation is an effective tool in alleviating the memory and computational burdens of kernel methods and sampling, as the mainstream of such algorithms, has drawn considerable attention in both theory and practice. This paper ...
- research-articleJuly 2008
A quasi-Newton approach to non-smooth convex optimization
ICML '08: Proceedings of the 25th international conference on Machine learningPages 1216–1223https://doi.org/10.1145/1390156.1390309We extend the well-known BFGS quasi-Newton method and its limited-memory variant LBFGS to the optimization of non-smooth convex objectives. This is done in a rigorous fashion by generalizing three components of BFGS to subdifferentials: The local ...
- research-articleJuly 2008
Preconditioned temporal difference learning
ICML '08: Proceedings of the 25th international conference on Machine learningPages 1208–1215https://doi.org/10.1145/1390156.1390308This paper extends many of the recent popular policy evaluation algorithms to a generalized framework that includes least-squares temporal difference (LSTD) learning, least-squares policy evaluation (LSPE) and a variant of incremental LSTD (iLSTD). The ...
-
- research-articleJuly 2008
Democratic approximation of lexicographic preference models
ICML '08: Proceedings of the 25th international conference on Machine learningPages 1200–1207https://doi.org/10.1145/1390156.1390307Previous algorithms for learning lexicographic preference models (LPMs) produce a "best guess" LPM that is consistent with the observations. Our approach is more democratic: we do not commit to a single LPM. Instead, we approximate the target using the ...
- research-articleJuly 2008
Efficiently learning linear-linear exponential family predictive representations of state
ICML '08: Proceedings of the 25th international conference on Machine learningPages 1176–1183https://doi.org/10.1145/1390156.1390304Exponential Family PSR (EFPSR) models capture stochastic dynamical systems by representing state as the parameters of an exponential family distribution over a shortterm window of future observations. They are appealing from a learning perspective ...
- research-articleJuly 2008
Deep learning via semi-supervised embedding
ICML '08: Proceedings of the 25th international conference on Machine learningPages 1168–1175https://doi.org/10.1145/1390156.1390303We show how nonlinear embedding algorithms popular for use with shallow semi-supervised learning techniques such as kernel methods can be applied to deep multilayer architectures, either as a regularizer at the output layer, or on each layer of the ...
- research-articleJuly 2008
Graph transduction via alternating minimization
ICML '08: Proceedings of the 25th international conference on Machine learningPages 1144–1151https://doi.org/10.1145/1390156.1390300Graph transduction methods label input data by learning a classification function that is regularized to exhibit smoothness along a graph over labeled and unlabeled samples. In practice, these algorithms are sensitive to the initial set of labels ...
- research-articleJuly 2008
Sparse multiscale gaussian process regression
ICML '08: Proceedings of the 25th international conference on Machine learningPages 1112–1119https://doi.org/10.1145/1390156.1390296Most existing sparse Gaussian process (g.p.) models seek computational advantages by basing their computations on a set of m basis functions that are the covariance function of the g.p. with one of its two inputs fixed. We generalise this for the case ...
- research-articleJuly 2008
A semiparametric statistical approach to model-free policy evaluation
ICML '08: Proceedings of the 25th international conference on Machine learningPages 1072–1079https://doi.org/10.1145/1390156.1390291Reinforcement learning (RL) methods based on least-squares temporal difference (LSTD) have been developed recently and have shown good practical performance. However, the quality of their estimation has not been well elucidated. In this article, we ...
- research-articleJuly 2008
Training restricted Boltzmann machines using approximations to the likelihood gradient
ICML '08: Proceedings of the 25th international conference on Machine learningPages 1064–1071https://doi.org/10.1145/1390156.1390290A new algorithm for training Restricted Boltzmann Machines is introduced. The algorithm, named Persistent Contrastive Divergence, is different from the standard Contrastive Divergence algorithms in that it aims to draw samples from almost exactly the ...
- research-articleJuly 2008
ν-support vector machine as conditional value-at-risk minimization
ICML '08: Proceedings of the 25th international conference on Machine learningPages 1056–1063https://doi.org/10.1145/1390156.1390289The ν-support vector classification (ν-SVC) algorithm was shown to work well and provide intuitive interpretations, e.g., the parameter ν roughly specifies the fraction of support vectors. Although ν corresponds to a fraction, it cannot take the entire ...
- research-articleJuly 2008
The many faces of optimism: a unifying approach
ICML '08: Proceedings of the 25th international conference on Machine learningPages 1048–1055https://doi.org/10.1145/1390156.1390288The exploration-exploitation dilemma has been an intriguing and unsolved problem within the framework of reinforcement learning. "Optimism in the face of uncertainty" and model building play central roles in advanced exploration methods. Here, we ...
- research-articleJuly 2008
Apprenticeship learning using linear programming
ICML '08: Proceedings of the 25th international conference on Machine learningPages 1032–1039https://doi.org/10.1145/1390156.1390286In apprenticeship learning, the goal is to learn a policy in a Markov decision process that is at least as good as a policy demonstrated by an expert. The difficulty arises in that the MDP's true reward function is assumed to be unknown. We show how to ...
- research-articleJuly 2008
A least squares formulation for canonical correlation analysis
ICML '08: Proceedings of the 25th international conference on Machine learningPages 1024–1031https://doi.org/10.1145/1390156.1390285Canonical Correlation Analysis (CCA) is a well-known technique for finding the correlations between two sets of multi-dimensional variables. It projects both sets of variables into a lower-dimensional space in which they are maximally correlated. CCA is ...
- research-articleJuly 2008
The asymptotics of semi-supervised learning in discriminative probabilistic models
ICML '08: Proceedings of the 25th international conference on Machine learningPages 984–991https://doi.org/10.1145/1390156.1390280Semi-supervised learning aims at taking advantage of unlabeled data to improve the efficiency of supervised learning procedures. For discriminative models however, this is a challenging task. In this contribution, we introduce an original methodology ...
- research-articleJuly 2008
Sample-based learning and search with permanent and transient memories
ICML '08: Proceedings of the 25th international conference on Machine learningPages 968–975https://doi.org/10.1145/1390156.1390278We present a reinforcement learning architecture, Dyna-2, that encompasses both sample-based learning and sample-based search, and that generalises across states during both learning and search. We apply Dyna-2 to high performance Computer Go. In this ...
- research-articleJuly 2008
Data spectroscopy: learning mixture models using eigenspaces of convolution operators
ICML '08: Proceedings of the 25th international conference on Machine learningPages 936–943https://doi.org/10.1145/1390156.1390274In this paper we develop a spectral framework for estimating mixture distributions, specifically Gaussian mixture models. In physics, spectroscopy is often used for the identification of substances through their spectrum. Treating a kernel function K(x, ...
- research-articleJuly 2008
SVM optimization: inverse dependence on training set size
ICML '08: Proceedings of the 25th international conference on Machine learningPages 928–935https://doi.org/10.1145/1390156.1390273We discuss how the runtime of SVM optimization should decrease as the size of the training data increases. We present theoretical and empirical results demonstrating how a simple subgradient descent approach indeed displays such behavior, at least for ...