Export Citations
Save this search
Please login to be able to save your searches and receive alerts for new content matching your search criteria.
- research-articleJune 2009
On primal and dual sparsity of Markov networks
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1265–1272https://doi.org/10.1145/1553374.1553536Sparsity is a desirable property in high dimensional learning. The l1-norm regularization can lead to primal sparsity, while max-margin methods achieve dual sparsity. Combining these two methods, an l1-norm max-margin Markov network (l1-M3N) can achieve ...
- research-articleJune 2009
MedLDA: maximum margin supervised topic models for regression and classification
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1257–1264https://doi.org/10.1145/1553374.1553535Supervised topic models utilize document's side information for discovering predictive low dimensional representations of documents; and existing models apply likelihood-based estimation. In this paper, we present a max-margin supervised topic model for ...
- research-articleJune 2009
Multi-instance learning by treating instances as non-I.I.D. samples
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1249–1256https://doi.org/10.1145/1553374.1553534Previous studies on multi-instance learning typically treated instances in the bags as independently and identically distributed. The instances in a bag, however, are rarely independent in real tasks, and a better performance can be expected if the ...
- research-articleJune 2009
Learning non-redundant codebooks for classifying complex objects
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1241–1248https://doi.org/10.1145/1553374.1553533Codebook-based representations are widely employed in the classification of complex objects such as images and documents. Most previous codebook-based methods construct a single codebook via clustering that maps a bag of low-level features into a fixed-...
- research-articleJune 2009
Learning instance specific distances using metric propagation
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1225–1232https://doi.org/10.1145/1553374.1553530In many real-world applications, such as image retrieval, it would be natural to measure the distances from one instance to others using instance specific distance which captures the distinctions from the perspective of the concerned instance. However, ...
-
- research-articleJune 2009
Discovering options from example trajectories
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1217–1224https://doi.org/10.1145/1553374.1553529We present a novel technique for automated problem decomposition to address the problem of scalability in reinforcement learning. Our technique makes use of a set of near-optimal trajectories to discover options and incorporates them into the learning ...
- research-articleJune 2009
Compositional noisy-logical learning
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1209–1216https://doi.org/10.1145/1553374.1553528We describe a new method for learning the conditional probability distribution of a binary-valued variable from labelled training examples. Our proposed Compositional Noisy-Logical Learning (CNLL) approach learns a noisy-logical distribution in a ...
- research-articleJune 2009
Piecewise-stationary bandit problems with side observations
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1177–1184https://doi.org/10.1145/1553374.1553524We consider a sequential decision problem where the rewards are generated by a piecewise-stationary distribution. However, the different reward distributions are unknown and may change at unknown instants. Our approach uses a limited number of side ...
- research-articleJune 2009
Learning structural SVMs with latent variables
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1169–1176https://doi.org/10.1145/1553374.1553523We present a large-margin formulation and algorithm for structured output prediction that allows the use of latent variables. Our proposal covers a large range of application problems, with an optimization problem that can be solved efficiently using ...
- research-articleJune 2009
Stochastic search using the natural gradient
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1161–1168https://doi.org/10.1145/1553374.1553522To optimize unknown 'fitness' functions, we present Natural Evolution Strategies, a novel algorithm that constitutes a principled alternative to standard stochastic search methods. It maintains a multinormal distribution on the set of solution ...
- research-articleJune 2009
Non-monotonic feature selection
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1145–1152https://doi.org/10.1145/1553374.1553520We consider the problem of selecting a subset of m most informative features where m is the number of required features. This feature selection problem is essentially a combinatorial optimization problem, and is usually solved by an approximation. ...
- research-articleJune 2009
Optimal reverse prediction: a unified perspective on supervised, unsupervised and semi-supervised learning
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1137–1144https://doi.org/10.1145/1553374.1553519Training principles for unsupervised learning are often derived from motivations that appear to be independent of supervised learning. In this paper we present a simple unification of several supervised and unsupervised training principles through the ...
- research-articleJune 2009
Herding dynamical weights to learn
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1121–1128https://doi.org/10.1145/1553374.1553517A new "herding" algorithm is proposed which directly converts observed moments into a sequence of pseudo-samples. The pseudo-samples respect the moment constraints and may be used to estimate (unobserved) quantities of interest. The procedure allows us ...
- research-articleJune 2009
Evaluation methods for topic models
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1105–1112https://doi.org/10.1145/1553374.1553515A natural evaluation metric for statistical topic models is the probability of held-out documents given a trained model. While exact computation of this probability is intractable, several estimators for this probability have been used in the topic ...
- research-articleJune 2009
Model-free reinforcement learning as mixture learning
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1081–1088https://doi.org/10.1145/1553374.1553512We cast model-free reinforcement learning as the problem of maximizing the likelihood of a probabilistic mixture model via sampling, addressing both the infinite and finite horizon cases. We describe a Stochastic Approximation EM algorithm for ...
- research-articleJune 2009
Information theoretic measures for clusterings comparison: is a correction for chance necessary?
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1073–1080https://doi.org/10.1145/1553374.1553511Information theoretic based measures form a fundamental class of similarity measures for comparing clusterings, beside the class of pair-counting based and set-matching based measures. In this paper, we discuss the necessity of correction for chance for ...
- research-articleJune 2009
More generality in efficient multiple kernel learning
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1065–1072https://doi.org/10.1145/1553374.1553510Recent advances in Multiple Kernel Learning (MKL) have positioned it as an attractive tool for tackling many supervised learning tasks. The development of efficient gradient descent based optimization schemes has made it possible to tackle large scale ...
- research-articleJune 2009
Ranking with ordered weighted pairwise classification
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1057–1064https://doi.org/10.1145/1553374.1553509In ranking with the pairwise classification approach, the loss associated to a predicted ranked list is the mean of the pairwise classification losses. This loss is inadequate for tasks like information retrieval where we prefer ranked lists with high ...
- research-articleJune 2009
Factored conditional restricted Boltzmann Machines for modeling motion style
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1025–1032https://doi.org/10.1145/1553374.1553505The Conditional Restricted Boltzmann Machine (CRBM) is a recently proposed model for time series that has a rich, distributed hidden state and permits simple, exact inference. We present a new model, based on the CRBM that preserves its most important ...
- research-articleJune 2009
Kernelized value function approximation for reinforcement learning
ICML '09: Proceedings of the 26th Annual International Conference on Machine LearningPages 1017–1024https://doi.org/10.1145/1553374.1553504A recent surge in research in kernelized approaches to reinforcement learning has sought to bring the benefits of kernelized machine learning techniques to reinforcement learning. Kernelized reinforcement learning techniques are fairly new and different ...