|
Machine Learning
On the Convergence of Bound Optimization Algorithms
-
Many practitioners who use EM and related algorithms complain
that they are sometimes slow. When does this happen,
and what can be done about it? In this paper,
we study the general class of {\em bound optimization}
algorithms -- including EM,
Iterative Scaling, Non-negative Matrix Factorization, CCCP -- and
their relationship to direct optimization algorithms such as
gradient-based methods for parameter learning.
We derive a general relationship between the updates performed by
bound optimization methods and those of gradient and second-order methods
and identify analytic conditions under which
bound optimization algorithms exhibit quasi-Newton behavior, and
conditions under which they possess poor, first-order convergence.
Based on this analysis, we consider several specific
algorithms, interpret and analyze their convergence properties
and provide some recipes for preprocessing input to these algorithms
to yield faster convergence behavior.
We report empirical results supporting our analysis and showing that simple
data preprocessing can result in dramatically improved performance
of bound optimizers in practice.
- Here is a preprint of a paper, written with Sam Roweis
and Zoubin Ghahramani:
On the Convergence of Bound Optimization Algorithms
.
Direct Optimization in Latent Variable Models
-
We show a close relationship between the
Expectation - Maximization (EM) algorithm and
direct optimization algorithms such as
gradient-based methods for parameter learning.
We identify analytic conditions under which
EM exhibits Quasi-Newton behavior, and
conditions under which it possesses poor,
first-order convergence.
Based on this analysis, we propose
two novel algorithms for maximum
likelihood estimation of latent variable models, and report empirical
results showing that, as predicted by the theory,
the proposed new algorithms can substantially outperform
standard EM in terms of speed of convergence in certain cases.
- Here is a preprint of a paper, written with Sam Roweis
and Zoubin Ghahramani:
Optimization with EM and Expectation-Conjugate-Gradient
.
- Here is a poster presentation at the 12th Annual
Canadian Conference on Intelligent Systems, May 2002, Calgary,
Canada:
Is EM Really Better?
.
Adaptive Overrelaxed Bound Optimization Algorithms
-
We study a class of overrelaxed bound optimization algorithms,
and their relationship to standard
bound optimizers, such as Expectation-Maximization,
Iterative Scaling, CCCP and Non-Negative Matrix Factorization.
We provide a theoretical analysis of the convergence properties of
these optimizers and identify analytic conditions under which
they are expected to outperform the standard versions.
Based on this analysis, we propose a novel, simple adaptive overrelaxed
scheme for practical optimization and report empirical results on
several synthetic and real-world data sets showing that these new
adaptive methods exhibit superior performance
(in certain cases by several orders of magnitude)
compared to their traditional counterparts. Our ``drop-in'' extensions
are simple to implement, apply to a wide variety of algorithms, almost
always give a substantial speedup, and do not require any theoretical
analysis of the underlying algorithm.
- Here is a preprint of a paper, written with Sam Roweis:
Adaptive Overrelaxed Bound Optimization Methods
.
Unsupervised Text Segmentation with Constrained HMMs
-
Here I am working with Sam Roweis
on a model that learns to segment
(or group) relevant sequential data together
in completely unsupervised fasion. The model
possesses hierarchical structure, but is nevertheless
extremely efficient in its greedy way of performing
training as well as inference. We also report
preliminary empirical results that suggest that this
approach to the task of text segmentation can be promising.
This project is in progress.
- Here is a poster presentation I gave at our
ML group meeting entitled
An Efficient Greedy Way to Perform Unsupervised Text Segmentation.
.
[
Home |
Information |
Research |
Teaching |
Background
]
Ruslan Salakhutdinov, Artificial Intelligence Group, Toronto
Computer Science || www.cs.toronto.edu/~rsalakhu
|