Research Interests
I completed my Ph.D. working with
Craig Boutilier in 2013. The focus of my doctoral work was using preference elicitation techniques to aid sequential decision making.
My main interests lie in the intersection between preference elicitation, reasoning under uncertainty and mechanism design. I have also done work with multi-agent systems, reputation and trust models, and electronic markets.
I am currently working at Google.
Select Publications
IJCAI 11
Robust Online Optimization of Reward-uncertain MDPs
Kevin Regan and Craig Boutilier
Twenty-Second Joint Conference on Artificial Intelligence
Imprecise-reward Markov decision processes (IRMDPs) are MDPs in which the
reward function is only partially specified (e.g., by some elicitation process).
Recent work using minimax regret to solve IRMDPs has shown, despite theoretical
intractability, how the set of policies that are nondom- inated w.r.t. reward
uncertainty can be exploited to accelerate regret computation. However,
the number of nondominated policies is generally so large as to undermine this
leverage. In this paper, we show how the quality of the approximation can be
improved online by pruning/adding nondominated policies during reward
elicitation, while maintain- ing computational tractability. Drawing insights
from the POMDP literature, we also develop a new anytime algorithm for
constructing the set of nondominated policies with provable (anytime)
error bounds.
@article{ijcai2011:rb:a,
author = {Kevin Regan and Craig Boutilier},
title = {
Robust Online Optimization of Reward-uncertain MDPs
},
journal = {
Twenty-Second Joint Conference on Artificial Intelligence (IJCAI 2011)
},
year = {2011}
}
IJCAI 11
Eliciting Additive Reward Functions for Markov Decision Processes
Kevin Regan and Craig Boutilier
Twenty-Second Joint Conference on Artificial Intelligence
Specifying the reward function of a Markov decision process (MDP) can be
demanding, requiring human assessment of the precise quality of, and tradeoffs
among, various states and actions. However, reward functions often possess
considerable structure which can be leveraged to streamline their
specification. We develop new, decision-theoretically sound heuristics for
eliciting rewards for factored MDPs whose reward functions exhibit additive
independence. Since we can often find good policies without complete reward
specification, we also develop new (exact and approximate) algorithms for
robust optimization of imprecise-reward MDPs with such additive reward. Our
methods are evaluated in two domains: autonomic computing and assistive
technology.
@article{ijcai2011:rb:b,
author = {Kevin Regan and Craig Boutilier},
title = {
Eliciting Additive Reward Functions for Markov Decision Processes
},
journal = {
Twenty-Second Joint Conference on Artificial Intelligence (IJCAI 2011)
},
year = {2011}
}
AAAI 10
Robust Policy Computation in Reward-uncertain MDPs using Nondominated Policies
Kevin Regan and Craig Boutilier
Twenty-Fourth AAAI Conference on Artificial Intelligence
The precise specification of reward functions for
Markov decision processes (MDPs)
is often extremely difficult, motivating research into both
reward elicitation and the robust solution of MDPs with imprecisely
specified reward (IRMDPs). We develop new techniques for the
robust optimization of IRMDPs, using the minimax regret
decision criterion, that exploit the set of nondominated policies,
i.e., policies that are optimal for some instantiation of the
imprecise reward function. Drawing parallels to POMDP value functions,
we devise a Witness-style algorithm for identifying
nondominated policies. We also examine several new algorithms for
computing minimax regret using the nondominated set, and examine
both practically and theoretically the impact of approximating
this set. Our results suggest that a small subset of the nondominated
set can greatly speed up computation, yet yield very tight approximations
to minimax regret.
@article{aaai2010rb,
author = {Kevin Regan and Craig Boutilier},
title = {
Robust Policy Computation in Reward-uncertain MDPs using Nondominated Policies
},
journal = {
Twenty-Fourth AAAI Conference on Artificial Intelligence (AAAI 2010)
},
year = {2010}
}
ICML 09
Online Feature Elicitation in Interactive Optimization
Craig Boutilier, Kevin Regan and Paolo Viappiani
International Conference on Machine Learning
Most models of utility elicitation in decision support
and interactive optimization assume a predefined set of
"catalog" features over which user
preferences are expressed. However, users may
differ in the features over which they are most
comfortable expressing their preferences. In this
work we consider the problem of feature elicitation:
a user's utility function is expressed using
features whose definitions (in terms of "catalog"
features) are unknown. We cast this as a
problem of concept learning, but whose goal is
to identify only enough about the concept to
enable a good decision to be recommended. We
describe computational procedures for identifying
optimal alternatives w.r.t. minimax regret in
the presence of concept uncertainty; and describe
several heuristic query strategies that focus on
reduction of relevant concept uncertainty.
@article{icml09brv,
author = {Craig Boutilier, Kevin Regan and Paolo Viappiani},
title = {
Online Feature Elicitation in Interactive Optimization
},
journal = {
ICML-09 Iternational Conference on Machine Learning
},
year = {2009}
}
Select Awards
- 2010 ⋅ Ontario Graduate Scholarship
- 2006 ⋅ NSERC Canadian Graduate Scholarship (2006 - 2009)
- 2006 ⋅ Award for Outstanding Achievement in Graduate Studies
- 2005 ⋅ Best Student Paper - Privacy Security & Trust 05