photo
Research Interests
I completed my Ph.D. working with Craig Boutilier in 2013. The focus of my doctoral work was using preference elicitation techniques to aid sequential decision making. My main interests lie in the intersection between preference elicitation, reasoning under uncertainty and mechanism design. I have also done work with multi-agent systems, reputation and trust models, and electronic markets.
I am currently working at Google.
Select Publications
IJCAI 11
Robust Online Optimization of Reward-uncertain MDPs
Kevin Regan and Craig Boutilier
Twenty-Second Joint Conference on Artificial Intelligence
Imprecise-reward Markov decision processes (IRMDPs) are MDPs in which the reward function is only partially specified (e.g., by some elicitation process). Recent work using minimax regret to solve IRMDPs has shown, despite theoretical intractability, how the set of policies that are nondom- inated w.r.t. reward uncertainty can be exploited to accelerate regret computation. However, the number of nondominated policies is generally so large as to undermine this leverage. In this paper, we show how the quality of the approximation can be improved online by pruning/adding nondominated policies during reward elicitation, while maintain- ing computational tractability. Drawing insights from the POMDP literature, we also develop a new anytime algorithm for constructing the set of nondominated policies with provable (anytime) error bounds.
@article{ijcai2011:rb:a, author = {Kevin Regan and Craig Boutilier}, title = { Robust Online Optimization of Reward-uncertain MDPs }, journal = { Twenty-Second Joint Conference on Artificial Intelligence (IJCAI 2011) }, year = {2011} }
IJCAI 11
Eliciting Additive Reward Functions for Markov Decision Processes
Kevin Regan and Craig Boutilier
Twenty-Second Joint Conference on Artificial Intelligence
Specifying the reward function of a Markov decision process (MDP) can be demanding, requiring human assessment of the precise quality of, and tradeoffs among, various states and actions. However, reward functions often possess considerable structure which can be leveraged to streamline their specification. We develop new, decision-theoretically sound heuristics for eliciting rewards for factored MDPs whose reward functions exhibit additive independence. Since we can often find good policies without complete reward specification, we also develop new (exact and approximate) algorithms for robust optimization of imprecise-reward MDPs with such additive reward. Our methods are evaluated in two domains: autonomic computing and assistive technology.
@article{ijcai2011:rb:b, author = {Kevin Regan and Craig Boutilier}, title = { Eliciting Additive Reward Functions for Markov Decision Processes }, journal = { Twenty-Second Joint Conference on Artificial Intelligence (IJCAI 2011) }, year = {2011} }
AAAI 10
Robust Policy Computation in Reward-uncertain MDPs using Nondominated Policies
Kevin Regan and Craig Boutilier
Twenty-Fourth AAAI Conference on Artificial Intelligence
The precise specification of reward functions for Markov decision processes (MDPs) is often extremely difficult, motivating research into both reward elicitation and the robust solution of MDPs with imprecisely specified reward (IRMDPs). We develop new techniques for the robust optimization of IRMDPs, using the minimax regret decision criterion, that exploit the set of nondominated policies, i.e., policies that are optimal for some instantiation of the imprecise reward function. Drawing parallels to POMDP value functions, we devise a Witness-style algorithm for identifying nondominated policies. We also examine several new algorithms for computing minimax regret using the nondominated set, and examine both practically and theoretically the impact of approximating this set. Our results suggest that a small subset of the nondominated set can greatly speed up computation, yet yield very tight approximations to minimax regret.
@article{aaai2010rb, author = {Kevin Regan and Craig Boutilier}, title = { Robust Policy Computation in Reward-uncertain MDPs using Nondominated Policies }, journal = { Twenty-Fourth AAAI Conference on Artificial Intelligence (AAAI 2010) }, year = {2010} }
ICML 09
Online Feature Elicitation in Interactive Optimization
Craig Boutilier, Kevin Regan and Paolo Viappiani
International Conference on Machine Learning
Most models of utility elicitation in decision support and interactive optimization assume a predefined set of "catalog" features over which user preferences are expressed. However, users may differ in the features over which they are most comfortable expressing their preferences. In this work we consider the problem of feature elicitation: a user's utility function is expressed using features whose definitions (in terms of "catalog" features) are unknown. We cast this as a problem of concept learning, but whose goal is to identify only enough about the concept to enable a good decision to be recommended. We describe computational procedures for identifying optimal alternatives w.r.t. minimax regret in the presence of concept uncertainty; and describe several heuristic query strategies that focus on reduction of relevant concept uncertainty.
@article{icml09brv, author = {Craig Boutilier, Kevin Regan and Paolo Viappiani}, title = { Online Feature Elicitation in Interactive Optimization }, journal = { ICML-09 Iternational Conference on Machine Learning }, year = {2009} }
Select Awards
  • 2010 ⋅ Ontario Graduate Scholarship
  • 2006 ⋅ NSERC Canadian Graduate Scholarship (2006 - 2009)
  • 2006 ⋅ Award for Outstanding Achievement in Graduate Studies
  • 2005 ⋅ Best Student Paper - Privacy Security & Trust 05
My favourite Donoku site.