Be Considerate: Avoiding Negative Side Effects in Reinforcement Learning
We consider the problem of avoiding negative side effects on other agents, similarly to in our AAAI 2022 paper, but in the context of reinforcement learning. Parand Alizadeh Alamdari, Toryn Q. Klassen, Rodrigo Toro Icarte, and Sheila A. McIlraith AAMAS 2022
AI systems may cause negative side effects because their given objectives don't capture everything that they should not do. We consider how to avoid side effects in the context of symbolic planning, including by finding plans that don't interfere with possible goals or plans of other agents. Toryn Q. Klassen, Sheila A. McIlraith, Christian Muise, and Jarvis Xu AAAI 2022
Reward Machines: Exploiting Reward Function Structure in Reinforcement Learning
This is the expanded version of our ICML 2018 paper that introduced reward machines, which give structured representations of reward functions. This paper uses a slightly different definition and introduces the CRM algorithm, a simpler variant of the QRM algorithm from the original paper. Rodrigo Toro Icarte, Toryn Q. Klassen, Richard Valenzano, and Sheila A. McIlraith JAIR, Volume 73, 2022
Representing Plausible Beliefs about States, Actions, and Processes
This thesis deals with the topic of modelling an agent’s beliefs about a dynamic world in a way that allows for changes in beliefs, including retracting of beliefs. It elaborates on work from the KR 2018 and KR 2020 papers below, and also has material regarding beliefs about environmental processes. Toryn Q. Klassen PhD thesis, University of Toronto, 2021
We provide an account of explanation in terms of the beliefs of agents and the mechanisms by which agents revise their beliefs. The account allows for explanations to refer to beliefs. Maayan Shvo, Toryn Q. Klassen, and Sheila A. McIlraith EXTRAAMAS 2020
In the task of plan recognition, an observer infers the plan and goal of an actor. We introduce the notion of epistemic plan recognition, which uses epistemic logic to model the observer in a plan recognition setting, represent agent beliefs, and allow for the recognition of epistemic goals. Maayan Shvo, Toryn Q. Klassen, Shirin Sohrabi, and Sheila A. McIlraith AAMAS 2020
Specifying Plausibility Levels for Iterated Belief Change in the Situation Calculus
This paper describes a qualitative model of plausibility based on counting the extensions of certain predicates. Toryn Q. Klassen, Sheila A. McIlraith, and Hector J. Levesque KR 2018
Using Reward Machines for High-Level Task Specification and Decomposition in Reinforcement Learning
We introduce reward machines -- a form of automaton that gives a structured description of a reward function. This structure can be exploited by reinforcement learning algorithms to learn faster (analogously to how the structure of formulas was used in the AAMAS 2018 paper below). Rodrigo Toro Icarte, Toryn Q. Klassen, Richard Valenzano, and Sheila A. McIlraith ICML 2018
Advice-Based Exploration in Model-Based Reinforcement Learning
Linear temporal logic (LTL) formulas and a heuristic are used to guide exploration during reinforcement learning. Note that the slides have embedded videos that may not play on some systems. Rodrigo Toro Icarte, Toryn Q. Klassen, Richard Valenzano, and Sheila A. McIlraith Canadian AI 2018
Towards Representing What Readers of Fiction Believe
We use a temporal modal logic to describe a reader's beliefs about the reading process. We also discuss some ideas on how to model how a reader "carries over" real-world knowledge into fictional stories. Toryn Q. Klassen, Hector J. Levesque, and Sheila A. McIlraith Commonsense 2017
Towards Tractable Inference for Resource-Bounded Agents
This paper, written during my master's program, considers a formal model of belief that was meant to avoid attributing unlimited reasoning power to agents. Toryn Q. Klassen, Sheila A. McIlraith, and Hector J. Levesque Commonsense 2015
This theory paper about properties of hash functions resulted from my undergraduate research in theoretical computer science. Toryn Q. Klassen and Philipp Woelfel LATIN 2012