Featured image of post Continual Subtask Learning

Continual Subtask Learning

Adam White

Logistics

Time: 4:00-5:00 PM; Wednesday 12/06/2023
Location: Packard 202

Presenter

Adam White
Assistant Professor,
Unversity of Alberta

Abstract

In many real-world problems the agent is much smaller than the vast world in which it must operate. In such scenarios, the world appears non-stationary to the agent, and thus we require agents capable of stable, non-convergent, never-ending learning. Successful agents must balance specializing their learning to the current situation with the need to learn many things over time which can be combined to learn yet new things--a concept known as scaffolding. We will start with an empirical study in Atari demonstrating how current deep reinforcement learning agents fail to learn continually, and in fact catastrophically unlearn over time. The remainder of the talk will focus on new architectures and algorithms for learning many things--subtasks formulated as general value functions--in parallel from a single stream of experience.

Reference

https://arxiv.org/abs/2303.07507

Bio

Adam White is an Assistant Professor at the University of Alberta. He is a Fellow and Director of the Alberta Machine Intelligence Institute and a Principle Investigator of Reinforcement Learning and Artificial Intelligence group at the University of Alberta. Adam is a Canada CIFAR Chair in Artificial Intelligence. Adam is the co-creator of the Reinforcement Learning Specialization on Coursera. Adam's research is focused on understanding the fundamental principles of learning in young humans, animals, and artificial agents in both simulated worlds and real industrial control applications. Adam's group is deeply passionate about good empirical practices and new methodologies to help determine if our algorithms are ready for deployment in the real world.

Recording

Logo designed by Seohyun Jeon
Theme Stack designed by Jimmy