Image
Stanford EE

Continual subtask learning

Summary
Adam White
Packard 202
Dec
6
Date(s)
Content

Abstract: In many real-world problems the agent is much smaller than the vast world in which it must operate. In such scenarios, the world appears non-stationary to the agent, and thus we require agents capable of stable, non-convergent, never-ending learning. Successful agents must balance specializing their learning to the current situation with the need to learn many things over time which can be combined to learn yet new things--a concept known as scaffolding. We will start with an empirical study in Atari demonstrating how current deep reinforcement learning agents fail to learn continually, and in fact catastrophically unlearn over time. The remainder of the talk will focus on new architectures and algorithms for learning many things--subtasks formulated as general value functions--in parallel from a single stream of experience.

Bio: Adam White is an Assistant Professor at the University of Alberta. He is a Fellow and Director of the Alberta Machine Intelligence Institute and a Principle Investigator of Reinforcement Learning and Artificial Intelligence group at the University of Alberta. Adam is a Canada CIFAR Chair in Artificial Intelligence. Adam is the co-creator of the Reinforcement Learning Specialization on Coursera. Adam's research is focused on understanding the fundamental principles of learning in young humans, animals, and artificial agents in both simulated worlds and real industrial control applications. Adam's group is deeply passionate about good empirical practices and new methodologies to help determine if our algorithms are ready for deployment in the real world.