Wandering Within a World: Online Contextualized Few-Shot Learning
release_n7kh3zro7zdtpodn63btex5jti
by
Mengye Ren, Michael L. Iuzzolino, Michael C. Mozer, Richard S. Zemel
2020
Abstract
We aim to bridge the gap between typical human and machine-learning
environments by extending the standard framework of few-shot learning to an
online, continual setting. In this setting, episodes do not have separate
training and testing phases, and instead models are evaluated online while
learning novel classes. As in the real world, where the presence of
spatiotemporal context helps us retrieve learned skills in the past, our online
few-shot learning setting also features an underlying context that changes
throughout time. Object classes are correlated within a context and inferring
the correct context can lead to better performance. Building upon this setting,
we propose a new few-shot learning dataset based on large scale indoor imagery
that mimics the visual experience of an agent wandering within a world.
Furthermore, we convert popular few-shot learning approaches into online
versions and we also propose a new contextual prototypical memory model that
can make use of spatiotemporal contextual information from the recent past.
In text/plain
format
Archived Files and Locations
application/pdf 3.2 MB
file_o3ddfshhorfs3bx6ktq3gv4loe
|
arxiv.org (repository) web.archive.org (webarchive) |
2007.04546v2
access all versions, variants, and formats of this works (eg, pre-prints)