Object discovery and representation networks
release_vkem4cnhpnhf3bqebzusk4o45m
by
Olivier J. Hénaff, Skanda Koppula, Evan Shelhamer, Daniel Zoran, Andrew Jaegle, Andrew Zisserman, João Carreira, Relja Arandjelović
2022
Abstract
The promise of self-supervised learning (SSL) is to leverage large amounts of
unlabeled data to solve complex tasks. While there has been excellent progress
with simple, image-level learning, recent methods have shown the advantage of
including knowledge of image structure. However, by introducing hand-crafted
image segmentations to define regions of interest, or specialized augmentation
strategies, these methods sacrifice the simplicity and generality that makes
SSL so powerful. Instead, we propose a self-supervised learning paradigm that
discovers this image structure by itself. Our method, Odin, couples object
discovery and representation networks to discover meaningful image
segmentations without any supervision. The resulting learning paradigm is
simpler, less brittle, and more general, and achieves state-of-the-art transfer
learning results for object detection and instance segmentation on COCO, and
semantic segmentation on PASCAL and Cityscapes, while strongly surpassing
supervised pre-training for video segmentation on DAVIS.
In text/plain
format
Archived Files and Locations
application/pdf 4.7 MB
file_chqo4wsh7nbwteapivjzrlwdwy
|
arxiv.org (repository) web.archive.org (webarchive) |
2203.08777v2
access all versions, variants, and formats of this works (eg, pre-prints)