Efficient Temporal Consistency for Streaming Video Scene Analysis
release_kum37btpuzb3lldoljw6tp4hli
by
Ondrej Miksik, Daniel Munoz, J. Andrew Bagnell, Martial Hebert
2018
Abstract
We address the problem of image-based scene analysis from streaming video, as would be seen from a moving platform, in order to efficiently generate spatially and temporally consistent predictions of semantic categories over time. In contrast to previous techniques which typically address this problem in batch and/or through graphical models, we demonstrate that by learning visual similarities between pixels across frames, a simple filtering algorithm is able to achieve high performance predictions in an efficient and online/causal manner. Our technique is a meta-algorithm that can be efficiently wrapped around any scene analysis technique that produces a per-pixel semantic label distribution. We validate our approach over three different scene analysis techniques on three different datasets that contain different semantic object categories. Our experiments demonstrate our approach is very efficient in practice and substantially improves the quality of predictions over time.
In text/plain
format
Archived Files and Locations
application/pdf 1.4 MB
file_sfsabu2iizbmpljmv4vgicbysm
|
s3-eu-west-1.amazonaws.com (publisher) web.archive.org (webarchive) |
article-journal
Stage
published
Date 2018-06-16
access all versions, variants, and formats of this works (eg, pre-prints)
Datacite Metadata (via API)
Worldcat
wikidata.org
CORE.ac.uk
Semantic Scholar
Google Scholar