Efficient Temporal Consistency for Streaming Video Scene Analysis release_kum37btpuzb3lldoljw6tp4hli

by Ondrej Miksik, Daniel Munoz, J. Andrew Bagnell, Martial Hebert

Published by Figshare.

2018  

Abstract

We address the problem of image-based scene analysis from streaming video, as would be seen from a moving platform, in order to efficiently generate spatially and temporally consistent predictions of semantic categories over time. In contrast to previous techniques which typically address this problem in batch and/or through graphical models, we demonstrate that by learning visual similarities between pixels across frames, a simple filtering algorithm is able to achieve high performance predictions in an efficient and online/causal manner. Our technique is a meta-algorithm that can be efficiently wrapped around any scene analysis technique that produces a per-pixel semantic label distribution. We validate our approach over three different scene analysis techniques on three different datasets that contain different semantic object categories. Our experiments demonstrate our approach is very efficient in practice and substantially improves the quality of predictions over time.
In text/plain format

Archived Files and Locations

application/pdf  1.4 MB
file_sfsabu2iizbmpljmv4vgicbysm
s3-eu-west-1.amazonaws.com (publisher)
web.archive.org (webarchive)
Read Archived PDF
Preserved and Accessible
Type  article-journal
Stage   published
Date   2018-06-16
Work Entity
access all versions, variants, and formats of this works (eg, pre-prints)
Catalog Record
Revision: b48ecef5-7719-40fb-8d39-89f92c518032
API URL: JSON