Consistent Mesh Colors for Multi-View Reconstructed 3D Scenes
release_3kxi36pdoja4hlthdooiehwvau
by
Mohamed Dahy Elkhouly, Alessio Del Bue, Stuart James
2021
Abstract
We address the issue of creating consistent mesh texture maps captured from
scenes without color calibration. We find that the method for aggregation of
the multiple views is crucial for creating spatially consistent meshes without
the need to explicitly optimize for spatial consistency. We compute a color
prior from the cross-correlation of observable view faces and the faces per
view to identify an optimal per-face color. We then use this color in a
re-weighting ratio for the best-view texture, which is identified by prior mesh
texturing work, to create a spatial consistent texture map. Despite our method
not explicitly handling spatial consistency, our results show qualitatively
more consistent results than other state-of-the-art techniques while being
computationally more efficient. We evaluate on prior datasets and additionally
Matterport3D showing qualitative improvements.
In text/plain
format
Archived Files and Locations
application/pdf 4.7 MB
file_4um4qj6lnre2bg366y5rfcpfjy
|
arxiv.org (repository) web.archive.org (webarchive) |
2101.10734v1
access all versions, variants, and formats of this works (eg, pre-prints)