Consistent Mesh Colors for Multi-View Reconstructed 3D Scenes release_3kxi36pdoja4hlthdooiehwvau

by Mohamed Dahy Elkhouly, Alessio Del Bue, Stuart James

Released as a article .

2021  

Abstract

We address the issue of creating consistent mesh texture maps captured from scenes without color calibration. We find that the method for aggregation of the multiple views is crucial for creating spatially consistent meshes without the need to explicitly optimize for spatial consistency. We compute a color prior from the cross-correlation of observable view faces and the faces per view to identify an optimal per-face color. We then use this color in a re-weighting ratio for the best-view texture, which is identified by prior mesh texturing work, to create a spatial consistent texture map. Despite our method not explicitly handling spatial consistency, our results show qualitatively more consistent results than other state-of-the-art techniques while being computationally more efficient. We evaluate on prior datasets and additionally Matterport3D showing qualitative improvements.
In text/plain format

Archived Files and Locations

application/pdf  4.7 MB
file_4um4qj6lnre2bg366y5rfcpfjy
arxiv.org (repository)
web.archive.org (webarchive)
Read Archived PDF
Preserved and Accessible
Type  article
Stage   submitted
Date   2021-01-26
Version   v1
Language   en ?
arXiv  2101.10734v1
Work Entity
access all versions, variants, and formats of this works (eg, pre-prints)
Catalog Record
Revision: 6ac0ed6d-43d2-43d4-9e3e-6210fa41d22d
API URL: JSON