Multi-fidelity Generative Deep Learning Turbulent Flows
release_kdevdvfsoff2ppsra6egg327p4
by
Nicholas Geneva, Nicholas Zabaras
2020
Abstract
In computational fluid dynamics, there is an inevitable trade off between
accuracy and computational cost. In this work, a novel multi-fidelity deep
generative model is introduced for the surrogate modeling of high-fidelity
turbulent flow fields given the solution of a computationally inexpensive but
inaccurate low-fidelity solver. The resulting surrogate is able to generate
physically accurate turbulent realizations at a computational cost magnitudes
lower than that of a high-fidelity simulation. The deep generative model
developed is a conditional invertible neural network, built with normalizing
flows, with recurrent LSTM connections that allow for stable training of
transient systems with high predictive accuracy. The model is trained with a
variational loss that combines both data-driven and physics-constrained
learning. This deep generative model is applied to non-trivial high Reynolds
number flows governed by the Navier-Stokes equations including turbulent flow
over a backwards facing step at different Reynolds numbers and turbulent wake
behind an array of bluff bodies. For both of these examples, the model is able
to generate unique yet physically accurate turbulent fluid flows conditioned on
an inexpensive low-fidelity solution.
In text/plain
format
Archived Content
There are no accessible files associated with this release. You could check other releases for this work for an accessible version.
Know of a fulltext copy of on the public web? Submit a URL and we will archive it
access all versions, variants, and formats of this works (eg, pre-prints)
Crossref Metadata (via API)
Worldcat
wikidata.org
CORE.ac.uk
Semantic Scholar
Google Scholar