POD-DL-ROM
The POD-DL-ROM is a technique that combines a classical Reduced Basis Method (RBM) with a Deep Learning-based Reduced Order Model (DL-ROM). The goal is to overcome different limitations shared by classical RBMs, mainly mitigating the Kolmogorov barrier to reducibility, i.e. deal with situations where the dimension of the reduced linear space is too large.
DL-ROMs techniques learn in a non-intrusive way both the nonlinear trial manifold and the nonlinear dynamics as a function of the time coordinate and the parameters. It relies on deep neural networks.
The first step of the POD-DL-ROM consists in applying a Proper Orthogonal Decomposition (POD) to further reduce the costs of the offline stage of the algorithm, compared to DL-ROMs. This prior dimensionality reduction through a randomized singular value decomposition (rSVD) avoids to feed training data of large dimension $\mathcal{N}_h$ (even if the number of modes $N$ required for the POD is big, we assume here that $N << \mathcal{N}_h$ where $\mathcal{N}_h$ denotes the number of degrees of freedom. Indeed, although extremely efficient at testing time, when evaluating the PDE solution for any new testing-parameter instance, DL-ROMs require an expensive training stage, because of the extremely large number of network parameters (weights/biases) to be estimated. When $N$ is large (i.e. when the dimension of the linear trial subspace becomes very large), this ROM technique shows efficient computational performances during both offline and online stages, compared to classical projection-based ROMs.
To sum up this ROM technique, a two-step dimensionality reduction is performed: first, a POD (realized through rSVD) is applied on a set of High-Fidelity (HF) snapshots, then a DL-ROM is built to approximate the map between the parameters and the POD coordinates.
Offline
- A POD procedure (randomSVD)
- A Deep Learning-based Reduced Order Model through an encoder, a Deep Feedforward Neural Network (DFNN) and a decoder.
Online
An approximation is found with the DFNN and the decoder.
Codes:
Details:
A model problem
Here, $\mu=(\mu_1,\mu_2,\mu_3,\mu_4)$ is our parameter of interest. We employ P1 elements with a backward time scheme of order 2.
The parameter set is given by $\mathcal{G}=[0.002, 0.005] \times [30, 70] \times [0.4, 0.6]^2$.
POD-DL-ROM
OFFLINE STAGE
After the data compression with a rSVD, the training strategy consists in:
- 1/ encoding the compressed solution with a CAE that goes from $\mathbb{R}^N \to \mathbb{R}^n$ with the dimension of the latent space $n$ as close as possible to the number of parameters ($n << N$),
- 2/ approximate the latent space with a DFNN,
- 3/ decoding this approximation to find the original solution.
Many algorithm parameters can be adapted to enhance the accuracy or to reduce the offline part runtimes (Number of training parameters, maximum iterations, learning rate, number of minibatches, latent space dimension, number of convolution filters …)
See more details on the basic concepts of deep learning in Appendix A of “A Comprehensive Deep Learning-Based Approach to Reduced Order Modeling of Nonlinear Time-Dependent Parametrized PDEs”.
A pretraining stage consisting in initializing the network parameters can be added with a simplified model to reduce the number of required iterations.
Online algorithm
An approximation is found with the DFNN and the decoder. The key ingredient here is the DFNN that avoids the use of the encoder during the online time, part that is costly in time.
Furthermore, the POD compression reduce the offline costs compared to DL-ROMs: Indeed, the compressed snapshots (of dimension the number of modes $N$) are considered as the action of the first layer of the CAE instead of the true snapshots (the dimension is the number of degrees of freedom $>>N$).