Reduced Basis Methods
1. Model Order Reduction
1.1. Definition
1.1.1. Problem statement
Goal : Replicate input-output behavior of large-scale system \(\Sigma\) over a certain (restricted) range of
- 
forcing inputs and
 - 
parameter inputs
 
Given large-scale system \(\Sigma_\mathcal{N}\) of dimension \(\mathcal{N}\), find a reduced order model \(\Sigma_N\) of dimension \(N << \mathcal{N}\) such that: The approximation error is small, i.e., there exists a global error bound such that :
- 
\(\|u(\mu) - u_N (\mu)\| \leq \varepsilon_{\mathrm{des}}\), and \(|s(\mu) - s_N (\mu)| \leq \varepsilon^s_{\mathrm{des}} , \forall \mu \in D^{\mu}\).
 - 
Stability (and passivity) is preserved.
 - 
The procedure is computationally stable and efficient.
 
1.2. Motivation
Generalized Inverse Problem :
- 
Given PDE(\(\mu\)) constraints, find value(s) of parameter \(\mu\) which:
- 
(OPT) minimizes (or maximizes) some functional;
 - 
(EST) agrees with measurements;
 - 
(CON) makes the system behave in a desired manner;
 - 
or some combination of the above
 
 - 
 - 
Full solution computationally very expensive due to a repeated evaluation for many different values of \(\mu\)
 - 
Goal: Low average cost or real-time online response.
 
1.3. Methodologies
- 
Reduced Basis Methods
 - 
Proper Orthogonal Decomposition
 - 
Balanced Truncation
 - 
Krylov Subspace Methods
 - 
Proper Generalized Decomposition
 - 
Modal Decomposition
 - 
Physical model reduction
 - 
…
 
| 
 Disclaimer on Model Order Reduction Techniques 
  | 
1.4. Summary
Many problems in computational engineering require many or real-time evaluations of PDE(\(\mu\))-induced + input-output relationships.
Model order reduction techniques enable certified, real-time calculation + of outputs of PDE(\(\mu\)) + for parameter estimation, optimization, and control.
2. Reduced Basis Method
2.1. Problem Statement
A model order reduction technique that allows efficient and reliable reducedorder approximations for a large class of parametrized partial differentialequations (PDEs) in real-time or in the limit of many queries.
- 
Comparison to other model reduction techniques:
- 
Parametrized problems(material, constants, geometry,…)
 - 
A posteriori error estimation
 - 
Offline-online decomposition
 - 
Greedy algorithm (to construct reduced basis space)
 
 - 
 - 
Motivation :
- 
Efficient solution of optimization and optimal control problems governed by parametrized PDEs.
 
 - 
 
2.1.1. Problem statement
Given a parameter \(\underbrace{\mu}_{\text{parameter}} \in \underbrace{D^\mu}_\text{parameter domain}\), evaluate \(\underbrace{s(\mu)}_\text{output} = L(\mu)^T \underbrace{u(\mu)}_\text{field variable}\) where \(u(\mu)\in\underbrace{X}_\text{FE space}\) satisfies a PDE \(A(\mu) u(\mu) = f(\mu)\).
Difficulties :
- 
Need to solve \(\textit{PDE}_{FE}(\mu)\) numerous times at different values of \(\mu\),
 - 
Finite element space \(X\) has a large dimension \(\mathcal{N}\).
 
2.1.3. General Problem Statement
Given a system \(\Sigma_\mathcal{N}\) of large dimension N,
where
- 
\(u(\mu, t) \in \mathbb{R}^{\mathcal{N}}\), the state
 - 
\(s(\mu, t)\), the outputs of interest
 - 
\(g(t)\), the forcing or control inputs
 
are functions of
- 
\(\mu \in D\), the parameter inputs
 - 
\(t\), time
 
and the matrices \(M\), \(A\), \(B\), and \(L\) also depend on \(\mu\).
Construct a reduced order system \(\Sigma_N\) of dimension \(N \ll \mathcal{N}\),
where \(u_N(\mu) \in \mathbb{R}^N\) is the reduced state.
We start by considering \(\dot{u} = 0\)
Full Model
\(\begin{align} A(\mu) u(\mu)& = & F(\mu)\\ s(\mu)&=&L^T(\mu) u(\mu) \end{align}\)
Reduced Model
\(\begin{align} A_N(\mu) u_N(\mu)& = & F_N(\mu)\\ s_N(\mu)&=&L^T_N(\mu) u_N(\mu) \end{align}\)
2.2. Key ingredients
2.2.1. Approximation
- 
Take « snapshots » at different \(\mu\)-values: \(u(\mu_i), i = 1 \ldots N\), and let \(Z_N=[\xi_1,\ldots,\xi_N] \in \mathbb{R}^{\mathcal{N}\times N}\) where the basis/test functions, \(\xi_i\) « \(=\) » \(u(\mu_i)\), are orthonormalized
 - 
For any new \(\mu\), approximate \(u\) by a linear combination of the \(\xi_i\) \(u(\mu) \approx \sum_{i=1}^N u_{N,i}(\mu) \xi_i = Z_N u_N(\mu)\) determined by Galerkin projection, i.e.,
 
2.2.2. A posteriori error estimation
- 
Assume well-posedness; \(A(\mu)\) positive and definite with a minimal eigenvalue \(\alpha_a :=\lambda_1 >0\), where \(A \xi=\lambda X \xi\) and \(X\) corresponds to the \(X\)-inner product, \((v, v)_X = \|v\|_X^2\)
 - 
Let \(\underbrace{e_N = u - Z_N\ u_N}_{\text{error}}\) , and \(\underbrace{r = F - A\ Z_N\ u_N}_{\text{residual}}, \forall \mu \in D^\mu\), so that \(A(\mu) e_N (\mu) = r(\mu)\)
 - 
Then (LAX-MILGRAM) for any \(\mu \in D^\mu\), \(\|u(\mu)- Z_N u_N(\mu) \|_X \leq \frac{\|r(\mu)\|_{X'}}{\alpha_{LB}(\mu)} =: \Delta_N(\mu)\), \(|s(\mu)-s_N(\mu)| \leq \|L\|_{X'} \Delta_N(\mu) =: \Delta^s_N(\mu)\) where \(\alpha_{LB}(\mu)\) is a lower bound to \(\alpha_a(\mu)\), and \(\|r\|_{X'}=r^T X^{-1} r\).
 
2.2.3. Offline-Online decomposition
How do we compute \(u_N\), \(s_N\), \(\Delta_N^s\) for any \(\mu\) efficiently online ?
We assue \(A(\mu) = \displaystyle\sum_{q=1}^Q \theta^q(\mu)A^q\) where
- 
\(\theta^q(\mu)\) are parameter depandent coefficients,
 - 
\(A^q\) are parameter independent matrices
 
so that \(A_N(\mu) = Z_N^T A(\mu)Z_N = \displaystyle \sum_{q=1}^Q \theta^q(\mu)\underbrace{Z_N^T A^q Z_N}_\text{parameter independant}\)
Since all required quantities can be decomposed in this manner, we can split the process in two phases :
- 
OFFLINE : Form and store \(\mu\)-independant quantities at cost \(O(\mathcal{N})\),
 - 
ONLINE : For any \(\mu\in D^\mu\), compute approximation and error obunds at cost \(O(QN^2+N^3)\) and \(O(Q^2N^2)\).
 
2.2.4. Greedy Algorithm
How do we choose the sample points \(\mu_i\) (snapshots) optimally ?
Given \(Z_{N=2} « = » [u(\mu_1), u(\mu_2)]\), we choos \(\mu_{N+1}\) as follows
and \(Z_{N+1} := [u(\mu_1),\cdots, u(\mu_{N+1})\).
- 
Key : \(\Delta_N(\mu)\) is sharp and inexpensive to compute (online)
 - 
Error bound gives « optimal » samples, so we get a good approximation \(u_N(\mu)\).
 
2.3. Summary
2.3.1. Reduced Basis Opportunities Computational Opportunities
- 
We restrict our attention to the typically smooth and low-dimensional manifold induced by the parametric dependence.
\(\Rightarrow\) Dimension reduction - 
We accept greatly increased offline cost in exchange for greatly decreased online cost.
\(\Rightarrow\) Real-time and/or many-query context 
2.3.2. Reduced Basis Relevance
Real-Time Context (control,\(\ldots\)): \(\begin{align} \mu & \rightarrow & s_N(\mu), \Delta^s_N(\mu) & \\ t_0 \text{(« input »)} & & & t_0+\delta t_{\mathrm{comp}} (\text{« response »}) \end{align}\)
Many-Query Context (design,\(\ldots\)): \(\begin{align} \mu_j & \rightarrow & s_N(\mu_j), \Delta^s_N(\mu_j),\quad j=1\ldots J \\ t_0 & & t_0+\delta t_{\mathrm{comp}} J\quad (J \rightarrow \infty) \end{align}\)
\(\Rightarrow\) Low parginal (real-time) and/or low average (many-query) cost.
2.3.3. Reduced Basis Challenges
- 
A Posteriori error estimation
- 
Rigorous error bounds for outputs of interest
 - 
Lower bounds to the stability « constants »
 
 - 
 - 
Offline-online computational procedures
- 
Full decoupling of finite element and reduced basis spaces
 - 
A posteriori error estimation
 - 
Nonaffine and nonlinear problems
 
 - 
 - 
Effective sampling strategies
- 
High parameter dimensions
 
 - 
 
2.3.4. Reduced Basis Outline
- 
Affine Elliptic Problems
- 
(non)symmetric, (non)compliant, (non)coercive
 - 
(Convection)-diffusion, linear elasticity, Helmholtz
 
 - 
 - 
Affine Parabolic Problems
- 
(Convection)-diffusion equation
 
 - 
 - 
Nonaffine and Nonlinear Problems
- 
Nonaffine parameter dependence, nonpolynomial nonlinearities
 
 - 
 - 
Reduced Basis (RB) Method for Fluid Flow
- 
Saddle-Point Problems (Stokes)
 - 
Navier-Stokes Equations
 
 - 
 - 
Applications
- 
Parameter Optimization and Estimation (Inverse Problems)
 - 
Optimal Control
 
 -