A nonintrusive reduced order model for nonlinear transient thermal problems with nonparametrized variability

In this work, we consider a transient thermal problem, with a nonlinear term coming from the radiation boundary condition and a nonparametrized variability in the form complex scenarios for the initial condition and the convection coefficients and external temperatures. We use a posteriori reduced order modeling by snapshot Proper Orthogonal Decomposition. To treat the nonlinearity, hyperreduction is required in our case, since precomputing the polynomial nonlinearities becomes too expensive for the radiation term. We applied the Empirical Cubature Method, originally proposed for nonlinear structural mechanics, to our particular problem. We apply the method to the design of high-pressure compressors for civilian aircraft engines, where a fast evaluation of the solution temperature is required when testing new configurations. We also illustrate that when using in the reduced solver the same model as the one from the high-fidelity code, the approximation is very accurate. However, when using a commercial code to generate the high-fidelity data, where the implementation of the model and solver is unknown, the reduced model is less accurate but still within engineering tolerances in our tests. Hence, the regularizing property of reduced order models, together with a nonintrusive approach, enables the use of commercial software to generate the data, even under some degree of uncertainty in the proprietary model or solver of the commercial software.


Introduction
When designing an aircraft engine, a particular attention must be paid to prevent inflight compressor surge. A cutaway schematic of a civilian aircraft engine is shown in Fig. 1, with a highlight on the high-pressure compressor. The design of this part requires engineering expertise (e.g., for the heat transfer coefficients), analysis of experiments, and typically 2D and 3D finite element model simulations. In particular, transient evolutions of the temperature field in the compressor need to be computed rapidly and accurately to ensure efficient design iterations.  [1], with an emphasis on the high-pressure compressor Many numerical algorithms are called "reduced order modeling" methods: for instance, considering a coarser mesh or making use of the symmetry in a problem to reduced the number of degrees of freedom can be considered as reduced order models. Statistical meta-models in the form of nonlinear regressions can also be called reduced order models, see [2][3][4] for reviews in machine learning. Other methods consist in expressing the solution in the form tensor decompositions, see [5][6][7]. In this work, we consider a posteriori reduced order modeling, which consist in having the reduced model solve the equations of the physics, usually in the form of a Galerkin problem written on a reduced order basis. Among these methods, the reduced basis method [8,9] is well adapted to parametrized problems for which we can derive an efficient a posteriori error bounds. In what follows, we use the snapshot Proper Orthogonal Decomposition, where a low rank structure is search in the set of high-fidelity solutions by means of an eigenvalue problem written on the matrix of the correlations between these solutions, see [10,11]. Depending on the problem and the considered variabilities, a particular treatment, usually called hyperreduction, must be added to obtain an efficient reduced model. Three methods have been introduced simultaneously for the hyperreduction of nonlinear problems: the Missing Point Estimation (MPE) [12], the Empirical Interpolation Method (EIM) [13], and the hyperreduction [14], which coined the expression. Other methods have been introduced later, such as the Best Point Interpolation Method (BPIM) [15], the Discrete Empirical Interpolation Method (DEIM) [16], the Gauss-Newton with Approximated Tensors method (GNAT) [17], the Energy Conserving Sampling and Weighting method (ECSW) [18], the Empirical Cubature Method (ECM) [19] and the LP empirical quadrature procedure [20].
The contribution of this work consists in the adaptation of the Empirical Cubature Method to a nonlinear transient thermal problem with nonparametrized variability, and the implementation of the reduced order model in a nonintrusive fashion, in a framework developed at Safran [21]. While any of the previous cited hyperreduction method could have been used, we choose the ECM for its ability to easily derive an approximation for the assembling of a boundary term only, the regular distribution of the reduced integration points on this boundary, and the fact that the approximated reduced operators conserve the spectral properties of their high-fidelity counterpart.
We also illustrate that using a nonintrusive approach allows to reduce a problem without knowing precisely the model, thanks to the underlying regularization property of reduced order methods. Indeed, using the finite element software Z-set [22], for which we know precisely the models, we were able to construct a ROM solving the same model and obtained very accurate results. However, the models and solvers in the commercial software Abaqus [23] are proprietary, and using the same ROM leads to less accurate results, but still within engineering accuracy. Thus, the nonintrusivity of the method enables the use of commercial software, even without knowing precisely the implementation of the model or solver of the high-fidelity code.
In what follows, we first present the problem of interest in "High-fidelity model for transient thermal problem with convection and radiation boundary conditions" section. Then, the a posteriori reduced order modeling of this problem is introduced in "Reduced order model" section, in the context of nonparametrized variability. Finally, the features and performance of the reduced order model are illustrated by numerical experiments in "Numerical applications" section and conclusions are drawn in "Conclusion and prospects" section.

High-fidelity model for transient thermal problem with convection and radiation boundary conditions
We denote our stucture of interest , such that the boundary ∂ is partitioned in d = 39 surfaces: From the first law of thermodynamics, the energy balance in the absence of work reads: where n is the exterior normal on ∂ , u is the volumic internal energy (in J m −3 ), q is the heat flux density (in J s −1 m −2 , q · n| ∂V is the heat lost per second to the exterior through the boundary), and r a volumic heat source (in J s −1 m −3 ). In our case, r = 0. The divergence theorem applied to the surface integral in (1) yields the local relation ∂u ∂t (x, t) + ∇ · q(x, t) = 0. Let U be the massic internal energy (in J kg −1 ), so that u = ρU , where ρ is the density (in kg m −3 ). In our case, the density is assumed to be uniform and constant in time, and the massic internal energy is assumed to depend only on the temperature T, so that ∂u ∂t = ρc p ∂T ∂t , where the massic heat capacity c p (in J kg −1 K −1 ) is supposed uniform and constant in time. The heat-flux satisfies the Fourier law: q(x, t) = −λ∇T (x, t), where the heat conductivity λ (in J s −1 m −1 K −1 ) is also supposed uniform and constant in time. Moreover, we model the heat exchanges between the structure and the exterior by convection and radiation boundary conditions, namely q(x, t) · n(x) = h(x, t) (T (x, t) − T 1,e (x, t)) + where h denotes the convection coefficient (in J s −1 m −2 K −1 ), σ the Stefan-Boltzmann constant (in J s −1 m −2 K −4 ), is the emissivity coefficient (dimensionless), T 1,e and T 2,e are external temperature values. These coefficients are uniform on each surface (i) , 1 ≤ i ≤ d, and the radiation coefficients are also constant in time: Hence, T (x, t), x ∈ , t ∈ [0, t f ], is solution of the following system of equations: 1,e (t) Denote H 1 ( ) := {u ∈ L 2 ( )|∇u ∈ L 2 ( )}, where L 2 ( ) if the set of square integrable functions set from to R, i.e. v ∈ L 2 ( ) if the L 2 ( ) inner product The space discretization is obtained by using finite elements: consider a finite dimensional space V as a conforming approximation of H 1 ( ): V ⊂ H 1 ( ), V = Span{ϕ k } 1≤k≤N , with {ϕ k } 1≤k≤N the considered finite element basis, and N the number of nodes on the mesh. The solution temperature is now searched in V: Denote T (s) ∈ R N , the vector of components T k (s). Using a backwards Euler for the time discretization on {0, dt, ..., Jdt} and a Galerkin method for the space discretization on V, the problem writes: find T (s + 1) ∈ R N , 0 ≤ s ≤ J − 1, such that where, for At time step s+1, the system of nonlinear equations (4) is solved by a Newton algorithm: where T (p) (s + 1) ∈ R N denotes the p-th iteration at time step s + 1, T (s) is the known solution at the previous time step s, and where DF s DT T (p) (s + 1) ∈ R N ×N is such that We denote b ext (s + 1) ∈ R N the vector of external loading such that 1,e (s + 1) The iterations of the Newton algorithm (6) are stopped when ≤ HF tol , where · 2 denotes here the Euclidean norm on R N . The system of nonlinear equations (4), solved by the Newton algorithm (6), is our reference high-fidelity model.
The evolution of the coefficients (2) are computed using a 1D nodal network (in-house code), and the nonlinear transient thermal problem is solved using finite element solvers: in this study, we consider Abaqus and Z-set commercial softwares.

Reduced order model
Reduced order techniques can be used to accelerate the computation of the high-fidelity problem (4). We define the speedup as the ratio between the durations of the computation of the high-fidelity model (HFM) and the reduced order model (ROM). ROM techniques are usually decomposed in two stages: the offline stage, where information from the high-fidelity model is learned, and the online stage, where the reduced order model is constructed and exploited. In the offline stage occur computationally demanding tasks, whereas the online stage is required to be efficient, in the sense that only operations in computational complexity independent of the number N of degrees of freedom of the high-fidelity model are usually allowed.

Nonparametrized variability
In the offline stage, the high-fidelity problem (4) is solved for given configurations. In the general case, the variations between the candidate configurations are quantified using a low-dimensional parametrization, leading to a parametrized reduced order model. In this work, we consider nonparametrized variations between the configurations of interest, which we call variability and denote μ. The variability contains the time steps s, as well as a nonparametrized description of the configuration, which in our case is the loading referred to with a label "conf" for configuration. For instance, μ = (s; "computation 1"), means that we consider the time step s and a description conf = "computation 1" of the configuration, consisting of The surface partitioning (i) , 1 ≤ i ≤ d, and the coefficients ρ, c p , λ, σ , and T 2,e are supposed constant between the considered variabilities. We denote P off . the set of variabilities encountered during the offline stage.

Remark 1 (nonparametrized variability)
One could argue that the variabilities are actually parametrized: N parameters for T init , and 2dJ parameters for h and T 1,e . However, the dimension of the parametrization is too large for a strategy based on precomputing the variabilities to be efficient: many terms would need to be recombined in the online stage to derive the reduced loading, which would strongly penalize the speedup.

Derivation of the reduced order model
In what follows, we consider a posteriori reduced order modeling, which means that our reduced model involves an efficient Galerkin method no longer written on the finite element basis {ϕ k } 1≤k≤N , but on a reduced order basis {ψ k } 1≤k≤n , with n N , adapted to the problem at hand. The reduced solution temperature is written In what follows, we no longer write explicitly the variability conf for clarity of the presentation. The same derivations as the previous section are carried-out on this basis. The obtained reduced Newton algorithm at time step s + 1 and variability conf reads: where the vectorF s T (p) (s + 1) ∈ R n and the matrix DF s DV T (p) (s + 1) ∈ R n×n are obtained respectively as (5) and (7) by replacing the finite element basis {ϕ k } 1≤k≤N by the reduced order basis {ψ k } 1≤k≤n . Define also the reduced external loading vectorb ext (s + 1) in the same fashion as (8). The iterations of the reduced Newton algorithm (10) are stopped Following the vocabulary introduced in [21], the offline stage is composed of the three following steps: • data generation: this corresponds to the generation of the numerical approximation of the solutions to (4), using the Newton algorithm (6). Multiple temporal solutions are considered, for different loading conditions. The set of theses solutions {T (·, μ i )} 1≤i≤N c is called the snapshots set, where N c is the number of high-fidelity solutions computed at the considered variabilities μ = (s; conf). • data compression: this corresponds to the generation of the reduced order basis, obtained by looking for a hidden low-rank structure of the snapshots set. In this work, we employ the snapshot Proper Orthogonal Decomposition (POD), using the L 2 ( )-inner product to compute the correlations between the snapshots. The POD modes are obtained by truncating the eigenvalue decomposition of the correlation matrix: we denote POD the tolerance for the truncation. • operator compression: this step contains all the precomputations required to ensure that the reduced problem is computed in complexity independent of N . In our nonlinear setting, this task requires hyperreduction: we use the Empirical Cubature Method (ECM).
ECM was originally proposed for structural mechanics, we derive in this work a version for transient nonlinear thermal problems. In our problem, only the radiation term in (4) is nonlinear. Hence, we can precompute the assembling of all the other terms: define These tensors are computed by integrating directly the reduced order basis over the domain and keeping a numerical complexity linear with respect to the number of elements of the mesh. Once these tensors have been precomputed, the matrix DF s DV T (p) (s + 1) and the vectorF s T (p) (s + 1) in (10) can be computed as: where, for andb ext In (12) and (14), there remains some terms whose assembling still require integrations over the high-fidelity mesh. Due to the nonlinear dependence of these terms with respect to the reduced temperature, we cannot precompute the assembling of these terms with order-2 tensors. Consider the radiation part of the reduced internal forces vector: Recall the notation for the snapshots set {T (·, μ j )} 1≤j≤N c , corresponding to the set of highfidelity solutions. The reduced internal forces vector taken at the snapshots is in practice computed using quadrature formulas adapted to the used finite-element basis: where E denotes the set of elements of the mesh, n e denotes the number of integration points for the element e, ω k and x k are the integration weights and points of the considered element. ECM consists in replacing this high-fidelity quadrature (17) by an approximation adapted to the snapshots set {T (·, μ j )} 1≤j≤N c and the reduced order basis {ψ i } 1≤i≤n , and involving a small number of integration points: where d N G := e∈E n e (N G is the total number of integration points), the reduced integration pointsx k , 1 ≤ k ≤ d, are taken among the integration points of the highfidelity quadrature (17) and the reduced integration weightsω k are positive. In (18), the term in brackets corresponds to the reduced prediction taken at the reduced integration points. An approximation of the radiation part of the reduced internal forces vector can now be assembled in complexity independent of N (more precisely independent of the number of nodes on the boundary ∂ ).
We now briefly present how this reduced quadrature formula is obtained. We denote h q := −σ T 4 (·, μ (q//n)+1 )ψ (q%n)+1 ∈ L 2 (∂ ), where // and % are respectively the quotient and the remainder of the Euclidean division, Z is a subset of [1; N G ] of size d, with N G the number of integration points, and J Z ∈ R nN c ×d and g ∈ N nN c are such that for all 1 ≤ q ≤ nN c and all 1 ≤ k ≤ d, where Z k denotes the k -th element of Z and where we recall that n is the number of POD modes. Letω ∈ R + d . From the introduced notation, (J Zω ) q = −σ d k =1ω k T 4 (x Z k , μ (q//n)+1 )ψ (q%n)+1 (x Z k ), 1 ≤ q ≤ nN c , which is a candidate approximation for −σ ∂ T 4 (·, μ (q//n)+1 )ψ (q%n)+1 = g q , 1 ≤ q ≤ nN c . The best reduced quadrature formula of length d for the reduced internal forces vector is obtained as where · 2 stands for the Euclidean norm. If we wish to optimize the length of the reduced quadrature as well in the objective function, we obtain a NP-hard optimization problem.
In [18] a Sparse NonNegative Least-Squares (NNLS) algorithm, proposed in [24], is used to  [26] adapted to the nonnegative requirement. This algorithm consists in a greedy construction of the reduced quadrature scheme, with a stopping criterion on the accuracy of the quadrature over the values of the radiation part of the internal forces vector for the available snapshots, with a tolerance denoted ECM . In Newton algorithms, the tangent operator does not need to be computed very accurately. Hence, we use the previously obtained reduced quadrature formula to compute the radiation contribution to the reduced tangent operator (12).

Remark 2 (Polynomial nonlinearities)
In the literature, some polynomial nonlinearities have been treated without hyperreduction, for example, the advection term in fluid dynamics (with a ROM based on a Galerkin method in our context) only requires the precomputation of an order-3 tensor in the form ψ i · ψ j · ∇ ψ k , 1 ≤ i, j, k ≤ n, see [27] for the reduction of the nonlinear Navier-Stokes equations, with an exact operator compression step. Other examples can be found in structural dynamics with geometric nonlinearities, where order-2 and 3 tensors can be precomputed, see [28,Section 3.2] and [29]. Such a strategy can be derived in our case also. Define The nonlinear terms reduced by ECM in (12) and (14) can be precomputed as and ∂ n k=1T (p) However, the fact that the tensor θ ∂ i,j,k,l,m is of order five is very detrimental to the computation time, both for the offline stage for the computation of this tensor, and for the online stage, where the coefficients of this tensor are recombined to assemble the reduced problem. In practice in our numerical experiments, for more than 20 POD modes, the offline stage becomes untractable (computing the fifth-order tensor takes more than 70 times the duration of the high-fidelity solves) and no practical speedup is obtain in the online stage. For these reasons, the use of hyperreduction is mandatory for our applications.

Remark 3 (Initial condition)
Without any parametrization for the initial condition, the first step of the reduced Newton algorithm (10) at the first time step (s = 0) requires to compute scalar products on the mesh , whose computational complexity depends on N :T (0) (1; conf ) =T (0; conf ), whereT k (0; conf ) = (T init (conf ), ψ k ) L 2 ( ) , 1 ≤ k ≤ n. However, this computation is made only once per online run. If the variability of T init with respect to conf was parametrized in an affine fashion (or if we considered only a pre-known set of initial conditions T init in the online stage), the scalar products (·, ·) L 2 ( ) could have been precomputed during the offline stage.

Nonintrusive reduced order modeling library
This work has been implemented in an in-house nonintrusive reduced order modeling library developed at Safran, see [21]. Thanks to the nonintrusive implementation, the high-fidelity snapshots can be generated by a commercial software. In our case, we use the commercial versions of Abaqus [23] and Z-set [22].
The nonintrusivity is gained from having coded routines to read solutions and meshes from the used finite element software into an in-house format, and having recoded all the finite element assembling routines in our library. Then, considering nonparametrized variability is now readily available: we can, at the beginning of the online stage, for a new variability, assemble the reduced loading terms using these available assembling routines.
We will see in "Advantages of the nonintrusivity" section that the nonintrusivity of our ROM library, together with a regularizing property of ROM techniques, enables to reduce problems using snapshot generated by a commercial software under some model uncertainty.

Data format for the reduced solution
In our application, the quantity of interest is the temperature field over the complete structure. As a consequence, the time spent to write the reduced prediction over the complete structure should be included in the duration of the ROM for the definition of the speedup. In our applications, this time for writing the solution on disk takes 10% of the duration of the HFM. Hence, to obtain speedups larger than 10, we use the PXDMF format [30], where the POD modes are written on disk during the offline phase, and only the coefficientsT k (s; conf), 1 ≤ k ≤ n, in (9) are saved during the online stage. The solution is reconstructed on the complete mesh each time it is needed, for instance for visualization purposes, the reconstruction is carried out on the fly by a paraview [31] reader, see Fig. 3.

Numerical applications
In this section, we consider the model order reduction of the problem (4) using the procedure described in "Reduced order model" section.

Simple test case with convection only using Z-set
We consider a very simple test case consisting of a square mesh with N = 676 degrees of freedom, a uniform initial condition of 100 • C, 100 time step for a duration of 1000 s, and a convection boundary condition with h = 1000 J s −1 m −2 K −1 and T 1,e oscillating between   100 and 1000 • C with a period of 600 s. A HFM solution is computed using Z-set, then a ROM is computed on the same configuration for validation purposes. The training data contains 100 snapshots. The POD selects 11 modes with POD = 10 −6 , and ECM is not computed here since radiation is not considered. The durations for the offline and online stages are given in Table 1, where "Remaining" refers to all the remaining operations including the reading of the snapshots and the operations required for the nonintrusive approach and where "compute online loading" refers to the computation of the reduced external forces vectors and the extraction of the sequences of coefficients h (i) (t) and T 1,e (t) from the input files, see "Nonintrusive reduced order modeling library" section. The HFM duration is 3.82 s, leading to a speedup of 48 in this case.
The comparison provided in Fig. 4 shows that the ROM is able to correctly reduce computations featuring convection boundary conditions.

Simple test case with radiation only using Z-set
We consider again a very simple test case consisting of the same square mesh as the previous section, with N = 676 degrees of freedom, a uniform initial condition of 1000 • C, 100 time step for a duration of 1000 s, and a convection boundary condition with σ = 3.9683 × 10 −8 J s −1 m −2 K −4 and T 2,e oscillating between 100 and 1000 • C with a period of 600 s. A HFM solution is computed using Z-set, then a ROM is computed on the same configuration again for validation purposes. The training data contains 100 snapshots. The POD selects 9 modes with POD = 10 −6 , and ECM selects 19 integration points with ECM = 10 −6 . The durations for the offline and online stages are given in Table 2. The HFM duration is 5.87 s, leading to a speedup of 37 in this case. The comparison provided in Fig. 5 shows that the ROM is able to correctly reduce computation featuring radiation boundary conditions.

Industrial application
In this section, we apply our nonintrusive reduced order strategy on an industrial study of the high-pressure compressor presented in "Introduction" section, where we recall that the number of degrees of freedom is N = 4785 . Various temporal scenarios of convection coefficients h (i) and external temperatures T (i) 1,e for each of the 39 surfaces (i) are considered, see Fig. 6. The physical duration of the simulation is now t f = 600 s, and Abaqus is chosen to compute the snapshots. We use the high-fidelity data from the scenarios 1 to 7 (defined in from Fig. 6) in the offline stage, while the accuracy of the ROM will be tested at the variability of scenario 8. We precise that the variability of our model is not parametrized by T min , T max and τ : these are parameters for a precomputing step handled by the in-house 1D nodal network code, which generates the nonparametrized time series h (i) (t) and T (i) 1,e (t), 1 ≤ i ≤ 39, which are the input of our problem. The POD selects 13 modes with POD = 10 −3 , and ECM selects 39 integration points with ECM = 10 −3 . The durations for the offline and online stages are given in Table 3. In this case, the POD part is relatively long (the computation of the matrix containing the correlations between the snapshot takes 38.2 s alone), because the training data contains 7 scenarios of 300 snapshots each, which means that the correlation matrix is a dense 2100 × 2100 matrix. The HFM duration is 52 s, leading to a speedup of 91 in this case.   The comparison provided in Fig. 7 shows the prediction of the ROM for the variability of the scenario 8, which was not in the training data. The points B and C are chosen to monitor the largest discrepancies, and an error close to −11 • C has been measured at t = 24 s (which correspond to a relative error of 2.4% at this time).
We now carry out an error analysis on this industrial test case, to assess the accuracy of the prediction with respect to the hyperparameter POD .  Table 4 and Fig. 8, where ECM = 10 −3 . In our experiments, the value of ECM does not have a significant impact (for values 10 −2 , 10 −3 or 10 −4 ), the reasons could be that the errors are dominated by the POD truncation or that the polynomial nonlinearities in our case are well approximated by few reduced integration points. In [21], where we considered nonlinear structural mechanics with elastoviscoplastic materials, the tolerance ECM had a significant impact on the accuracy of the ROM. We notice also that for values of POD smaller than 10 −3 the speedup is not interesting, and for values below 10 −4 the accuracy of the prediction does not increase anymore.  Fig. 9 Absolute error between the HFM and the ROM at the last time step (top) and over time on three nodes (bottom) using Z-set generated snapshots (left) and Abaqus generated snapshots (right)

Advantages of the nonintrusivity
In this section, we consider the same high-pressure compressor as the previous section, and simulate the evolution of the temperature during 200 s. We compute the offline stage using only one scenario (of initial and boundary conditions), and a ROM prediction for the same scenario, then compare this ROM prediction with the HF prediction, using both Z-set and Abaqus. Since Z-set cannot deal with time dependent convection coefficients h, we keep them constant in time. The tolerance coefficients for the reduction and hyperreduction are respectively POD = 10 −6 and ECP = 10 −5 , which should lead to very accurate results. The comparisons are illustrated in Fig. 9.
We observe that the results with Z-set are 3 orders of magnitude more accurate that the ones with Abaqus. Even though we used the commercial version of Z-set in our nonintrusive framework, we have the source of the code at our disposal, and we have checked that our reduced model, in the online stage, implements the same model as the one in Z-set. Abaqus being a commercial code, the exact implementation of the solvers and models is proprietary and not known. However the accuracy of the ROM using Abaqus generated snapshot is still within engineering tolerances, even in the context of the nonparametrized variability detailed in "Industrial application" section. This robustness with respect to the model is provided by the a posteriori reduced order modeling technique, where regularity features are transported from the snapshots to the modes. In [32], a reduced order model is constructed from Large Eddy Simulations snapshots, but a resolution of the Navier-Stokes equations with no turbulence model is carried out in the online stage, with accurate results. In our application, we illustrate that the nonintrusive framework enables, thanks to the model robustness provided by the ROM technology, to reduced computations made by software without knowing the exact model and solver.

Conclusion and prospects
In this work, we propose a nonintrusive reduced order modeling strategy for nonlinear transient thermal problems with convection and radiation boundary conditions, under nonparametrized variability in the form of complex scenarios for the initial and boundary conditions. The Empirical Cubature Method is applied to derive a reduced quadrature scheme for the radiation boundary term, ensuring a fast construction of the reduced problem. In a 2D setting representative of the design process of a high-pressure compressor, the approach can successfully reduce the transient temperature prediction under the nonparametrized variability of the boundary condition scenarios, coming from a weak coupling with an in-house 1D code, with a speedup of 91 and an accuracy within engineering tolerances. The nonintrusivity of the approach enables the use of a commercial code to generate the snapshots, and the construction a reduced model without the knowledge of the exact high-fidelity model and solver.
In a future work, we plan to extend the method to a fully nonlinear setting, where the heat capacity and the heat conductivity depend on the temperature in a nonlinear fashion. The method needs to be challenged on larger 3D cases, for which better speedups are expected.