Machine Learning in Materials Processing & Characterization
Unit 9: Inverse Problems and Process Maps
FAU Erlangen-Nürnberg
After this unit, you will be able to:
The Fundamental Asymmetry
The forward problem is what we have been doing in previous units:
\[\mathbf{y} = f(\mathbf{x}) + \varepsilon\]
Given process parameters \(\mathbf{x}\), predict the resulting structure or properties \(\mathbf{y}\).
Examples:
| Input \(\mathbf{x}\) | Model \(f\) | Output \(\mathbf{y}\) |
|---|---|---|
| Laser power, scan speed | Thermal simulation | Melt pool depth |
| Alloy composition | CALPHAD | Phase fractions |
| Annealing time, temperature | Diffusion model | Grain size |
The inverse problem reverses the arrow:
\[\mathbf{x} = f^{-1}(\mathbf{y})\]
Given the desired output \(\mathbf{y}^*\), find the input \(\mathbf{x}^*\) that produces it.
The question engineers actually ask:
If the forward map \(f\) is many-to-one, then \(f^{-1}\) is one-to-many:
\[f(\mathbf{x}_1) = f(\mathbf{x}_2) = \ldots = f(\mathbf{x}_k) = \mathbf{y}^*\]
Materials example:
A grain size of \(d = 10\,\mu\text{m}\) can be achieved by:

Train a standard neural network to map \(\mathbf{y} \rightarrow \mathbf{x}\) using MSE loss:
\[\mathcal{L} = \frac{1}{N}\sum_{i=1}^{N} \|\mathbf{x}_i - \hat{\mathbf{x}}(\mathbf{y}_i)\|^2\]
If multiple valid solutions \(\{\mathbf{x}_1, \mathbf{x}_2, \mathbf{x}_3\}\) exist for the same \(\mathbf{y}^*\):
\[\hat{\mathbf{x}}_\text{NN} \approx \frac{1}{k}\sum_{j=1}^{k}\mathbf{x}_j \quad \text{(the mean of valid solutions)}\]
The mean of valid solutions is often not itself a valid solution!
The Averaging Trap
A standard neural network trained with MSE loss on an inverse problem will predict the average of all valid solutions. This average may lie in a region of parameter space that produces defective parts — it is a physically meaningless compromise.
Taming Unstable Solutions
A problem is well-posed if and only if all three conditions hold:
A problem that violates any of these conditions is called ill-posed.
| Condition | Forward Problem | Inverse Problem |
|---|---|---|
| Existence | Usually yes | May fail (no process gives exactly \(\mathbf{y}^*\)) |
| Uniqueness | Usually yes | Often fails (many-to-one forward map) |
| Stability | Usually yes | Often fails (noise amplification) |
Stability violation — noise amplification:
If the forward model maps a large range of inputs to a narrow range of outputs, then the inverse must map a narrow range of outputs back to a large range of inputs.
\[\text{Forward: } \Delta x = 100 \;\rightarrow\; \Delta y = 1\] \[\text{Inverse: } \Delta y = 1 \;\rightarrow\; \Delta x = 100\]
A \(1\%\) measurement error in \(y\) causes a \(100\%\) error in the inferred \(x\)!
Materials example:
You have an X-ray CT scanner and take projection images of a metal part from 180 angles.
The forward problem is straightforward:
\[p_\theta(s) = \int_{\text{ray}} \mu(x, y)\, dl\]
Each projection is a line integral of the attenuation coefficient \(\mu(x, y)\).
Question 1: Is the inverse problem (reconstructing \(\mu(x, y)\) from projections) well-posed?
\(\rightarrow\) It depends! With infinite noiseless projections, it is well-posed (Radon transform is invertible). With finitely many noisy projections, it is ill-posed — many images are consistent with the data.
Question 2: What happens if you reduce from 180 projections to 10?
\(\rightarrow\) The problem becomes severely ill-posed. The null space of the measurement operator grows — there are many structures that produce the same 10 projections. You need strong priors (sparsity, smoothness, physics) to regularize.
Core idea: We cannot solve the inverse problem from data alone — we need to add prior knowledge about what a “good” solution looks like.
\[\hat{\mathbf{x}} = \arg\min_{\mathbf{x}} \underbrace{\|f(\mathbf{x}) - \mathbf{y}_\text{obs}\|^2}_{\text{data fidelity}} + \lambda \underbrace{R(\mathbf{x})}_{\text{regularization}}\]
| Regularizer \(R(\mathbf{x})\) | Prior assumption | Effect |
|---|---|---|
| \(\|\mathbf{x}\|_2^2\) (L2 / Ridge) | Parameters are small | Smooth solutions, shrinks all parameters |
| \(\|\mathbf{x}\|_1\) (L1 / LASSO) | Solution is sparse | Feature selection, sets parameters to zero |
| \(\|\nabla \mathbf{x}\|_2^2\) | Solution is smooth | Suppresses oscillations |
| Total Variation \(\|\nabla \mathbf{x}\|_1\) | Piecewise constant | Preserves sharp edges |
Instead of generic mathematical penalties, we can use physical laws as regularizers:
\[\hat{\mathbf{x}} = \arg\min_{\mathbf{x}} \|\mathbf{y}_\text{obs} - f(\mathbf{x})\|^2 + \lambda_\text{PDE} \underbrace{\left\|\mathcal{N}[\mathbf{u}(\mathbf{x})]\right\|^2}_{\text{PDE residual}} + \lambda_\text{BC} \underbrace{\left\|\mathcal{B}[\mathbf{u}(\mathbf{x})]\right\|^2}_{\text{boundary conditions}}\]
Physical constraints in materials science:
Advantage over L1/L2:
Tikhonov regularization (also called Ridge regression in ML) adds an L2 penalty:
\[\hat{\mathbf{x}} = \arg\min_{\mathbf{x}} \|\mathbf{A}\mathbf{x} - \mathbf{y}\|_2^2 + \lambda \|\mathbf{L}\mathbf{x}\|_2^2\]
For the linear case \(\mathbf{y} = \mathbf{A}\mathbf{x}\), the solution is:
\[\hat{\mathbf{x}}_\lambda = (\mathbf{A}^T\mathbf{A} + \lambda \mathbf{L}^T\mathbf{L})^{-1}\mathbf{A}^T\mathbf{y}\]
Effect on the singular values:

Instead of finding a single “best” \(\mathbf{x}\), compute the posterior distribution:
\[p(\mathbf{x} \mid \mathbf{y}) = \frac{p(\mathbf{y} \mid \mathbf{x})\, p(\mathbf{x})}{p(\mathbf{y})}\]
| Bayesian Term | Inverse Problem Meaning |
|---|---|
| \(p(\mathbf{x} \mid \mathbf{y})\) — Posterior | Probability of parameters given observed data |
| \(p(\mathbf{y} \mid \mathbf{x})\) — Likelihood | How well the forward model fits the data |
| \(p(\mathbf{x})\) — Prior | Our regularization / domain knowledge |
| \(p(\mathbf{y})\) — Evidence | Normalization constant |
Connection to regularization:
Why Go Bayesian?
The posterior distribution gives us not just a point estimate but uncertainty quantification: we know which process parameters are well-constrained by the data and which are uncertain. This is essential for risk-aware manufacturing.
Navigating the Parameter Space
A process map is a visualization of the parameter space that classifies regions by outcome quality:

The engineering goal: Find the region in parameter space where all quality criteria are simultaneously satisfied — the safe operating window.
Each defect mechanism defines a constraint boundary in parameter space:
| Defect | Mechanism | Boundary condition |
|---|---|---|
| Keyholing | Vapor depression collapse | \(E_v > E_\text{keyhole}^*\) (energy density too high) |
| Lack of fusion | Insufficient melting | \(E_v < E_\text{fusion}^*\) (energy density too low) |
| Balling | Rayleigh instability | \(v > v_\text{ball}^*(P)\) (speed too high for given power) |
| Cracking | Thermal stress | \(\dot{T} > \dot{T}_\text{crack}^*\) (cooling rate too high) |
The safe operating window is the intersection:
\[\Omega_\text{safe} = \{\mathbf{x} : E_v(\mathbf{x}) \in [E_\text{fusion}^*, E_\text{keyhole}^*] \;\wedge\; v < v_\text{ball}^*(P) \;\wedge\; \dot{T} < \dot{T}_\text{crack}^*\}\]
A process corridor extends the static process map to account for temporal drift:

The corridor concept:
Real manufacturing processes have many more than 2 parameters:
LPBF example — at least 8 key parameters:
| Parameter | Typical Range |
|---|---|
| Laser power \(P\) | 100 – 500 W |
| Scan speed \(v\) | 200 – 2000 mm/s |
| Hatch spacing \(h\) | 50 – 150 μm |
| Layer thickness \(t\) | 20 – 100 μm |
| Spot diameter \(d\) | 50 – 200 μm |
| Preheat temperature \(T_0\) | 25 – 500 °C |
| Scan strategy | Stripe, island, meander |
| Gas flow rate | 1 – 10 m/s |
Idea: Train a classifier to predict defect type from process parameters:
\[\hat{c}(\mathbf{x}) = \text{Classifier}(P, v, h, t, d, T_0, \ldots)\]
where \(c \in \{\text{dense}, \text{keyhole}, \text{LOF}, \text{balling}, \text{cracking}\}\)
Common approaches:
flowchart LR
A["Experimental Database<br>(parameters, outcomes)"] --> B["Train Classifier<br>(RF, SVM, GP, NN)"]
B --> C["Decision Boundaries<br>in Parameter Space"]
C --> D["Process Map<br>with Confidence"]
D --> E["Safe Operating Window<br>+ Uncertainty"]
style A fill:#3498db,color:#fff
style B fill:#9b59b6,color:#fff
style C fill:#e67e22,color:#fff
style D fill:#2ecc71,color:#fff
style E fill:#1abc9c,color:#fffSensitivity analysis identifies which process parameters have the strongest influence on quality:
Local sensitivity (gradient-based):
\[S_i = \frac{\partial \hat{y}}{\partial x_i}\bigg|_{\mathbf{x}_0}\]
How much does the output change when parameter \(i\) is perturbed?
Global sensitivity (Sobol indices):
\[S_i = \frac{\text{Var}[\mathbb{E}[\hat{y} \mid x_i]]}{\text{Var}[\hat{y}]}\]
What fraction of the total output variance is attributable to parameter \(i\)?
Materials insight: If the safe operating window is narrow along parameter \(x_i\) but wide along \(x_j\):
Key concepts:
From Data to Physics
Recall from Unit 7: A Physics-Informed Neural Network represents the solution \(u(x, t)\) as a neural network and trains it by minimizing:
\[\mathcal{L} = \underbrace{\mathcal{L}_\text{data}}_{\text{match observations}} + \underbrace{\lambda_\text{PDE}\, \mathcal{L}_\text{PDE}}_{\text{satisfy physics}} + \underbrace{\lambda_\text{BC}\, \mathcal{L}_\text{BC}}_{\text{boundary conditions}}\]
Forward PINN: All parameters of the PDE are known. The network learns \(u(x,t)\).
Inverse PINN: Some parameters of the PDE are unknown. The network learns \(u(x,t)\) and the unknown parameters simultaneously.
Key insight: Unknown physical parameters (diffusivity, conductivity, viscosity) become trainable parameters of the optimization, alongside the network weights.
The heat equation:
\[\frac{\partial T}{\partial t} = \alpha \nabla^2 T\]
where \(\alpha\) is the thermal diffusivity — unknown.
Data: Temperature measurements \(T_\text{obs}(x_i, t_j)\) at scattered sensor locations.
Inverse PINN setup:
\[\mathcal{L} = \frac{1}{N_d}\sum_{i=1}^{N_d}\left(\hat{T}(x_i, t_i) - T_\text{obs}^i\right)^2 + \frac{\lambda}{N_c}\sum_{j=1}^{N_c}\left(\frac{\partial \hat{T}}{\partial t} - \alpha \nabla^2 \hat{T}\right)^2_{\!(x_j, t_j)}\]
flowchart TB
subgraph Inputs
X["Spatial coordinates x"]
T["Time t"]
end
subgraph Network["Neural Network NN(x,t; θ)"]
H1["Hidden layers"]
end
subgraph Outputs
U["Predicted field û"]
end
subgraph AutoDiff["Automatic Differentiation"]
DU["∂û/∂t, ∇²û"]
end
subgraph Loss["Total Loss"]
LD["J_data = ||û - u_obs||²"]
LP["J_PDE = ||∂û/∂t - α∇²û||²"]
end
subgraph Params["Trainable Parameters"]
TH["Network weights θ"]
AL["Unknown physics α"]
end
X --> Network
T --> Network
Network --> U
U --> AutoDiff
U --> LD
AutoDiff --> LP
AL --> LP
LD --> |"Backprop"| TH
LP --> |"Backprop"| TH
LP --> |"Backprop"| AL
style AL fill:#e74c3c,color:#fff
style LD fill:#3498db,color:#fff
style LP fill:#2ecc71,color:#fffTraditional approach — least-squares curve fitting:
Problems with curve fitting:
PINN advantages:
Many material properties depend on temperature or composition:
\[\frac{\partial c}{\partial t} = \nabla \cdot [D(T)\, \nabla c]\]
where \(D(T)\) is the temperature-dependent diffusivity — unknown functional form.
PINN approach:
\[\mathcal{L} = \underbrace{\sum_i (c_\text{obs}^i - \hat{c}^i)^2}_{\text{data}} + \lambda \underbrace{\sum_j \left(\frac{\partial \hat{c}}{\partial t} - \nabla \cdot [\hat{D}(\hat{T}) \nabla \hat{c}]\right)^2_j}_{\text{PDE}} + \mu \underbrace{\sum_k \left(\frac{d\hat{D}}{dT}\right)^2_k}_{\text{smoothness prior on } D}\]
Result: We discover the full functional relationship \(D(T)\) from data — not just a single number!
Scenario: You have cross-sectional images of melt pool shapes from LPBF experiments at various process parameters.
Forward model: Steady-state heat equation with convection:
\[\rho c_p (\mathbf{v} \cdot \nabla T) = \nabla \cdot [k(T)\, \nabla T] + Q_\text{laser}\]
Known: Laser power \(Q_\text{laser}\), scan speed \(\mathbf{v}\), melt pool boundary (solidus isotherm location)
Unknown: Temperature-dependent thermal conductivity \(k(T)\)
Inverse PINN:

Typical accuracy for parameter discovery:
| Problem | Unknown | Relative Error | Data Points |
|---|---|---|---|
| 1D Heat equation | \(\alpha\) (scalar) | \(< 1\%\) | 100 |
| 2D Diffusion | \(D\) (scalar) | \(1\text{–}3\%\) | 500 |
| Navier-Stokes | \(\nu\) (viscosity) | \(2\text{–}5\%\) | 1000 |
| Variable \(D(T)\) | \(D(T)\) (function) | \(5\text{–}10\%\) | 2000 |
Convergence challenges:
Practical Tip
Start the inverse PINN training with a good initial guess for the unknown parameter. If \(\alpha_\text{true} \approx 10^{-5}\), initializing at \(\alpha_0 = 10^{-7}\) may converge to a wrong local minimum. Use rough estimates from literature or dimensional analysis.
| Criterion | Traditional (Adjoint/FEM) | PINNs |
|---|---|---|
| Mesh requirement | Yes (FEM mesh) | No (meshfree) |
| Analytical solution | Sometimes needed | Never needed |
| Complex geometry | Challenging meshing | Easy (point sampling) |
| Nonlinear PDEs | Iterative, expensive | Natural (backprop) |
| Spatially varying parameters | Parameterization needed | Network approximation |
| Uncertainty quantification | Adjoint-based | Ensemble/Bayesian NN |
| Computational cost | One FEM solve per iteration | One training run |
| Maturity | Decades of development | Rapidly evolving |
| Industrial adoption | Widespread | Early stage |
Bottom line: PINNs are not always better — but they offer unique advantages for complex, multi-physics, data-scarce problems where meshing or analytical solutions are impractical.
Classical reconstruction:
\[\hat{\mu}(x, y) = \text{FBP}[p_\theta(s)] \quad \text{(Filtered Backprojection)}\]
Works well with many projections, but fails with limited/noisy data.
ML inverse approaches:
Why it matters for materials:
Real manufacturing involves coupled physics:
\[\text{Thermal} \longleftrightarrow \text{Mechanical} \longleftrightarrow \text{Microstructural} \longleftrightarrow \text{Fluid}\]
Multi-physics inverse PINN:
\[\mathcal{L} = \mathcal{L}_\text{data} + \lambda_1 \mathcal{L}_\text{heat} + \lambda_2 \mathcal{L}_\text{Navier-Stokes} + \lambda_3 \mathcal{L}_\text{phase-field} + \lambda_4 \mathcal{L}_\text{mechanics}\]
Example — LPBF multi-physics inversion:
Challenge: Balancing loss terms \(\lambda_1, \ldots, \lambda_4\) becomes critical — active research area (adaptive weighting, gradient normalization)
A digital twin is a continuously updated computational model of a physical system:

The inverse problem is at the heart of every digital twin:
Key concepts:
Discovering Governing Equations from Data
Standard regression: Fit parameters of a known model
\[y = a_0 + a_1 x + a_2 x^2 \quad \text{(we chose the form, fit the } a_i\text{)}\]
Symbolic regression: Discover both the form and the parameters
\[y = ??? \quad \text{(find the equation itself from data)}\]
The dream:
| Input Data | Discovered Equation |
|---|---|
| Planetary orbits | \(F = \frac{Gm_1 m_2}{r^2}\) |
| Pendulum motion | \(\ddot{\theta} = -\frac{g}{l}\sin\theta\) |
| Cooling curves | \(\frac{dT}{dt} = -h(T - T_\infty)\) |
| Stress-strain data | \(\sigma = K\varepsilon^n\) |
Challenge: The space of possible equations is combinatorially vast — brute force search is infeasible
Observation: Most physical laws involve only a few terms from a large space of possible terms.
\[\frac{d\mathbf{x}}{dt} = f(\mathbf{x}) = \boldsymbol{\Theta}(\mathbf{x})\, \boldsymbol{\xi}\]
where:
flowchart LR
A["Measurement Data<br>x(t)"] --> B["Compute Derivatives<br>ẋ(t)"]
B --> C["Build Library<br>Θ(x) = [1, x, x², sin(x), ...]"]
C --> D["Sparse Regression<br>ẋ = Θξ, minimize ||ξ||₁"]
D --> E["Discovered Equation<br>ẋ = -0.5x + 0.1x³"]
style A fill:#3498db,color:#fff
style B fill:#9b59b6,color:#fff
style C fill:#e67e22,color:#fff
style D fill:#e74c3c,color:#fff
style E fill:#2ecc71,color:#fffFor state variables \(\mathbf{x} = [x_1, x_2]\), a typical library includes:
\[\boldsymbol{\Theta}(\mathbf{x}) = \begin{bmatrix} 1 & x_1 & x_2 & x_1^2 & x_1 x_2 & x_2^2 & x_1^3 & \ldots & \sin(x_1) & \cos(x_1) & \ldots \end{bmatrix}\]
Design choices:
| Choice | Tradeoff |
|---|---|
| Polynomials up to degree \(d\) | Higher \(d\) → more expressive but more columns → harder sparsity |
| Trigonometric functions | Include if oscillatory behavior expected |
| Cross-terms \(x_i x_j\) | Capture interactions between state variables |
| Domain-specific terms | \(e^{-E_a/RT}\) for Arrhenius, \(x(1-x)\) for logistic growth |
The library should be:
The optimization problem:
\[\hat{\boldsymbol{\xi}} = \arg\min_{\boldsymbol{\xi}} \|\dot{\mathbf{X}} - \boldsymbol{\Theta}(\mathbf{X})\boldsymbol{\xi}\|_2^2 + \lambda \|\boldsymbol{\xi}\|_1\]
Alternative: Sequential Thresholded Least Squares (STLS)
STLS is the original SINDy algorithm — simpler and often more robust than LASSO for equation discovery.
You measure the angular position \(\theta(t)\) of a damped pendulum at 1000 time points over 30 seconds.
You build a library:
\[\boldsymbol{\Theta} = [1,\; \theta,\; \dot{\theta},\; \theta^2,\; \theta\dot{\theta},\; \dot{\theta}^2,\; \theta^3,\; \sin\theta,\; \cos\theta]\]
Question 1: What equation should SINDy discover?
\(\rightarrow\) The damped pendulum: \(\ddot{\theta} = -\frac{g}{l}\sin\theta - b\dot{\theta}\)
So \(\boldsymbol{\xi}\) should have nonzero entries only for \(\sin\theta\) and \(\dot{\theta}\).
Question 2: What if you forgot to include \(\sin\theta\) in your library?
\(\rightarrow\) SINDy would find the best sparse approximation using available terms, likely \(\ddot{\theta} \approx -c_1\theta + c_3\theta^3 - b\dot{\theta}\) — a polynomial approximation to sine. The library must contain the true terms!
Question 3: What if the data is very noisy?
\(\rightarrow\) Computing \(\ddot{\theta}\) by numerical differentiation amplifies noise. This is itself an ill-posed inverse problem! Solutions: smoothing, total variation differentiation, or weak-form SINDy.
Setup (McClarren Ch 2.5):
Library (10 candidate terms):
\[\boldsymbol{\Theta} = [1,\; \theta,\; \dot{\theta},\; \theta^2,\; \theta\dot{\theta},\; \dot{\theta}^2,\; \theta^3,\; \theta^2\dot{\theta},\; \sin\theta,\; \cos\theta]\]
Discovered coefficients after STLS:
| Term | \(\xi_i\) for \(\ddot{\theta}\) |
|---|---|
| \(\sin\theta\) | \(-9.81\) ✓ |
| \(\dot{\theta}\) | \(-0.10\) ✓ |
| All others | \(0\) ✓ |
Discovered equation: \(\ddot{\theta} = -9.81\sin\theta - 0.10\dot{\theta}\) — matches the ground truth exactly!
Problem: Given experimental stress-strain curves, discover the constitutive model.
Library for 1D plasticity:
\[\boldsymbol{\Theta} = [\varepsilon,\; \varepsilon^2,\; \varepsilon^{1/2},\; \varepsilon^n,\; \ln(1+\varepsilon),\; \dot{\varepsilon},\; \dot{\varepsilon}^m,\; T,\; e^{-Q/RT}]\]
Potential discoveries:
| Material behavior | Discovered law |
|---|---|
| Linear elastic | \(\sigma = E\varepsilon\) |
| Power-law hardening | \(\sigma = K\varepsilon^n\) |
| Johnson-Cook | \(\sigma = (A + B\varepsilon^n)(1 + C\ln\dot{\varepsilon}^*)(1 - T^{*m})\) |
| Voce hardening | \(\sigma = \sigma_s - (\sigma_s - \sigma_0)e^{-\varepsilon/\varepsilon_0}\) |
Advantage: No assumption about which model is “right” — the data decides!
William of Occam (c. 1287–1347):
“Entities should not be multiplied beyond necessity.”
In equation discovery:
L1 regularization is the mathematical implementation of Occam’s Razor:
\[\min \|\dot{\mathbf{X}} - \boldsymbol{\Theta}\boldsymbol{\xi}\|_2^2 + \lambda\|\boldsymbol{\xi}\|_1\]
Why This Matters for Materials Science
A discovered equation like \(\dot{d} = k_0 d^{-1} e^{-Q/RT}\) tells you about the mechanism (grain boundary diffusion-controlled growth). A neural network with 10,000 parameters that fits the same data tells you nothing about the physics.
Idea: Use a PINN to learn the solution field, and SINDy to discover the equation from the learned field.
Two-stage approach:
Why combine?
Integrated approach (PDE-FIND, DeepMoD):
| Criterion | Black-Box NN | SINDy / Symbolic |
|---|---|---|
| Prediction accuracy | High (on training distribution) | High (if correct form found) |
| Extrapolation | Poor — fails outside training range | Good — physics extrapolates |
| Interpretability | None — “black box” | Complete — explicit equation |
| Data efficiency | Needs large datasets | Works with moderate data |
| Parameter count | Thousands to millions | Typically < 10 |
| Physical insight | None | Reveals mechanisms |
| Regulatory acceptance | Difficult to certify | Equation can be validated |
For materials science and engineering:
Key concepts:
Putting It All Together
Optimal experimental design (OED) for inverse problems: choose measurements that maximally constrain the unknown parameters.
Key principle:
\[\text{Information} \propto \text{sensitivity of observation to unknown parameter}\]
If \(\frac{\partial y_\text{obs}}{\partial \theta}\) is large, the observation \(y_\text{obs}\) is informative about \(\theta\).
Practical guidelines:
Materials example: To determine thermal diffusivity from cooling data:
Data limitations:
Model limitations:
Physics limitations:
The responsible approach: Always validate inversions against independent data, report uncertainties, and state the assumptions explicitly.
| Approach | Discovers | Best for |
|---|---|---|
| Regularized optimization | Parameter values | Linear/mildly nonlinear problems |
| Bayesian inference | Posterior distributions | Uncertainty quantification |
| Process maps | Feasible regions | Manufacturing process design |
| PINNs (inverse) | Parameters + fields | Multi-physics, complex geometry |
| SINDy | Governing equations | Interpretable dynamics models |
Next unit: Uncertainty Quantification and Bayesian Methods
Textbooks:
Key papers:
Software:
pip install pysindypip install deepxde
© Philipp Pelz - Machine Learning in Materials Processing & Characterization