FAU Erlangen-Nürnberg
What can a CNN already do for us?
What this unit is not.
Recap — Unit 3
Today — Unit 4
By the end of 90 minutes, you can:
Standards-grade descriptors
Standards-grade ≠ lossless
Three families
Each family answers questions you knew to ask.
Question: can we keep more information without drowning in \(10^6\)-dim raw pixels?
Answer: structured vectors — \(S_2\), descriptor stacks, learned embeddings — sized to data and task.
Systems where one scalar per mechanism breaks
Consequence
Up next: what changes when the descriptor is learned, not chosen.

Azimi et al., Sci. Rep. 2018 Azimi, Seyed Majid et al., (2018), doi:10.1038/s41598-018-20037-5
Note
The 45-point jump is the headline of this whole unit.

Holm et al., MMTA 2020 — review Holm, Elizabeth A. et al., (2020), doi:10.1007/s11661-020-06008-4
| Classical | Modern (learned) | |
|---|---|---|
| Input | Image \(\to\) metrics | Image / signal \(\to\) representation |
| Features | Hand-crafted, named | Learned (or correlation-based) |
| Bottleneck | Information loss | Data + compute + validation discipline |
Ethics carry over
Before training: map microstructure to tensor \(\mathbf{X}\).
Principle: encoding upper-bounds what physics the hypothesis class can express Neuer, Michael et al., (2024).
Garbage encoding \(\Rightarrow\) garbage in, regardless of architecture.
| Encoding | Shape | What the model sees |
|---|---|---|
| Hand-crafted | \(\mathbb{R}^D\), small | Pre-distilled features |
| \(S_2\) / patches | \(\mathbb{R}^{D'}\) | Correlations / local stats |
| Eigen-modes | \(\mathbb{R}^{K}\) | Linear modes of structure |
| Image + conv | \(\mathbb{R}^{H \times W \times C}\) | Spatial features (Unit 5) |
MLP turf
\[S_2(\mathbf{r}) = P\!\bigl(\text{phase}(\mathbf{x})=\alpha \,\wedge\, \text{phase}(\mathbf{x}+\mathbf{r})=\alpha\bigr)\]
Why MLP-friendly
Typical chain
Why it works
Idea. Stack registered microstructure fields (phase indicator, orientation channels) into a design matrix; PCA on standardised columns yields dominant modes of structural variation — “eigen-microstructures.”
Why standardise first?
Connect: Unit 5 CNNs learn spatial features end-to-end; eigen-modes are the linear baseline to beat.
MLP on flattened pixels?
1-D CNN is the natural architecture
Note
“CNN” is not a synonym for “image network.”
| Input type | Typical \(D\) | First-line model |
|---|---|---|
| Composition + process | 10–50 | MLP |
| Morphology scalars | 5–50 | MLP |
| \(S_2\) / MKS | \(10^2\)–\(10^3\) | MLP / shallow 1-D conv |
| 1-D spectrum | \(10^3\)–\(10^4\) | 1-D CNN |
| 2-D micrograph | \(10^4\)–\(10^7\) | CNN (Unit 5) |
| 3-D volume | \(10^6\)–\(10^9\) | 3-D CNN / U-Net |
Decision rule
Six task families
Common pattern
raw signal → CNN → label / property
No hand-crafted descriptors. Same architecture family across SEM, EBSD, TEM, XRD, X-ray CT.



materialsdata.nist.gov).










Same recipe across ten cases
What changed across cases
What did not change
Six cases spanning:
Pattern
process sensor → CNN → quality
decision
The time scale changes: characterization is offline, processing is online — sometimes at video rate.






Two CNNs in series
Six cases, same toolkit
Common ceiling
All sixteen cases fit:
Raw signal → Encoding → CNN
→ Loss → Label / property
What still varies


CNN wins when
\(S_2\) / MKS still competitive when
Invalid protocol
Reality
Rule. Group ID = whatever is exchangeable at deployment.
Mitigations (preview Unit 6)
Mitigations
Mitigations
The Unit 5 punchline
Use simpler models when
Practical rule
Note
“Use the simplest model that survives grouped CV.”
Forward pass / activations / training loop don’t change.
What changes in materials ML
Next week
Carry forward from today
Beyond Unit 5
Reading
Exercises

© Philipp Pelz - Machine Learning in Materials Processing & Characterization