Machine Learning in Materials Processing & Characterization
Unit 4: From Classical Metrics to Learned Representations

Prof. Dr. Philipp Pelz

FAU Erlangen-Nürnberg

FAU Logo IMN Logo CENEM Logo ERC Logo Eclipse Logo

§0 · Frame

01. Today’s Question

What can a CNN already do for us?

  • Sixteen real, published applications across characterization and processing.
  • Same convolutional toolkit, deployed everywhere from SEM to LPBF cameras.

What this unit is not.

  • Not a re-derivation of perceptron / MLP / activations — that is MFML Unit 4.
  • This deck assumes the forward pass and training loop are familiar.

02. Where We Are

Recap — Unit 3

Today — Unit 4

  • Turn microstructure into model-ready tensors.
  • Tour ten characterization + six processing case studies.
  • Diagnose pitfalls before next week’s CNN deep-dive.

03. Learning Outcomes

By the end of 90 minutes, you can:

  1. Quantify information loss when a micrograph collapses to one stereological scalar.
  2. Choose between tabular, \(S_2\), eigen-mode, image, and 1-D spectral encodings.
  3. Recognise published CNN applications across SEM, EBSD, TEM, XRD, X-ray CT, AM cameras, and welding sensors.
  1. Name failure modes that erase apparent CNN gains (specimen splits, lab shift, imbalance, segmentation noise, raw-pixel MLPs).
  2. Articulate why Unit 5 is about CNNs — locality + weight sharing as the right inductive bias.

§1 · Why Classical Metrics Aren’t Enough

04. Stereology in One Slide

Standards-grade descriptors

  • \(V_V\) — volume fraction per phase.
  • \(S_V\) — interface area per unit volume.
  • Mean intercept / ASTM grain-size \(G\).

Standards-grade ≠ lossless

05. Hand-Crafted Descriptor Families

Three families

  • Shape: aspect ratio, circularity, tortuosity.
  • Distribution: nearest-neighbour spacing, clustering indices.
  • Texture: ODF coefficients, pole figures.

Each family answers questions you knew to ask.

  • Strength: physical names, peer-reviewable, auditable.
  • Weakness: you only recover structure you designed the scalar to see — unknown mechanisms stay invisible.

06. The Information Bottleneck

  • Micrograph: \(\mathcal{O}(10^6)\) pixels of state.
  • ASTM-style scalar: one number per channel of interest.
  • Compression ratio: \(\sim 10^6\) : \(1\).

Question: can we keep more information without drowning in \(10^6\)-dim raw pixels?

Answer: structured vectors — \(S_2\), descriptor stacks, learned embeddings — sized to data and task.

07. Where ASTM Hits a Wall

Systems where one scalar per mechanism breaks

  • High-entropy alloys — multi-phase, partitioning, sluggish diffusion.
  • Additively manufactured parts — spatially varying solidification, not a stationary field.
  • Hierarchical composites — nm–µm length scales coexisting in one image.

Consequence

  • Scalar summaries assume stationarity and known relevant descriptors.
  • Modern materials violate both routinely.

Up next: what changes when the descriptor is learned, not chosen.

08. Hero Result — Steel Phase Classification

VGG-style FCNN architecture used by Azimi et al.

FCNN architecture: cropped SEM/LOM constituent → 224×224 → VGG-style conv stack → fully connected → softmax over phase classes Azimi, Seyed Majid et al., (2018), doi:10.1038/s41598-018-20037-5.

Azimi et al., Sci. Rep. 2018 Azimi, Seyed Majid et al., (2018), doi:10.1038/s41598-018-20037-5

  • Dual-phase steel constituents on SEM micrographs (martensite, bainite, pearlite, …).
  • Fully Convolutional Net + superpixel max-voting.
  • Prior SOTA: 48.9% → FCNN: 93.94%.
  • Same images. No new physics — representation change alone.

Note

The 45-point jump is the headline of this whole unit.

09. Where Hand-Crafted Hits a Wall — A Wider View

Target micrograph and three nearest CNN-feature neighbours.

CNN-based image retrieval on UHCS micrographs. A query micrograph (top-left) and the three nearest neighbours by CNN feature distance — same-class structures retrieved without any human-defined descriptor Holm, Elizabeth A. et al., (2020), doi:10.1007/s11661-020-06008-4.

Holm et al., MMTA 2020 — review Holm, Elizabeth A. et al., (2020), doi:10.1007/s11661-020-06008-4

  • Surveys CV/ML across classification, semantic segmentation, object detection, instance segmentation.
  • Pattern: where labels exist, learned representations match or beat hand-crafted features.
  • The bottleneck has moved: from “which descriptor?” to “which labels and which split?”

10. The Paradigm Shift

Classical Modern (learned)
Input Image \(\to\) metrics Image / signal \(\to\) representation
Features Hand-crafted, named Learned (or correlation-based)
Bottleneck Information loss Data + compute + validation discipline

Ethics carry over

  • Specimen splits, leakage, calibration — all unchanged.
  • Scientist still owns labels, splits, metrics, physics checks Neuer, Michael et al., (2024).

§2 · Encoding Microstructure for ML

11. The Encoding Question

Before training: map microstructure to tensor \(\mathbf{X}\).

Principle: encoding upper-bounds what physics the hypothesis class can express Neuer, Michael et al., (2024).

Garbage encoding \(\Rightarrow\) garbage in, regardless of architecture.

Encoding Shape What the model sees
Hand-crafted \(\mathbb{R}^D\), small Pre-distilled features
\(S_2\) / patches \(\mathbb{R}^{D'}\) Correlations / local stats
Eigen-modes \(\mathbb{R}^{K}\) Linear modes of structure
Image + conv \(\mathbb{R}^{H \times W \times C}\) Spatial features (Unit 5)

12. Tabular: Composition + Process

  • Often no image in \(\mathbf{X}\):
    • Composition fractions in \(\mathbb{R}^{d_{\text{el}}}\).
    • Process: temperature, time, cooling rate, atmosphere.
    • History: ordered steps (embedded or binned).

MLP turf

  • \(D \sim 10\)\(50\), well-defined units.
  • Standardise per train fold; freeze \((\mu, \sigma)\) at inference Neuer, Michael et al., (2024).
  • Watch: mass fractions sum to 1 → drop one column or use compositional geometry.

13. Two-Point Statistics \(S_2\)

\[S_2(\mathbf{r}) = P\!\bigl(\text{phase}(\mathbf{x})=\alpha \,\wedge\, \text{phase}(\mathbf{x}+\mathbf{r})=\alpha\bigr)\]

  • Translation-averaged correlation.
  • Captures length scales, anisotropy, clustering — far more than one scalar, far less than full pixels Sandfeld, Stefan et al., (2024).

Why MLP-friendly

  • Fixed-length vector after binning \(\mathbf{r}\) on a grid in the unit cell / ROI.
  • Pairs naturally with standardised inputs (Unit 3).
  • \(D \sim 10^2\)\(10^3\) — tractable on materials sample counts.

14. MKS Pipeline (Materials Knowledge Systems)

Typical chain

  1. Segment / phase-label microstructure.
  2. Compute \(S_2\) on a fixed grid of \(\mathbf{r}\).
  3. Standardise correlation components using train statistics only (Unit 3).
  4. Train MLP (or linear map) \(g_\theta(S_2) \approx\) property Sandfeld, Stefan et al., (2024).

Why it works

  • Bakes in translation invariance before the net sees data.
  • Keeps \(D\) in the hundreds — matches typical materials sample counts.
  • Strong baseline before escalating to CNNs on raw pixels.

15. Eigen-Microstructures

Idea. Stack registered microstructure fields (phase indicator, orientation channels) into a design matrix; PCA on standardised columns yields dominant modes of structural variation — “eigen-microstructures.”

Why standardise first?

  • Without it, PC1 often tracks brightness, thickness, detector gain — not microstructure.
  • With per-feature z-scores fit on train only, PCs more often reflect shape variation Sandfeld, Stefan et al., (2024).

Connect: Unit 5 CNNs learn spatial features end-to-end; eigen-modes are the linear baseline to beat.

16. Image as Tensor

  • 2-D micrograph: \(\mathbf{X} \in \mathbb{R}^{H \times W \times C}\).
  • 3-D tomography: \(\mathbf{X} \in \mathbb{R}^{D \times H \times W \times C}\).
  • \(C\) = channels: BSE/SE, EBSD orientation Euler angles, EDS element maps.

MLP on flattened pixels?

  • \(1024 \times 1024\) flattened → first dense layer ≈ \(10^9\) weights.
  • With \(N \sim 100\) specimens: spurious correlations win.
  • Solution: convolutional inductive bias (Unit 5).

17. Spectra as 1-D Signals

  • XRD pattern, EELS edge, Raman spectrum: \(\mathbf{x} \in \mathbb{R}^{N_{\text{channels}}}\).
  • Locality matters along the channel index — neighbouring bins describe the same peak.
  • The convolutional inductive bias applies in 1-D too.

1-D CNN is the natural architecture

  • Same shared weights, same locality argument as 2-D images.
  • Preview: Park 2017 (slide 28) — phase ID with a 1-D CNN.

Note

“CNN” is not a synonym for “image network.”

18. Encoding Decision Rule

Input type Typical \(D\) First-line model
Composition + process 10–50 MLP
Morphology scalars 5–50 MLP
\(S_2\) / MKS \(10^2\)\(10^3\) MLP / shallow 1-D conv
1-D spectrum \(10^3\)\(10^4\) 1-D CNN
2-D micrograph \(10^4\)\(10^7\) CNN (Unit 5)
3-D volume \(10^6\)\(10^9\) 3-D CNN / U-Net

Decision rule

  • Start with the smallest \(\mathbf{X}\) that passes physics + grouped CV.
  • Add representation capacity when grouped CV shows a persistent gap, not when train loss wants it McClarren, Ryan G., (2021).

20. Case 1 — Steel Phase Classification (Azimi 2018)

Side-by-side comparison of SEM input, ground-truth phase labels, and FCNN output.

Pixel-wise FCNN segmentation: SEM input → ground truth → predicted phase map (martensite, bainite, pearlite, ferrite). Same network, four different microstructures Azimi, Seyed Majid et al., (2018), doi:10.1038/s41598-018-20037-5.
  • Task. Classify constituents in dual-phase steel SEM micrographs.
  • Method. Fully Convolutional Net + max-voting on superpixels.
  • Data. Thousands of SEM tiles, expert-labelled.
  • Result. 93.94% vs prior SOTA 48.9%.
  • Lesson. Representation change alone unlocks the 45-pt jump.

21. Case 2 — UHCS Microstructure Manifold (DeCost & Holm)

Micrographs arranged on a 2-D embedding manifold.

2017 — Microstructure manifold. UHCS micrographs embedded by CNN features cluster spontaneously by heat-treatment class — no labels used to position them DeCost, Brian L. et al., (2017), doi:10.1016/j.actamat.2017.05.014.

Encoder + upsample + MLP segmentation pipeline.

2019 — High-throughput segmentation. Input SEM → CNN encoder (Conv1–5) + upsampling → per-pixel constituent map DeCost, Brian L. et al., (2019), doi:10.1017/S1431927618015635.
  • Dataset. 961 public UHCS micrographs (materialsdata.nist.gov).
  • Lesson. Pretrained CNN features cluster phase classes without labels — a transfer-learning preview (Unit 6) and the basis for Exercise 1.

22. Case 3 — U-Net for EBSD Phase Segmentation

Four EBSD-derived input channels and labels.

Inputs. Band-slope (BS), grain-boundary overlay, KAM map, manual labels — the EBSD-derived channels feeding the U-Net Martinez Ostormujof, T. et al., (2022), doi:10.1016/j.matchar.2021.111638.

Per-temperature segmentation accuracy from BS U-Net.

Cross-temperature results. Three test temperatures (T1–T3); top: BS micrograph; middle: prediction with accuracy 99.3% / 98.5% / 97.2%; bottom: error map Martinez Ostormujof, T. et al., (2022), doi:10.1016/j.matchar.2021.111638.
  • Task. Pixel-level martensite / ferrite-bainite segmentation across three tempering conditions.
  • Lesson. A standard U-Net on a single grayscale BS channel can reach the EBSD-quaternion baseline if augmentation is honest.

23. Case 4 — Complex Microstructure Inference (Durmaz et al. 2021)

LOM, SEM, and EBSD overlays of the same microstructure.

Multi-modal training data. Light-optical micrograph (LOM, top) and SEM (middle) of the same area, co-registered with EBSD-derived bainite-instance overlays (bottom) Durmaz, Ali Riza et al., (2021), doi:10.1038/s41467-021-26565-5.

Bainite instance segmentation results on LOM and SEM.

Lath-bainite instance segmentation. Top row: LOM input + SEM input; bottom row: per-instance prediction colored by lath family Durmaz, Ali Riza et al., (2021), doi:10.1038/s41467-021-26565-5.
  • Method. U-Net (semantic) + Mask R-CNN (instance) trained on EBSD-derived ground truth, deployed on LOM/SEM only at inference.
  • Lesson. EBSD-grade labels at training time → optical-microscopy throughput at inference time.

24. Case 5 — TEM Dislocation Segmentation (Govind et al. 2024)

Simulated bright-field TEM image used for training.

Simulated TEM dislocation training image — physics-based contrast with controllable label density Govind, Kishan et al., (2024), doi:10.1088/2632-2153/ad1a4e.

Real experimental TEM image with dislocation lines.

Real experimental TEM micrograph — dislocation arcs against grain background Govind, Kishan et al., (2024), doi:10.1088/2632-2153/ad1a4e.
  • Task. Instance segmentation of dislocations in TEM.
  • Method. YOLO-style + U-Net trained on simulated dislocation images, evaluated on real experiments.
  • Lesson. Simulation-augmented training bypasses the “never enough labels” bottleneck — standard wherever physics simulators are mature.

25. Case 6 — STEM Defects in Irradiated Steels (Roberts 2019)

DefectSegNet U-Net architecture.

DefectSegNet — encoder–decoder U-Net with trans-coder dense skip connections; 512×512 input → per-pixel probability map Roberts, Graham et al., (2019), doi:10.1038/s41598-019-49105-0.

Side-by-side STEM input, ground truth, and DefectSegNet predictions.

Per-class results: input STEM (DCI), ground truth, prediction, and overlay-comparison for precipitates and voids Roberts, Graham et al., (2019), doi:10.1038/s41598-019-49105-0.
  • Task. Semantic segmentation of voids, dislocation loops, precipitates in irradiated steels.
  • Result. ~85% IoU — matches inter-annotator variability.
  • Lesson. Once you hit the annotator floor, more model capacity buys nothing.

26. Case 7 — 3-D-CNN Composite Stiffness from RVEs

  • Result. >40% accuracy improvement over hand-engineered descriptors at a fraction of the FE cost.
  • Lesson. CNNs can act as homogenisation surrogates inside design loops where each FE call is too expensive.

27. Case 8 — Yield-Surface Prediction from Microstructure

  • Lesson. A learned representation lets one model output functional properties — anisotropic yield, stress-strain curves, dispersion relations.
  • Significance. Moves CNNs from “label predictors” to “constitutive surrogates.”

28. Case 9 — Lee/Park et al. 2020 — XRD Phase ID with 1-D CNN

1-D convolutional architecture for XRD phase identification.

1-D CNN architecture. XRD pattern (top-left) → stack of 1-D conv layers → flatten → fully-connected → multi-head outputs (existence + fraction of each phase) Lee, Jin-Woong et al., (2020), doi:10.1038/s41467-019-13749-3.
  • Task. Phase ID in multi-phase inorganic mixtures from XRD.
  • Train on simulation, test on real. ~\(10^5\) patterns from ICSD with augmentation for strain, texture, peak broadening.
  • Result. ~100% phase ID; ~86% three-phase quantification on real experiments.
  • Lesson. CNN \(\neq\) image network — convolution applies wherever there is locality (peak shape along \(2\theta\)).

29. Case 10 — 3-D U-Net for Li-ion Electrode Tomography

3-D X-ray CT volume and three-class segmentation.

X-ray nano-CT input → 3-class 3-D segmentation. Input volume (top), reconstruction (right), final segmentation: active material (grey), carbon-binder (orange), pore (blue) Müller, Simon et al., (2021), doi:10.1038/s41467-021-26480-9.

Per-slice segmentation quality comparison.

Per-slice quality. Same input, four output variants — only the deep-learning column recovers the carbon-binder phase consistently Müller, Simon et al., (2021), doi:10.1038/s41467-021-26480-9.
  • Method. 3-D U-Net trained partly on synthetic electrodes with known voxel-level ground truth.
  • Lesson. Carbon-binder vs pore has near-zero contrast — thresholding fails; simulation-augmented CNN succeeds.

32. Case 11 — LPBF Powder-Bed Quality (Xception Transfer Learning)

  • Task. Classify powder-bed defects (balling, incomplete spreading, groove, ridge, spatters, protruding part, scattered powder, homogeneous) from line-sensor recoater images during LPBF Fischer, Felix Gabriel et al., (2022), doi:10.1016/j.matdes.2022.111029.
  • Method. Xception pretrained on ImageNet, fine-tuned on a Fraunhofer-ILT dataset acquired under coaxial / dark-field / diffuse lighting.
  • Result. 99.15% classification accuracy across seven classes (dark-field condition); per-class F1 between 97.85% and 99.71%.
  • Lesson. ImageNet pretraining transfers astonishingly well even to grayscale recoater frames — a clean transfer-learning teaser for Unit 6.

33. Case 12 — Thermographic Porosity Prediction in LPBF

Schematic of thermographic camera and melt-pool keyhole.

Sensing geometry. SWIR camera images the melt pool from above; sub-surface keyhole pores form below the visible track Oster, Simon et al., (2024), doi:10.1007/s10845-023-02117-0.

Photograph of an LPBF staircase test specimen.

Specimen used for ground truth. LPBF Inconel 718 staircase sample — thermography during build, post-build µCT defines per-voxel porosity labels Oster, Simon et al., (2024), doi:10.1007/s10845-023-02117-0.
  • Method. Multi-layer thermographic feature stack → supervised CNN classifier; CT ground truth.
  • Result. Accuracy ~0.96, F1 ~0.86 for keyhole porosity in small sub-volumes.
  • Lesson. Thermal history is a proxy for porosity — CNNs decode it densely, below the resolution of point pyrometers.

34. Case 13 — Real-Time FSW U-Net at ~25 fps

  • Result. ~25 fps continuous inference; defect area + weld width streamed to closed-loop controller.
  • Lesson. CNNs are now fast enough for in-line process control — not just offline metrology.

35. Case 14 — Radiographic Weld Inspection (CNN-ViT)

End-to-end pipeline of the CNN-ViT weld inspection framework.

End-to-end pipeline. Data → preprocessing → comparative training of (a) CNN baseline and (b) hybrid CNN-ViT (CNN backbone + patch tokenization + transformer encoder) → testing → explainability layer with Grad-CAM and self-attention maps Parmar, Kumar et al., (2026), doi:10.1038/s41598-026-44874-x.

Grad-CAM saliency maps over weld radiographs.

Misclassification analysis with Grad-CAM. Original radiographs (top), expert ROIs (middle), and Grad-CAM heatmaps (bottom) for the four classes — the explainability layer reveals where the network looked when it got the answer wrong Parmar, Kumar et al., (2026), doi:10.1038/s41598-026-44874-x.
  • Result. CNN-ViT 98.56% vs CNN baseline 97.90%; ~31% reduction in misclassification rate.
  • Lesson. Hybrid CNN + ViT = local CNN features + global ViT context, with auditable Grad-CAM evidence per decision — a regulatory-grade design.

36. Case 15 — CNN as Crystal-Plasticity Surrogate

FE vs CNN von-Mises stress fields and error maps for three Voronoi RVEs.

Voronoi grain RVEs — three test microstructures. Row (a) DAMASK FE ground truth \(S_\mathrm{vM}\), (b) CNN prediction, (c) error map; field-level agreement to within ±0.4 GPa Mianroodi, Jaber Rezaei et al., (2021), doi:10.1038/s41524-021-00571-z.

FE vs CNN stress fields for various inclusion shapes.

Inclusion shape generalisation — circle, square, diamond, divided square. Same CNN extrapolates to morphologies it never saw during training Mianroodi, Jaber Rezaei et al., (2021), doi:10.1038/s41524-021-00571-z.
  • Method. 3-D CNN trained on CPFEM ground truth; orders of magnitude faster at inference.
  • Lesson. CNN now a viable surrogate inside design loops — replaces FE inner solves wherever speed matters more than the last fraction of a percent.

37. Case 16 — End-to-End PSP Closure

Two CNNs in series

  • Each CNN trained independently, then chained at inference.
  • Lesson. The PSPP backbone of the course is now fully learnable end-to-end Sandfeld, Stefan et al., (2024).
  • Caveat. Errors compound across the chain — Unit 12 will revisit uncertainty propagation.

§5 · The Pattern + Pitfalls

39. The Common Pattern

All sixteen cases fit:

Raw signal   →  Encoding   →  CNN
   →  Loss   →  Label / property
  • The architecture family is shared.
  • The encoding and deployment unit decide the project.

What still varies

  • \(\mathbf{X}\) — pixels, voxels, spectra, sensor streams.
  • \(y\) — class, mask, scalar, function.
  • Loss — cross-entropy, IoU, MSE, calibrated probabilistic.
  • Split — by specimen / build / day / instrument.

40. CNN vs 2-Point Statistics — When CNN Wins

3-D CNN architecture for microstructure-property regression.

3-D CNN architecture — voxelated microstructure → conv + pool stack → fully-connected head → effective stiffness \(C_{11}\) Mann, Andrew et al., (2022), doi:10.3389/fmats.2022.851085.

Three RVEs and corresponding 2-point statistics maps.

Microstructure ⇄ \(S_2\). Three voxelated two-phase RVEs (top) and their corresponding 2-point auto-correlation maps (bottom) — the classical input that the CNN replaces or complements Mann, Andrew et al., (2022), doi:10.3389/fmats.2022.851085.

CNN wins when

  • Spatial features are task-specific.
  • \(N \gtrsim 10^3\) specimens or simulation augmentation available.

\(S_2\) / MKS still competitive when

41. Specimen Splits Revisited

Invalid protocol

  1. 200 micrographs → 16 crops each → 3200 patches.
  2. Random 80/20 patch split.
  3. Report \(R^2 \approx 0.95\).

Reality

  • Train and test share specimens → correlated rows; metric is optimistic.
  • Specimen-level split on the same labels can collapse \(R^2\) to ~0.72 Sandfeld, Stefan et al., (2024).

Rule. Group ID = whatever is exchangeable at deployment.

42. Cross-Lab Distribution Shift

  • Train on microscope A: \(R^2 \approx 0.88\).
  • Test on microscope B (same alloy): \(R^2 \approx 0.45\).
  • CNN may latch onto contrast / vignetting / detector noise rather than grains.

Mitigations (preview Unit 6)

  • Physics-aware normalisation, harmonised imaging SOPs.
  • Domain randomisation / adaptation when train and deployment labs differ.
  • Saliency / attention checks that the network looks where physics says it should.

43. Class Imbalance on Rare Defects

  • Defect prevalence 2% → “always predict good” → 98% accuracy, 0% recall.
  • Materials goal is usually high recall on the rare class.

Mitigations

  • Use precision / recall / F1 or PR-AUC, not accuracy.
  • Stratified specimen-level splits.
  • Cost-sensitive losses; resample within train only; active labelling of hard negatives.

44. Label Noise from Upstream Segmentation

  • \(\mathbf{x}\) derived from segmentation v1.3; \(y\) from pristine tensile test.
  • Segmentation drift between v1.3 and v1.4 → false aleatory scatter → CNN fits artefacts.

Mitigations

  • Inter-annotator / inter-version study on a subset.
  • Ensemble segmentations; report label variance.
  • Uncertainty-aware losses (preview Unit 12).
  • Version-pin the entire upstream pipeline.

45. Raw-Pixel MLP Failure → CNN Motivation

  • \(1024 \times 1024\) RGB → first dense layer ≈ \(10^9\) weights.
  • \(N \sim 10^2\) specimens: spurious pixel correlations dominate.
  • Optimisation finds coupons, scratches, brightness gradients — not physics.

The Unit 5 punchline

  • CNNs: locality + weight sharing → effective parameter count drops by orders of magnitude.
  • The right inductive bias for spatial data Goodfellow, Ian et al., (2016).

46. When Not to Use a CNN

Use simpler models when

  • \(N\) is small (a few hundred specimens).
  • Inputs are tabular composition + process.
  • Regulatory / safety context demands coefficient-level audit.
  • Extrapolation outside training process window is required.

Practical rule

  • Always fit a serious linear / MLP / tree baseline first.
  • Escalate to CNN only when grouped-CV gain survives stress tests (shift, OOD batches) Sandfeld, Stefan et al., (2024).

Note

“Use the simplest model that survives grouped CV.”

§6 · Bridge & Wrap

47. The MFML Toolkit Applied Here

Forward pass / activations / training loop don’t change.

  • Same \(f_\theta\), same \(J\), same backprop.
  • What MFML proved: this toolkit is flexible.

What changes in materials ML

  • What feeds \(\mathbf{X}\) — pixels, descriptors, \(S_2\), process vector, spectra.
  • What \(y\) means — measurement chain, label noise, calibration.
  • How you split — specimens, batches, instruments.
  • What loss reflects deployment cost.

48. Looking Ahead — Unit 5 (CNNs)

Next week

  • Convolution = locality + weight sharing.
  • Architectures: VGG, ResNet, U-Net, ViT.
  • Where each architecture fits a materials task.

Carry forward from today

  • Specimen splits, normalisation, shift awareness.
  • CNNs multiply debugging surface — they don’t remove obligations Goodfellow, Ian et al., (2016).

Beyond Unit 5

  • Unit 6: transfer learning + domain shift.
  • Unit 12: uncertainty quantification.

49. Reading + Exercises

Reading

Exercises

  1. Reproduce Azimi-style classification on UHCS micrographs (NIST public). Compare a hand-crafted feature pipeline against a small CNN.
  2. Compute binned \(S_2\) on a binary microstructure set; train an MLP on \(S_2\) and a small CNN on raw images; compare grouped-CV scores.
  3. Repeat (2) with deliberate patch-level splitting; quantify how much \(R^2\) inflates vs the specimen-level baseline.

50. Key Takeaways

  • Hand-crafted metrics — interpretable, standardised, lossy by construction.
  • \(S_2\) / MKS / eigen-modes — principled middle ground between scalars and pixels.
  • Sixteen real applications across SEM, EBSD, TEM, XRD, X-ray CT, AM cameras, weld radiographs, RVE simulators — same convolutional toolkit, different encodings and heads.
  • Specimen splits, lab shift, imbalance, segmentation noise — still your responsibility, no matter how deep the network.

Continue

References

Materials data science, Stefan Sandfeld & others
Advanced steel microstructural classification by deep learning methods, Scientific Reports, Seyed Majid Azimi, Dominik Britz, Michael Engstler, Mario Fritz, & Frank Mücklich https://doi.org/10.1038/s41598-018-20037-5
Overview: Computer vision and machine learning for microstructural characterization and analysis, Metallurgical and Materials Transactions A, Elizabeth A. Holm, Ryan Cohn, Nan Gao, Andrew R. Kitahara, Thomas P. Matson, Bo Lei, & Srujana Rao Yarasi https://doi.org/10.1007/s11661-020-06008-4
Machine learning for engineers: Introduction to physics-informed, explainable learning methods for AI in engineering applications, Michael Neuer & others
Machine learning for engineers: Using data to solve problems for physical systems, Ryan G. McClarren
Exploring the microstructure manifold: Image texture representations applied to ultrahigh carbon steel microstructures, Acta Materialia, Brian L. DeCost, Toby Francis, & Elizabeth A. Holm https://doi.org/10.1016/j.actamat.2017.05.014
High throughput quantitative metallography for complex microstructures using deep learning: A case study in ultrahigh carbon steel, Microscopy and Microanalysis, Brian L. DeCost, Bo Lei, Toby Francis, & Elizabeth A. Holm https://doi.org/10.1017/S1431927618015635
Deep learning for automated phase segmentation in EBSD maps. A case study in dual phase steel microstructures, Materials Characterization, T. Martinez Ostormujof, R. R. P. Purushottam Raj Purohit, S. Breumier, N. Gey, M. Salib, & L. Germain https://doi.org/10.1016/j.matchar.2021.111638
A deep learning approach for complex microstructure inference, Nature Communications, Ali Riza Durmaz, Martin Müller, Bo Lei, Akhil Thomas, Dominik Britz, Elizabeth A. Holm, Chris Eberl, Frank Mücklich, & Peter Gumbsch https://doi.org/10.1038/s41467-021-26565-5
Deep learning of crystalline defects from TEM images: A solution for the problem of “never enough training data”, Machine Learning: Science and Technology, Kishan Govind, Daniela Oliveros, Antonin Dlouhy, Marc Legros, & Stefan Sandfeld https://doi.org/10.1088/2632-2153/ad1a4e
Deep learning for semantic segmentation of defects in advanced STEM images of steels, Scientific Reports, Graham Roberts, Simon Y. Haile, Rajat Sainju, Danny J. Edwards, Brian Hutchinson, & Yuanyuan Zhu https://doi.org/10.1038/s41598-019-49105-0
Deep learning approaches for mining structure-property linkages in high contrast composites from simulation datasets, Computational Materials Science, Zijiang Yang, Yuksel C. Yabansu, Reda Al-Bahrani, Wei-keng Liao, Alok N. Choudhary, Surya R. Kalidindi, & Ankit Agrawal https://doi.org/10.1016/j.commatsci.2018.05.014
Modeling structure-property relationships with convolutional neural networks: Yield surface prediction based on microstructure images, International Journal of Plasticity, Julian N. Heidenreich, Maysam B. Gorji, & Dirk Mohr https://doi.org/10.1016/j.ijplas.2022.103506
A deep-learning technique for phase identification in multiphase inorganic compounds using synthetic XRD powder patterns, Nature Communications, Jin-Woong Lee, Woon Bae Park, Jin Hee Lee, Satendra Pal Singh, & Kee-Sun Sohn https://doi.org/10.1038/s41467-019-13749-3
Deep learning-based segmentation of lithium-ion battery microstructures enhanced by artificially generated electrodes, Nature Communications, Simon Müller, Christina Sauter, Ramesh Shunmugasundaram, Nils Wenzler, Vincent De Andrade, Francesco De Carlo, Ender Konukoglu, & Vanessa Wood https://doi.org/10.1038/s41467-021-26480-9
Monitoring of the powder bed quality in metal additive manufacturing using deep transfer learning, Materials & Design, Felix Gabriel Fischer, Max Gero Zimmermann, Niklas Praetzsch, & Christian Knaak https://doi.org/10.1016/j.matdes.2022.111029
A deep learning framework for defect prediction based on thermographic in-situ monitoring in laser powder bed fusion, Journal of Intelligent Manufacturing, Simon Oster, Philipp P. Breese, Alexander Ulbricht, Gunther Mohr, & Simon J. Altenburg https://doi.org/10.1007/s10845-023-02117-0
Machine vision defect segmentation and geometric measurement for real time quality monitoring in friction stir welding, Journal of Manufacturing Systems, Naveen Loganathan, Joel Andersson, & Vivek Patel https://doi.org/10.1016/j.jmsy.2026.01.007
A misclassification-aware explainable hybrid CNN-vision transformer framework for radiographic weld inspection, Scientific Reports, Kumar Parmar, Rituraj Jain, P. T. Anitha, A. Jayanthi, Ramesh Babu Putchanuthala, Kamal Upreti, & Terefe Bayisa Ayele https://doi.org/10.1038/s41598-026-44874-x
Teaching solid mechanics to artificial intelligence—a fast solver for heterogeneous materials, npj Computational Materials, Jaber Rezaei Mianroodi, Nima H. Siboni, & Dierk Raabe https://doi.org/10.1038/s41524-021-00571-z
Development of a robust CNN model for capturing microstructure-property linkages and building property closures supporting material design, Frontiers in Materials, Andrew Mann & Surya R. Kalidindi https://doi.org/10.3389/fmats.2022.851085
Deep learning, Ian Goodfellow, Yoshua Bengio, & Aaron Courville