Mathematical Foundations of AI & ML

Author

Philipp Pelz

Published

January 20, 2026

Other Formats
Keywords

Machine Learning, Artificial Intelligence, Mathematics, Linear Algebra, Probability, Optimization

1 Mathematical Foundations of AI & ML – Unified Syllabus Overview (with ML-PC & MG)

Legend

  • First serious use — concept must be introduced in MFML before being used in ML-PC or MG
  • Reinforcement / application — concept is applied or deepened, but not introduced
  • (R) Refresher — topic was covered in a prior course and is only briefly revisited
  • MFML Mathematical Foundations of AI & ML
  • ML-PC Machine Learning in Materials Processing & Characterization
  • MG Materials Genomics
Week MFML – Mathematical Foundations (revised) ML-PC – ML in Materials Processing & Characterization (revised) MG – Materials Genomics (revised) Exercise (90 min, Python-based) Dependency Logic
1 Learning vs data analysis; models, loss functions, prediction vs explanation Role of ML in processing & characterization; ML vs physics models Role of ML in materials discovery; databases & targets NumPy refresher; vectors, dot products, simple loss (MSE) MFML defines “learning” as optimization, not statistics
2 Linear algebra refresher for learning: covariance, PCA/SVD (R) PCA as a tool for spectra & images (◎) PCA & low-D structure in materials spaces (◎) PCA refresher on known dataset; visualize variance directions PCA assumed known; MFML aligns notation & geometry
3 Regression as loss minimization; linear models revisited Regression as surrogate modeling for processes & properties (★) Regression & correlation in materials datasets (★) Linear regression from scratch via loss minimization Regression reframed explicitly as learning problem
4 Neural networks early: neuron, activations, universal approximation NN regression for materials properties (★) NN models for structure–property relations (★) Single-neuron + activation functions (manual forward pass) MFML must precede any NN usage
5 Backpropagation, gradients, training dynamics NN training stability & convergence (★) NN training pitfalls in materials data (◎) Manual backprop for shallow NN MFML supplies chain rule & gradient flow
6 Loss landscapes, conditioning, optimization behavior Hyperparameters, robustness, convergence issues (★) Model robustness & sensitivity (◎) Gradient descent experiments: learning rate & conditioning Optimization treated as learning dynamics
7 Generalization, bias–variance, regularization Overfitting control in models (★) Limits of high-D regression (★) Overfitting demo: polynomial vs NN models Critical conceptual gate for both applied courses
8 Probabilistic view of learning: noise & likelihood Noise-aware modeling & error propagation (◎) Noise & uncertainty in materials datasets (★) Noise injection; likelihood vs MSE comparison MFML reframes probability for ML
9 Representation learning: learned vs engineered features Feature learning in signals & images (★) Descriptor learning vs hand-crafted features (★) Feature learning with simple NN Transition from classical descriptors
10 Latent spaces: autoencoders & embeddings Compression & anomaly detection in processes (★) Latent materials spaces & embeddings (★) Autoencoder with framework (PyTorch/Keras) Core week for Materials Genomics
11 Unsupervised learning revisited (objectives, not algorithms) Clustering & process drift detection (◎) Clustering vs discovery in materials space (◎) Compare clustering vs AE embeddings Students reinterpret known clustering methods
12 Uncertainty in predictions (aleatoric vs epistemic); Gaussian Processes (conceptual) Trust & confidence in ML-assisted decisions; surrogate models (★) Discovery & screening with uncertainty; exploration vs exploitation (★) Predictive uncertainty: GP regression vs NN ensembles Enables responsible ML & accelerator concepts
13 Physics-informed & constrained learning Physics-informed ML for processes & characterization (★) Physical constraints in materials ML (◎) Constrained NN / penalty-based PINN demo MFML leads constraints & PINN concepts
14 Explainability, limits, scientific trust Integrated case studies & failure modes Limits & ethics of data-driven discovery Mini end-to-end synthesis project All courses converge conceptually

References

Bishop, Christopher M. 2006. Pattern Recognition and Machine Learning by Christopher m. Bishop. Vol. 400. Springer Science+ Business Media, LLC Berlin, Germany:
McClarren, Ryan G.. 2021. Machine Learning for Engineers: Using Data to Solve Problems for Physical Systems. Springer.
Murphy, Kevin P. 2012. Machine Learning: A Probabilistic Perspective. MIT press.
Neuer, Marcus J. 2024. Machine Learning for Engineers: Introduction to Physics-Informed, Explainable Learning Methods for AI in Engineering Applications. Springer Nature.