🤖

Physics Informed Neural Networks (PINNs) Overview

Jun 23, 2025

Overview

This lecture introduces Physics Informed Neural Networks (PINNs), explains their core concepts, advantages, limitations, and extensions, and discusses best practices for their application in physics-based machine learning tasks.

Introduction to PINNs

  • PINNs integrate known physical laws, expressed as partial differential equations (PDEs), into neural network models by adding physics-based loss terms.
  • They use automatic differentiation to calculate derivatives of outputs with respect to space and time inputs.
  • The PINN loss function includes terms for both data fitting and enforcing PDE constraints.

Core Architecture & Training

  • The model predicts physical fields (e.g., velocity, pressure) as functions of space and time.
  • Training uses both real data points and "virtual points"—locations where only the physics loss is evaluated.
  • PINNs can work well even with limited data due to the incorporation of physical laws.
  • The balance between data loss and physics loss is controlled by a hyperparameter.

Applications & Successes

  • Effective for reconstructing flow fields from sparse measurements, such as inferring internal flows in reactors with few sensors.
  • Outperforms naive neural networks (without physics-informed loss) in generalizing beyond the training data.
  • Useful for speeding up inference steps once trained.

Limitations & Cautions

  • Physics is promoted but not strictly enforced; perfect conservation is not guaranteed.
  • Optimization can be stiff and may not generalize to highly chaotic or discontinuous systems (e.g., with shock waves).
  • Tuning the physics-data loss balance is crucial and affects both error and physical accuracy.
  • Some parameter regimes and PDE types remain challenging for standard PINNs.

Extensions & Improvements

  • Fractional PINNs handle PDEs with fractional or integral operators by combining auto-diff and traditional discretization.
  • Delta PINNs use problem geometry (e.g., via Laplace–Beltrami eigenfunctions) to improve solutions on complex domains.
  • Curriculum regularization and sequence-to-sequence learning strategies can improve training robustness.

Key Terms & Definitions

  • PINN (Physics Informed Neural Network) — A neural network that incorporates PDE constraints into its loss function.
  • Auto-differentiation — Automatic computation of derivatives used for PDE constraints in PINNs.
  • Virtual points — Points in the input domain where only the physics loss is checked.
  • Loss function — Objective that combines data mismatch and PDE violation for network training.
  • Fractional PINNs — Extensions of PINNs for equations with fractional or integral operators.
  • Delta PINNs — PINNs that use eigenfunctions of operators to better exploit geometry.
  • Curriculum regularization — Gradually increasing the effect of the physics loss during training.

Action Items / Next Steps

  • Read assigned papers (original PINN papers, extension studies, and failure analysis).
  • Watch recommended tutorial videos and blog posts linked in the course materials.
  • Implement a simple PINN on test data (e.g., the mass-spring-damper example) to explore training and generalization.
  • Experiment with loss function hyperparameters to observe effects on data fit and physics enforcement.