Deep Learning

Undergraduate and Graduate course, University of Siegen, 2025

Winter term 2025 (~ 50 students per semester).

  • Supervised machine learning as an interpolation problem

  • Simple network architectures: Fully connected layers, rectified linear units, sigmoids, softmax

  • Gradient descent for nested functions: The chain rule and it’s implementation via backpropagation

  • Stochastic gradient descent on large data sets, acceleration via momentum and ADAM

  • Capacity, overfitting and underfitting of neural networks

  • Training, testing, and validation data sets

  • Improving generalization: data augmentation, dropout, early stopping

  • Working with images: Convolutions and pooling layers. Computing derivatives and adjoint linear operators

  • Getting the network to train: Data preprocessing, weight initialization schemes, and batch normalization

  • Applications and state-of-the-art architectures for image classification, segmentation, and denoising

  • Architecture designs: Encoder-decoder idea, unrolled algorithms, skip connections + residual learning, recurrent neural networks

  • Implementations in NumPy and PyTorch: Hands-on practical experience by implementing gradient descent on a fully connected network in NumPy.

  • Introduction to the deep learning framework PyTorch for training complex models on GPUs.