Deep Learning
Undergraduate and Graduate course, University of Siegen, 2025
Winter term 2025 (~ 50 students per semester).
-
Supervised machine learning as an interpolation problem
-
Simple network architectures: Fully connected layers, rectified linear units, sigmoids, softmax
-
Gradient descent for nested functions: The chain rule and it’s implementation via backpropagation
-
Stochastic gradient descent on large data sets, acceleration via momentum and ADAM
-
Capacity, overfitting and underfitting of neural networks
-
Training, testing, and validation data sets
-
Improving generalization: data augmentation, dropout, early stopping
-
Working with images: Convolutions and pooling layers. Computing derivatives and adjoint linear operators
-
Getting the network to train: Data preprocessing, weight initialization schemes, and batch normalization
-
Applications and state-of-the-art architectures for image classification, segmentation, and denoising
-
Architecture designs: Encoder-decoder idea, unrolled algorithms, skip connections + residual learning, recurrent neural networks
-
Implementations in NumPy and PyTorch: Hands-on practical experience by implementing gradient descent on a fully connected network in NumPy.
-
Introduction to the deep learning framework PyTorch for training complex models on GPUs.