Part III: Neural Networks
Neural networks are universal function approximators built from composing simple parametric transformations. This part derives everything from scratch: the perceptron learning rule, the full backpropagation algorithm, the engineering innovations that make deep networks trainable, and the convolutional inductive bias that powered the deep learning revolution in vision.
Chapter 7: Perceptrons & Backpropagation
From the single perceptron to multilayer networks — full derivation of the backpropagation algorithm via computational graphs and the chain rule.
Chapter 8: Deep Neural Networks
Activation functions, vanishing gradients, Batch Normalisation, Dropout, residual connections and weight initialisation — the engineering science of depth.
Chapter 9: Convolutional Neural Networks
Discrete convolution, parameter sharing, translation equivariance, pooling and modern architectures from LeNet to EfficientNet.
What you will learn
Prerequisites
Part I (linear algebra, calculus, probability) and Part II (supervised learning, gradient descent). You should be comfortable with matrix calculus and the concept of a loss function before beginning.