ickma.dev
ickma.dev — Notes on Deep Learning and Math
A growing collection of structured study notes and visual explanations — written for clarity, reproducibility, and long-term memory.
Latest Updates
∇ Deep Learning Book 34 chapters
My notes on the Deep Learning book by Ian Goodfellow, Yoshua Bengio, and Aaron Courville.
Chapter 9.6: Structured Outputs CNNs can generate high-dimensional structured objects through pixel-level predictions. Recurrent convolution refines predictions iteratively, producing dense outputs for segmentation, depth estimation, and flow prediction.
Chapter 9.5: Convolutional Functions Mathematical details of convolution operations: deep learning uses cross-correlation (not true convolution), multi-channel formula \(Z_{l,x,y} = \sum_{i,j,k} V_{i,x+j-1,y+k-1} K_{l,i,j,k}\), stride for downsampling, three padding strategies (valid/same/full), and gradient computation.
Chapter 9.4: Convolution and Pooling as an Infinitely Strong Prior Why CNNs work on images but not everywhere: architectural constraints (local connectivity + weight sharing) act as infinitely strong Bayesian priors, assigning probability 1 to translation-equivariant functions and 0 to all others.
📐 MIT 18.06SC Linear Algebra 36 lectures
My journey through MIT’s Linear Algebra course, focusing on building intuition and making connections between fundamental concepts.
Lecture 27: Positive Definite Matrices and Minima Connecting positive definite matrices to multivariable calculus and optimization: the Hessian matrix, second derivative tests, and the geometric interpretation of quadratic forms as ellipsoids.
Lecture 26: Complex Matrices and Fast Fourier Transform Extending linear algebra to complex vectors: Hermitian matrices, unitary matrices, and the Fast Fourier Transform algorithm that reduces DFT complexity from O(N²) to O(N log N).
Lecture 28: Similar Matrices and Jordan Form When matrices share eigenvalues but differ in structure: similar matrices represent the same transformation in different bases, and Jordan form reveals the canonical structure when diagonalization fails.
📐 MIT 18.065 Matrix Methods 1 lecture
My notes from MIT 18.065 Matrix Methods in Data Analysis, Signal Processing, and Machine Learning.
Lecture 8: Norms of Vectors and Matrices Understanding vector p-norms (\(\|v\|_p\)) and when they satisfy the triangle inequality (only for \(p \geq 1\)). The “\(\frac{1}{2}\)-norm” creates non-convex unit balls for strong sparsity.