Listen "Matrix Calculus in Deep Learning"
Episode Synopsis
Matrix calculus is essential for understanding and implementing deep learning. It provides the mathematical tools to optimize neural networks using gradient descent. The Jacobian matrix, a key concept, organizes partial derivatives of vector-valued functions. The vector chain rule simplifies derivative calculations in nested functions, common in neural networks. Automatic differentiation, used in modern libraries, relies on these principles. Grasping matrix calculus allows for a deeper understanding of model training and the implementation of custom neural networks.
More episodes of the podcast Large Language Model (LLM) Talk
Kimi K2
22/07/2025
Mixture-of-Recursions (MoR)
18/07/2025
MeanFlow
10/07/2025
Mamba
10/07/2025
LLM Alignment
14/06/2025
Why We Think
20/05/2025
Deep Research
12/05/2025
vLLM
04/05/2025
Qwen3: Thinking Deeper, Acting Faster
04/05/2025
ZARZA We are Zarza, the prestigious firm behind major projects in information technology.