13 - Backpropogation - lec. 18
ucla | CS M146 | 2023-05-31T13:59
Table of Contents
Supplemental
- the purpose of backprop - to calculat gradients for weights and perform gradient descent
- uptream gradient = local gradient * downstream gradient
- gradient of max function = 0 if max=0 or 1 if max=x
Lecture
Forward and Backward Pass
Architecture (DAG)
Backprop (DAG) - Scalar
Simple DAG
Common Gate Reduction
Torch Autograd
Backprop (Tensors) - Vector
Vector (Matrix) Derivatives
Jacobian Backprop - implicit
- must be computed implicitly (diagonally) as constructing jacobians for large data causes memory explosion
Matrix Multiplication - implicit Jacobian
Another viewpoint
- matrix mult is associative so w can do chain rule
Discussion
Resources
📌
**SUMMARY
**