Calemsy / Machine-Learning-2017-Fall

Hung-yi Lee
1 stars 0 forks source link

Part 8. BackPropagation #8

Open Calemsy opened 5 years ago

Calemsy commented 5 years ago

1 - Chain Rule

2 - BackPropagation

因为总的损失是各个样本损失的和:$L(\theta) = \sum{n=1}^{N}l^n(\theta)$,所以总的损失对参数的偏导是各个样本上损失对参数偏导的求和: $$\frac{\partial L(\theta)}{\partial w}=\sum{n=1}^{N}\frac{\partial l^n(\theta)}{\partial w}$$

如下图: $$\frac{\partial l}{\partial w} = \frac{\partial l}{\partial z}\frac{\partial z}{\partial w}$$ 其中

image

2.1 - Forward pass

$$\frac{\partial z}{\partial w_1} = x_1, \frac{\partial z}{\partial w_2} = x_2$$

image

2.2 - Backward pass

$$\frac{\partial l}{\partial z} = ?$$

image

$$\frac{\partial l}{\partial z} = \frac{\partial l}{\partial a}\frac{\partial a}{\partial z}$$

$$\frac{\partial l}{\partial a} = \frac{\partial z'}{\partial a}\frac{\partial l}{\partial z'} + \frac{\partial z''}{\partial a}\frac{\partial l}{\partial z''} =w_3 \frac{\partial l}{\partial z'} + w_4 \frac{\partial l}{\partial z''}$$

假设:$\frac{\partial l}{\partial z'}$和$\frac{\partial l}{\partial z''}$是已知的,那么$\frac{\partial l}{\partial z}$的计算公式如下以及下图所示:

$$\frac{\partial l}{\partial z} = \frac{\partial a}{\partial z}[w_3 \frac{\partial l}{\partial z'} + w_4 \frac{\partial l}{\partial z''}] = \sigma'(z)[w_3 \frac{\partial l}{\partial z'} + w_4 \frac{\partial l}{\partial z''}]$$

image

上面的计算基于$\frac{\partial l}{\partial z'}$和$\frac{\partial l}{\partial z''}$是已知的,那么$\frac{\partial l}{\partial z'}$和$\frac{\partial l}{\partial z''}$如何计算呢?

3 - Summary

image