Differences
This shows you the differences between two versions of the page.
Both sides previous revisionPrevious revision | |||
data_mining:neural_network:overfitting [2018/05/10 15:55] – [Inverted dropout] phreazer | data_mining:neural_network:overfitting [2018/05/10 16:03] (current) – [Inverted dropout] phreazer | ||
---|---|---|---|
Line 139: | Line 139: | ||
==== Inverted dropout ==== | ==== Inverted dropout ==== | ||
- | Layer $l=3$. | + | < |
+ | Layer l=3 | ||
- | $keep.prob = 0.8$ // probability that unit will be kept | + | keep.prob = 0.8 // probability that unit will be kept |
- | $d3 = np.random.rand(a3.shape[0], | + | d3 = np.random.rand(a3.shape[0], |
- | $a3 = np.multiply(a3, | + | a3 = np.multiply(a3, |
- | $a3 /= keep.prob$ // e.g. 50 units => 10 units shut off | + | a3 /= keep.prob // e.g. 50 units => 10 units shut off |
- | + | ||
- | $Z = Wa+b$ // reduced by 20% => standardize with 0.8 => expected value stays the same | + | |
+ | Z = Wa+b // reduced by 20% => standardize with 0.8 => expected value stays the same | ||
+ | </ | ||
Making predictions at test time: No drop out | Making predictions at test time: No drop out | ||