Show pageOld revisionsBacklinksBack to top This page is read only. You can view the source, but not change it. Ask your administrator if you think this is wrong. ====== NN initialization ====== ===== Random initialization ===== Weights need to be randomly initialized. For bias zero is ok. If weights are zero: in backprop => $dz_{1,2}$ are the same. Hidden units would compute same function (= are symmetric). Solution: $W^{[i]}=np.random.randn((2,2)) * 0.01$ $0.01$ because else we would end up at ends of activation function values (and slopes would be small), e.g. if values would be large. data_mining/neural_network/initialization.txt Last modified: 2017/08/19 20:38by phreazer