data_mining:neural_network:loss_functions

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revisionPrevious revision
Next revision
Previous revision
Last revisionBoth sides next revision
data_mining:neural_network:loss_functions [2019/11/10 00:10] – [Binary cross entropy] phreazerdata_mining:neural_network:loss_functions [2019/11/10 00:31] – [Binary cross entropy] phreazer
Line 6: Line 6:
  
 $-(y log(\hat{y}) + (1-y) log(1-\hat{y})$ $-(y log(\hat{y}) + (1-y) log(1-\hat{y})$
 +
 +or
 +
 +$-1/N \sum_{i=1}^N (y_i log(\hat{y_i}) + (1-y_i) log(1-\hat{y_i})$
  
 (Negative log, since log in [0,1] is < 0) (Negative log, since log in [0,1] is < 0)
Line 34: Line 38:
  
 For an algo we need to find closes $p$ for $q$ by minimizing cross-entropy. For an algo we need to find closes $p$ for $q$ by minimizing cross-entropy.
 +
 +In training we have $N$ samples. For one particular example the distribution is known to which class it belongs. The loss function should minimize the average cross-entropy.
  • data_mining/neural_network/loss_functions.txt
  • Last modified: 2019/11/10 00:40
  • by phreazer