ml:loss_functions
Differences
This shows you the differences between two versions of the page.
| Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
| ml:loss_functions [2023/10/27 21:07] – jmflanig | ml:loss_functions [2024/07/23 00:32] (current) – jmflanig | ||
|---|---|---|---|
| Line 11: | Line 11: | ||
| * Lots of different ways to write this loss function. | * Lots of different ways to write this loss function. | ||
| * The cross-entropy version writes it as $L(\mathcal{D}) = -\sum_{i=1}^{N}\sum_{y} p(y|x_i) log(p_\theta(y|x_i))$, | * The cross-entropy version writes it as $L(\mathcal{D}) = -\sum_{i=1}^{N}\sum_{y} p(y|x_i) log(p_\theta(y|x_i))$, | ||
| + | * Cross-entropy loss can be written as | ||
| + | \[ | ||
| + | L(\theta, | ||
| + | \] | ||
| + | * This is often call the Conditional Random Field (CRF) loss | ||
| * The minimum of cross-entropy loss does not always exist, and does not exist if the data training data can be completely separated. | * The minimum of cross-entropy loss does not always exist, and does not exist if the data training data can be completely separated. | ||
| * Perceptron loss \[ | * Perceptron loss \[ | ||
| Line 21: | Line 26: | ||
| * [[https:// | * [[https:// | ||
| * [[https:// | * [[https:// | ||
| + | * The softmax margin loss is obtained by replacing the max in the SVM loss with a softmax: \[ | ||
| + | L(\theta, | ||
| + | \] | ||
| * Risk\[ | * Risk\[ | ||
| L(\theta, | L(\theta, | ||
ml/loss_functions.1698440878.txt.gz · Last modified: 2023/10/27 21:07 by jmflanig