Cross Entropy
- Entropy
- Cross-Entropy is a measure from the field of information theory, building upon Entropy and generally calculating the difference between two Probability Distributions.
- It is closely related to but is different from KL Divergence that calculates the relative Entropy between two Probability Distributions, whereas cross-Entropy can be thought to calculate the total Entropy between the Distributions.
- implicit distribution $$ p(Y|x;\theta)
- Categorical CE
- 
- Classification
- Labels should be [One hot](One hot.md)
- $$
\mathscr{L}(\theta) = -\mathbb{E}{(x,y) \sim P(X,Y)} \Sigma{i=1}^C 1(y=i)log (p_{model}f_i(x|\theta))
- C is no of classes - $$ L(y, \hat y) = - \Sigma_{i}\Sigma_{c}y_{i}^{c} log(\hat y_{i}^{c})- [MSE](MSE.md)
- Regression
- $$
\mathscr{L}(\theta) = \frac{1}{2}\mathbb{E}_{(x,y) \sim P(X,Y)}||y-f(x;\theta)||
- [Binary Cross Entropy](Binary%20Cross%20Entropy.md) - The Cross Entropy Loss function is a popular loss function that is used in multi-class image classification tasks. Derived from the field of information theory, it uses the concept of entropy to quantifies the discrepancy between two given probability distributions. The formula for computing the loss is given by $$ \mathscr{l}(x,y) = L = \{l_{1}, ..., l_{N}\}^{T} $$ where $$ l_{n} = -w_{y_{n}}log \frac{exp(x_{n, y_{n}})}{\Sigma_{c=1}^{C}exp(x_{n,c})} $$, $x$ is the input, $y$ is the target, $C$ is the number of classes -  -  - Also [Negative Log Likelihood](Negative%20Log%20Likelihood.md)