![Hinge loss gives accuracy 1 but cross entropy gives accuracy 0 after many epochs, why? - PyTorch Forums Hinge loss gives accuracy 1 but cross entropy gives accuracy 0 after many epochs, why? - PyTorch Forums](https://discuss.pytorch.org/uploads/default/original/2X/2/2ad4119a40ee6e24f006aabae0f6d0981a20a9cf.png)
Hinge loss gives accuracy 1 but cross entropy gives accuracy 0 after many epochs, why? - PyTorch Forums
![Understanding Categorical Cross-Entropy Loss, Binary Cross-Entropy Loss, Softmax Loss, Logistic Loss, Focal Loss and all those confusing names Understanding Categorical Cross-Entropy Loss, Binary Cross-Entropy Loss, Softmax Loss, Logistic Loss, Focal Loss and all those confusing names](https://gombru.github.io/assets/cross_entropy_loss/intro.png)
Understanding Categorical Cross-Entropy Loss, Binary Cross-Entropy Loss, Softmax Loss, Logistic Loss, Focal Loss and all those confusing names
![NT-Xent (Normalized Temperature-Scaled Cross-Entropy) Loss Explained and Implemented in PyTorch | by Dhruv Matani | Towards Data Science NT-Xent (Normalized Temperature-Scaled Cross-Entropy) Loss Explained and Implemented in PyTorch | by Dhruv Matani | Towards Data Science](https://miro.medium.com/v2/resize:fit:1189/1*bXxtQ57ZoLNduhom2Fp2ew.png)
NT-Xent (Normalized Temperature-Scaled Cross-Entropy) Loss Explained and Implemented in PyTorch | by Dhruv Matani | Towards Data Science
![Why Softmax not used when Cross-entropy-loss is used as loss function during Neural Network training in PyTorch? | by Shakti Wadekar | Medium Why Softmax not used when Cross-entropy-loss is used as loss function during Neural Network training in PyTorch? | by Shakti Wadekar | Medium](https://miro.medium.com/v2/resize:fit:469/1*8Kvne7teaEVoq5X78DyRMA.png)
Why Softmax not used when Cross-entropy-loss is used as loss function during Neural Network training in PyTorch? | by Shakti Wadekar | Medium
![neural network - Why is the implementation of cross entropy different in Pytorch and Tensorflow? - Stack Overflow neural network - Why is the implementation of cross entropy different in Pytorch and Tensorflow? - Stack Overflow](https://i.stack.imgur.com/e6gKc.png)
neural network - Why is the implementation of cross entropy different in Pytorch and Tensorflow? - Stack Overflow
![Hinge loss gives accuracy 1 but cross entropy gives accuracy 0 after many epochs, why? - PyTorch Forums Hinge loss gives accuracy 1 but cross entropy gives accuracy 0 after many epochs, why? - PyTorch Forums](https://discuss.pytorch.org/uploads/default/original/2X/4/4ac4609aee5df6b686796f7490dcd242f463fe5b.png)