About 7,100,000 results
Open links in new tab
  1. python - `CrossEntropyLoss ()` in PyTorch - Stack Overflow

    The combination of nn.LogSoftmax and nn.NLLLoss is equivalent to using nn.CrossEntropyLoss. This terminology is a particularity of PyTorch, as the nn.NLLoss [sic] computes, in fact, the cross entropy …

  2. machine learning - What is cross-entropy? - Stack Overflow

    Cross entropy is one out of many possible loss functions (another popular one is SVM hinge loss). These loss functions are typically written as J (theta) and can be used within gradient descent, which …

  3. Comparing MSE loss and cross-entropy loss in terms of convergence

    Mar 16, 2018 · The point is that the cross-entropy and MSE loss are the same. The modern NN learn their parameters using maximum likelihood estimation (MLE) of the parameter space.

  4. python - How to correctly use Cross Entropy Loss vs Softmax for ...

    Cross Entropy H (p, q) Cross-entropy is a function that compares two probability distributions. From a practical standpoint it's probably not worth getting into the formal motivation of cross-entropy, though …

  5. Trying to understand cross_entropy loss in PyTorch

    Jul 23, 2019 · This is a very newbie question but I'm trying to wrap my head around cross_entropy loss in Torch so I created the following code: x = torch.FloatTensor([ [1.,0.,0.] ...

  6. In which cases is the cross-entropy preferred over the mean squared ...

    Apr 24, 2017 · Although both of the above methods provide a better score for the better closeness of prediction, still cross-entropy is preferred. Is it in every case or there are some peculiar scenarios …

  7. Cross Entropy Calculation in PyTorch tutorial - Stack Overflow

    As far as I know, the calculation of cross-entropy usually used between two tensors like: Target as [0,0,0,1], where 1 is the right class Output tensor as [0.1,0.2,0.3,0.4], where the sum as 1. So based …

  8. cross entropy - PyTorch LogSoftmax vs Softmax for CrossEntropyLoss ...

    Dec 8, 2020 · I understand that PyTorch's LogSoftmax function is basically just a more numerically stable way to compute Log(Softmax(x)). Softmax lets you convert the output from a Linear layer into …

  9. python - What are logits? What is the difference between softmax and ...

    The cross entropy is a summary metric: it sums across the elements. The output of tf.nn.softmax_cross_entropy_with_logits on a shape [2,5] tensor is of shape [2,1] (the first dimension …

  10. How to choose cross-entropy loss in TensorFlow? - Stack Overflow

    Classification problems, such as logistic regression or multinomial logistic regression, optimize a cross-entropy loss. Normally, the cross-entropy layer follows the softmax layer, which produces