March 8, 2022, 3:43 p.m. | Remy Lau

Towards Data Science - Medium towardsdatascience.com

Two closely related mathematical formulations widely used in data science, and notes on their implementations in PyTorch

Photo by Claudio Schwarz on Unsplash

TL;DR

  • Negative log-likelihood minimization is a proxy problem to the problem of maximum likelihood estimation.
  • Cross-entropy and negative log-likelihood are closely related mathematical formulations.
  • The essential part of computing the negative log-likelihood is to “sum up the correct log probabilities.”
  • The PyTorch implementations of CrossEntropyLoss and NLLLoss are slightly different in the expected input values. In short, …

cross-entropy data science entropy loss-function machine learning negative pytorch

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne