Published onMay 31, 2025(Preview) Cross Entropy From First Principlesmachine-learninginformation-theorylanguage-modelsCross entropy and Kullback-Leibler (KL) divergence are key concepts in training machine learning models. In this post, we derive these concepts from first principles.