바로가기메뉴

본문 바로가기 주메뉴 바로가기
 
 

logo

  • P-ISSN1738-6764
  • E-ISSN2093-7504
  • KCI

Statistical Analyses of Cross-Entropy Error Function with Probabilistic Target Encoding for Training Neural Networks

INTERNATIONAL JOURNAL OF CONTENTS / INTERNATIONAL JOURNAL OF CONTENTS, (P)1738-6764; (E)2093-7504
2025, v.21 no.1, pp.88-92
오상훈 (목원대학교)

Abstract

Training neural networks with softmax outputs requires assigning target values to output nodes. Due to its simplicity, we often use one-hot encoding, which adopts “one” or “zero” as the desired output values. However, when training neural networks to minimize the cross-entropy error function between the desired and actual output node values, overfitting of neural networks to training samples becomes a significant issue. A probabilistic target encoding has been proposed to mitigate the overfitting. In this paper, we derive the optimal solutions for output nodes that minimize the cross-entropy error function using the probabilistic target encoding. In the extreme case of the probabilistic target encoding, the analysis corresponds to the cross-entropy error function with one-hot encoding. The statistical analyses conducted to derive the optimal solutions provide considerable insights, including the interval of target values for the Bayes classifier.

keywords
Cross-entropy Loss Function, Probabilistic Target Encoding, Optimal Solution of Output Values, Softmax Function, Neural Networks

INTERNATIONAL JOURNAL OF CONTENTS