官术网_书友最值得收藏!

Categorical cross-entropy

Categorical cross-entropy is the most diffused classification cost function, adopted by logistic regression and the majority of neural architectures. The generic analytical expression is:

This cost function is convex and can be easily optimized using stochastic gradient descent techniques; moreover, it has another important interpretation. If we are training a classifier, our goal is to create a model whose distribution is as similar as possible to pdata. This condition can be achieved by minimizing the Kullback-Leibler pergence between the two distributions:

In the previous expression, pM is the distribution generated by the model. Now, if we rewrite the pergence, we get:

The first term is the entropy of the data-generating distribution, and it doesn't depend on the model parameters, while the second one is the cross-entropy. Therefore, if we minimize the cross-entropy, we also minimize the Kullback-Leibler pergence, forcing the model to reproduce a distribution that is very similar to pdata. This is a very elegant explanation as to why the cross-entropy cost function is an excellent choice for classification problems.

主站蜘蛛池模板: 清河县| 吉林省| 房产| 灵武市| 公安县| 揭西县| 内江市| 南部县| 屯昌县| 大名县| 西安市| 巩留县| 高尔夫| 女性| 叶城县| 石景山区| 黔东| 大化| 芮城县| 沙湾县| 天峨县| 民权县| 丹东市| 商南县| 武隆县| 镇江市| 芒康县| 湖北省| 增城市| 海伦市| 阿坝县| 黑山县| 深圳市| 泸州市| 滁州市| 贵州省| 靖州| 江阴市| 安龙县| 横山县| 西藏|