官术网_书友最值得收藏!

  • Hands-On Neural Networks
  • Leonardo De Marchi Laura Mitchell
  • 142字
  • 2021-06-24 14:00:15

Sigmoid

As you have already seen, the sigmoid function is a particular case of the logistic function, and it gives us something similar to a step function; therefore, it's useful for binary classifications, indicating a probability as a result. The function is differentiable; therefore, we can run gradient descent for every point. It's also monotonic, which means that it always increases or decreases, but its derivative does not; therefore, it will have a minima. It forces all output values to be between 0 and 1. Because of this, even very high values asymptotically tend to one and very low to 0. One problem that this creates is that the derivative in those points is approximately 0; therefore, the gradient descent process will not find a local minima for very high or very low values, as shown in the following diagram:

主站蜘蛛池模板: 肃南| 从江县| 勃利县| 咸丰县| 鄯善县| 斗六市| 东方市| 尚志市| 且末县| 马关县| 沙洋县| 论坛| 新丰县| 定结县| 登封市| 天台县| 普格县| 尚义县| 开平市| 丰顺县| 子长县| 滁州市| 梅州市| 夹江县| 乡宁县| 五原县| 蓬莱市| 临武县| 连城县| 济阳县| 新泰市| 余姚市| 灵石县| 隆德县| 资源县| 蕉岭县| 新绛县| 石棉县| 宁乡县| 辽源市| 礼泉县|