官术网_书友最值得收藏!

Activation functions

The activation function in each artificial neuron decides whether the incoming signals have reached the threshold and should output signals for the next level. It is crucial to set up the right activation function because of the gradient vanishing issue, which we will talk about later.

Another important feature of an activation function is that it should be differentiable. The network learns from the errors that are calculated at the output layer. A differentiable activation function is needed to perform backpropagation optimization while propagating backwards in the network to compute gradients of error (loss) with respect to weights, and then optimize weights accordingly, using gradient descent or any other optimization technique to reduce the error.

The following table lists a few common activation functions. We will dive into them a bit deeper, talk about the differences between them, and explain how to choose the right activation function:

主站蜘蛛池模板: 贵德县| 葵青区| 定兴县| 青川县| 宁远县| 乐陵市| 灵璧县| 屯留县| 甘肃省| 商洛市| 珠海市| 长春市| 北海市| 德钦县| 全椒县| 祁阳县| 股票| 邢台市| 泌阳县| 鄯善县| 特克斯县| 肥西县| 滦南县| 汨罗市| 沭阳县| 五家渠市| 阿鲁科尔沁旗| 保德县| 伊宁县| 柳林县| 沁源县| 大名县| 揭东县| 上思县| 美姑县| 军事| 会东县| 黎平县| 凯里市| 丹凤县| 堆龙德庆县|