官术网_书友最值得收藏!

How to choose the right activation function

The activation function is decided depending upon the objective of the problem statement and the concerned properties. Some of the inferences are as follows:

  • Sigmoid functions work very well in the case of shallow networks and binary classifiers. Deeper networks may lead to vanishing gradients.

  • The ReLU function is the most widely used, and try using Leaky ReLU to avoid the case of dead neurons. Thus, start with ReLU, then move to another activation function if ReLU doesn't provide good results.

  • Use softmax in the outer layer for the multi-class classification.

  • Avoid using ReLU in the outer layer.

主站蜘蛛池模板: 洞头县| 南充市| 巨野县| 安义县| 卢湾区| 江西省| 唐海县| 湖南省| 屯门区| 临泉县| 防城港市| 安吉县| 鄂州市| 桂平市| 元江| 元朗区| 比如县| 岳普湖县| 乌拉特前旗| 措勤县| 南郑县| 望都县| 伊宁市| 濮阳市| 靖西县| 驻马店市| 永川市| 辉县市| 肥城市| 宁武县| 吉安县| 建昌县| 舟曲县| 云阳县| 靖宇县| 仁寿县| 临桂县| 瑞安市| 麦盖提县| 桂林市| 库尔勒市|