- Deep Learning with PyTorch
- Vishnu Subramanian
- 138字
- 2021-06-24 19:16:27
Sigmoid
The sigmoid activation function has a simple mathematical form, as follows:
The sigmoid function intuitively takes a real-valued number and outputs a number in a range between zero and one. For a large negative number, it returns close to zero and, for a large positive number, it returns close to one. The following plot represents different sigmoid function outputs:

The sigmoid function has been historically used across different architectures, but in recent times it has gone out of popularity as it has one major drawback. When the output of the sigmoid function is close to zero or one, the gradients for the layers before the sigmoid function are close to zero and, hence, the learnable parameters of the previous layer get gradients close to zero and the weights do not get adjusted often, resulting in dead neurons.
- 計算機組裝與系統(tǒng)配置
- Apple Motion 5 Cookbook
- Hands-On Machine Learning with C#
- Practical Machine Learning with R
- 超大流量分布式系統(tǒng)架構(gòu)解決方案:人人都是架構(gòu)師2.0
- Managing Data and Media in Microsoft Silverlight 4:A mashup of chapters from Packt's bestselling Silverlight books
- Spring Cloud實戰(zhàn)
- 筆記本電腦芯片級維修從入門到精通(圖解版)
- FreeSWITCH Cookbook
- Spring Security 3.x Cookbook
- 計算機電路基礎(chǔ)(第2版)
- 微服務(wù)實戰(zhàn)(Dubbox +Spring Boot+Docker)
- ARM接口編程
- ActionScript Graphing Cookbook
- Spring Cloud微服務(wù)架構(gòu)開發(fā)