官术网_书友最值得收藏!

Training neural networks with backpropagation

Calculating the activation of a neuron, the forward part, or what we call feed-forward propagation, is quite straightforward to process. The complexity we encounter now is training the errors back through the network. When we train the network now, we start at the last output layer and determine the total error, just as we did with a single perceptron, but now we need to sum up all errors across the output layer. Then we need to use this value to backpropagate the error back through the network, updating each of the weights based on their contribution to the total error. Understanding the contribution of a single weight in a network with thousands or millions of weights could be quite complicated, except thankfully for the help of differentiation and the chain rule. Before we get to the complicated math, we first need to discuss the Cost function and how we calculate errors in the next section.

While the math of backpropagation is complicated and may be intimidating, at some point, you will want or need to understand it well. However, for the purposes of this book, you can omit or just revisit this section as needed. All the networks we develop in later chapters will automatically handle backpropagation for us. Of course, you can't run away from the math either; it is everywhere in deep learning.
主站蜘蛛池模板: 湾仔区| 南汇区| 陈巴尔虎旗| 兰考县| 区。| 南召县| 花莲市| 南宁市| 正蓝旗| 张家界市| 秦皇岛市| 横峰县| 水富县| 乌审旗| 永兴县| 华容县| 兖州市| 华容县| 蕉岭县| 馆陶县| 洪洞县| 芜湖县| 弋阳县| 枞阳县| 怀仁县| 清苑县| 新绛县| 福贡县| 石首市| 巴东县| 海宁市| 同仁县| 信宜市| 集贤县| 岳普湖县| 富平县| 东山县| 西乡县| 神农架林区| 开鲁县| 宜春市|