官术网_书友最值得收藏!

Using regularization to overcome overfitting

In the previous chapter, we saw the diminishing returns from further training iterations on neural networks in terms of their predictive ability on holdout or test data (that is, data not used to train the model). This is because complex models may memorize some of the noise in the data rather than learning the general patterns. These models then perform much worse when predicting new data. There are some methods we can apply to make our model generalize, that is, fit the overall patterns. These are called regularization and aim to reduce testing errors so that the model performs well on new data.

The most common regularization technique used in deep learning is dropout. However, we will also discuss two other regularization techniques that have a basis in regression and deep learning. These two regularization techniques are L1 penalty, which is also known as Lasso, and L2 penalty, which is also known as Ridge.

主站蜘蛛池模板: 读书| 松桃| 泰兴市| 长宁区| 壤塘县| 玉田县| 乐陵市| 南充市| 镇巴县| 景洪市| 西安市| 元阳县| 都安| 桐城市| 松阳县| 宿州市| 讷河市| 平度市| 南昌市| 茌平县| 新巴尔虎右旗| 钟祥市| 大丰市| 调兵山市| 布尔津县| 凤庆县| 若尔盖县| 丹寨县| 靖安县| 白山市| 岳阳市| 仁化县| 丹东市| 洮南市| 杭州市| 孟州市| 获嘉县| 若尔盖县| 马尔康县| 司法| 南木林县|