官术网_书友最值得收藏!

  • Deep Learning with R for Beginners
  • Mark Hodnett Joshua F. Wiley Yuxi (Hayden) Liu Pablo Maldonado
  • 149字
  • 2021-06-24 14:30:43

L2 penalty

The L2 penalty, also known as ridge regression, is similar in many ways to the L1 penalty, but instead of adding a penalty based on the sum of the absolute weights, the penalty is based on the squared weights. This means that larger absolute weights are penalized more. In the context of neural networks, this is sometimes referred to as weight decay. If you examine the gradient of the regularized objective function, there is a penalty such that, at every update, there is a multiplicative penalty to the weights. As for the L1 penalty, although they could be included, biases or offsets are usually excluded from this.

From the perspective of a linear regression problem, the L2 penalty is a modification to the objective function minimized, from ∑(yi - ?i) to ∑(yi - ?iλΘ2.

主站蜘蛛池模板: 隆林| 额敏县| 华亭县| 观塘区| 沈丘县| 南京市| 贡山| 馆陶县| 瑞安市| 三门峡市| 贵德县| 芦溪县| 库尔勒市| 商城县| 锦屏县| 莱州市| 昌宁县| 洱源县| 通江县| 寻甸| 民权县| 威海市| 团风县| 建瓯市| 宁阳县| 肥东县| 当涂县| 宿州市| 盖州市| 玛曲县| 高台县| 通城县| 富裕县| 泰和县| 阳城县| 宜君县| 年辖:市辖区| 泰顺县| 蓝山县| 西昌市| 宜城市|