- Deep Learning with R for Beginners
- Mark Hodnett Joshua F. Wiley Yuxi (Hayden) Liu Pablo Maldonado
- 149字
- 2021-06-24 14:30:43
L2 penalty
The L2 penalty, also known as ridge regression, is similar in many ways to the L1 penalty, but instead of adding a penalty based on the sum of the absolute weights, the penalty is based on the squared weights. This means that larger absolute weights are penalized more. In the context of neural networks, this is sometimes referred to as weight decay. If you examine the gradient of the regularized objective function, there is a penalty such that, at every update, there is a multiplicative penalty to the weights. As for the L1 penalty, although they could be included, biases or offsets are usually excluded from this.
From the perspective of a linear regression problem, the L2 penalty is a modification to the objective function minimized, from ∑(yi - ?i) to ∑(yi - ?i) + λΘ2.
- 數(shù)據(jù)庫基礎(chǔ)與應(yīng)用:Access 2010
- R數(shù)據(jù)科學(xué)實戰(zhàn):工具詳解與案例分析(鮮讀版)
- 數(shù)據(jù)要素五論:信息、權(quán)屬、價值、安全、交易
- 數(shù)據(jù)架構(gòu)與商業(yè)智能
- Hadoop大數(shù)據(jù)開發(fā)案例教程與項目實戰(zhàn)(在線實驗+在線自測)
- Chef Essentials
- Oracle高性能SQL引擎剖析:SQL優(yōu)化與調(diào)優(yōu)機(jī)制詳解
- Visual FoxPro數(shù)據(jù)庫技術(shù)基礎(chǔ)
- Scratch 2.0 Game Development HOTSHOT
- 大數(shù)據(jù)技術(shù)體系詳解:原理、架構(gòu)與實踐
- 大數(shù)據(jù)計算系統(tǒng)原理、技術(shù)與應(yīng)用
- Artificial Intelligence for Big Data
- 深入理解Flink:實時大數(shù)據(jù)處理實踐
- Learn Selenium
- Creating Mobile Apps with Appcelerator Titanium