官术网_书友最值得收藏!

Ridge regression (L2)

Due to ridge regression, we need to make some changes to the loss function. The original loss function gets added by a shrinkage component:

Now, this modified loss function needs to be minimized to adjust the estimates or coefficients. Here, the lambda is tuning the parameter that regularizes the loss function. That is, it decides how much it should penalize the flexibility of the model. The flexibility of the model is dependent on the coefficients. If the coefficients of the model go up, the flexibility also goes up, which isn't a good sign for our model. Likewise, as the coefficients go down, the flexibility is restricted and the model starts to perform better. The shrinkage of each estimated parameter makes the model better here, and this is what ridge regression does. When lambda keeps going higher and higher, that is, λ → ∞, the penalty component rises, and the estimates start shrinking. However, when λ → 0, the penalty component decreases and starts to become an ordinary least square (OLS) for estimating unknown parameters in a linear regression.

主站蜘蛛池模板: 黑水县| 新宾| 通化县| 浦县| 兰考县| 赣州市| 灌南县| 崇义县| 铁力市| 根河市| 临洮县| 方城县| 四子王旗| 庆阳市| 龙山县| 方正县| 清涧县| 垦利县| 丘北县| 杂多县| 招远市| 四会市| 扬州市| 广东省| 大埔县| 汕尾市| 荆门市| 札达县| 陵川县| 利辛县| 巴南区| 平乐县| 湖北省| 茂名市| 宝坻区| 赤城县| 永丰县| 四子王旗| 阿拉善右旗| 新巴尔虎左旗| 五大连池市|