!!! Overview [{$pagename}] !! [{$pagename}] in [Machine Learning] [{$pagename}] in [Machine Learning] penalizes the [weight] when it is too large.For [{$pagename}] with [Logistic Regression], you try to minimize the [Cost function] add lambda (λ), which is called the [regularization parameter] which is usually determined form the development set. [Regularization/l2-term.png] Above shows the L2 [{$pagename}] formula and then the [{$pagename}] [Parameter] added to the [Cost function] "__J__"L2 [{$pagename}] formula is just the square Euclidean norm of the prime to [vector] w which is called L^2 [{$pagename}] which is the most common. Two popular examples of [{$pagename}] methods for [Linear Regression] are: * [LASSO] Regression * [Ridge Regression] * [Dropout Regularization] These methods are effective to use when there is collinearity in your input values and [Ordinary Least Squares] would cause [Overfitting] the [Training dataset]. !! Misc Notes The [bias] is generally not [{$pagename}]. In [Python] lambda is a reserved word so often it is used as lambd. !! More Information There might be more information for this subject on one of the following: [{ReferringPagesPlugin before='*' after='\n' }] ---- * [#1] - [Frobenius Norm|http://mathworld.wolfram.com/FrobeniusNorm.html|target='_blank'] - based on information obtained 2018-01-03