Regularization - AshokBhat/ml GitHub Wiki

Description

  • Techniques used to avoid overfitting problem during training
  • Makes the fitted function smoother

Most widely used techniques

  • L1 regularization - Penalize weights in proportion to the sum of the absolute values of the weights
  • L2 regularization - Penalize weights in proportion to the sum of the squares of the weights
  • Dropout - randomly set a fraction of neurons to 0 at each training iteration
  • Weight decay

FAQ

See also