Different regularization

Different regularization methods have different effects on the learning process.

For example,

L2 regularization penalizes high weight values.

L1 regularization penalizes weight values that do not equal zero.

Adding noise to the weights during learning ensures that the learned hidden representations take extreme values.

Sampling the hidden representations regularizes the network by pushing the hidden representation to be binary during the forward pass which limits the modeling capacity of the network.

最后編輯于
?著作權歸作者所有,轉載或內容合作請聯系作者
平臺聲明:文章內容(如有圖片或視頻亦包括在內)由作者上傳并發布,文章內容僅代表作者本人觀點,簡書系信息發布平臺,僅提供信息存儲服務。

推薦閱讀更多精彩內容