Artificial Intelligence Asked by Ali KHalili on September 27, 2020
Why L2 loss is more commonly used in Neural Networks than other loss functions?
What is the reason to L2 being a default choice in Neural Networks?
I'll cover both L2 regularized loss, as well as Mean-Squared Error (MSE):
MSE:
L2 Regularization:
Using L2 regularization is equivalent to invoking a Gaussian prior (see https://stats.stackexchange.com/questions/163388/why-is-the-l2-regularization-equivalent-to-gaussian-prior) on your model/estimator. If modeling your problem as a Maximum A Posteriori Inference (MAP) problem, if your likelihood model (p(y|x)) is Gaussian, then your posterior distribution over parameters (p(x|y)) will also be Gaussian. From Wikipedia: "If the likelihood function is Gaussian, choosing a Gaussian prior over the mean will ensure that the posterior distribution is also Gaussian" (source: https://en.wikipedia.org/wiki/Conjugate_prior).
As in the case above, L2 loss is continuously-differentiable across any domain, unlike L1 loss.
Correct answer by Ryan Sander on September 27, 2020
1 Asked on January 5, 2022 by thepacker
1 Asked on January 5, 2022 by huzaifah-shamim
0 Asked on January 1, 2022
computer vision convolution geometric deep learning graph neural networks
1 Asked on December 30, 2021
machine learning natural language processing neural networks pattern recognition
1 Asked on December 30, 2021
1 Asked on December 30, 2021 by nim-py
0 Asked on December 30, 2021
0 Asked on December 30, 2021
classification convolutional neural networks image segmentation representation learning
1 Asked on December 27, 2021 by bestr
actor critic methods neural networks optimization pytorch reinforcement learning
1 Asked on December 27, 2021 by jaeger6
comparison l1 regularization l2 regularization machine learning regularization
0 Asked on December 27, 2021
computer vision deep learning neural networks papers terminology
1 Asked on December 27, 2021 by kao
deep learning hyper parameters hyperparameter optimization neural networks training
2 Asked on December 25, 2021 by vesko-vujovic
0 Asked on December 25, 2021 by beinando
0 Asked on December 25, 2021
convolution deep learning geometric deep learning graph neural networks
2 Asked on December 22, 2021
machine learning natural language processing optical character recognition pattern recognition
1 Asked on December 20, 2021
classification convolutional neural networks image recognition
1 Asked on December 18, 2021
deep learning generative adversarial networks machine learning neural networks recurrent neural networks
Get help from others!
Recent Questions
Recent Answers
© 2022 AnswerBun.com. All rights reserved. Sites we Love: PCI Database, MenuIva, UKBizDB, Menu Kuliner, Sharing RPP