TransWikia.com

Why do increasing regularization weights make objective function not monotonically decrease?

Cross Validated Asked on November 29, 2021

I run modified non-negative matrix factorization (NMF) and tune the regularization weight from 1e5 to 1e13.

The table below shows errors from calculating cost function for 25 iterations of all weights.

enter image description here

At weight 1e4, the objective function is monotonically decreased. But when I increase the weight more and more, the error profile of 25 iterations goes down and gradually up. What is happened here?
By the way, I think I got the best result is at weight 5e10 but the error profile is not monotonically decreased. In this case, which weight I should choose?

One Answer

Forgive me for not being an expert on NMF, but I am pretty experienced with other regularized techniques.

Regularization techniques usually use a form of gradient descent to find the optimal answer. In general, these are guaranteed to converge. However, they are not often guaranteed to do so monotonically.

The geometric interpretation of gradient descent is searching for the bottom of a bowl by walking down the sides of the bowl. Think about an image like this. The size of those steps depends on the algorithm used. It's usually computationally cheap to have a constant or monotonically decreasing step size. If the step size is too large, which is likely in the beginning of your algorithm, then you can "overshoot" your goal, and hence end up with an iteration larger than the previous one.

Answered by Tim Atreides on November 29, 2021

Add your own answers!

Ask a Question

Get help from others!

© 2024 TransWikia.com. All rights reserved. Sites we Love: PCI Database, UKBizDB, Menu Kuliner, Sharing RPP