TransWikia.com

How can a learning rate that is too large cause the output of the network (and the error) to go to infinity?

Artificial Intelligence Asked by user1477107 on August 24, 2021

It happened to my neural network, when I use a learning rate of <0.2 everything works fine, but when I try something above 0.4 I start getting "nan" errors because the output of my network keeps increasing.

From what I understand, what happens is that if I choose a learning rate that is too large, I overshoot the local minimum. But still, I am getting somewhere, and from there I’m moving in the correct direction. At worst my output should be random, I don’t understand what is the scenario that causes my output and error to approach infinity every time I run my NN with a learning rate that is too large (and it’s not even that large)

enter image description here
How does the red line go to infinity ever? I kind of understand it could happen if we choose a crazy high learning rate, but if the NN works for 0.2 and doesn’t for 0.4, I don’t understand that

Add your own answers!

Ask a Question

Get help from others!

© 2024 TransWikia.com. All rights reserved. Sites we Love: PCI Database, UKBizDB, Menu Kuliner, Sharing RPP