You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello. So I've been building a TextClassifier. I am up to the optimize learning rate step. However, I don't think the find_learning_rate function is working correctly, and I don't know quite enough about it to debug it very well.
1) The model is supposed to take awhile to train. It completes really fast when using this function.
It's supposed to try 100 different learning rates. I get 87 iterations, not 100, and they are all using the same starting learning rate.
I can't see what it's doing at all. There is no output of the train step. It just all of a sudden says its done, but it hasn't produced anything of value.
I don't know if this is a small bug, a user error, or something bigger.
It almost feels like it's running 87 minibatches at the starting learning rate, and then thinking it's done already when it's barely started.
I mean worst case scenario, I can use hyperopt to check a few learning rates over night. But it would be really cool if this method could be made to work. I need a little guidance to help debug this.
Thanks so much.
The text was updated successfully, but these errors were encountered:
I've been investigating and reading up on the theory. It's supposed to be starting out untrained and go by minibatch. That part is good. Point one is invalid.
The issue is that scheduler.step(1) is supposed to increment the learning rate each time. It is not. I will have to go deeper into scheduler.
For point 2, I feel this is still valid. If it runs out of mini-batches before iterations is reached, shouldn't it start over from the beginning of the batches? If I ask for 1000 iterations, I want 1000 iterations. It's up to me to decide whether that number makes any sense.
Hello. So I've been building a TextClassifier. I am up to the optimize learning rate step. However, I don't think the find_learning_rate function is working correctly, and I don't know quite enough about it to debug it very well.
1) The model is supposed to take awhile to train. It completes really fast when using this function.It's supposed to try 100 different learning rates. I get 87 iterations, not 100, and they are all using the same starting learning rate.
I can't see what it's doing at all. There is no output of the train step. It just all of a sudden says its done, but it hasn't produced anything of value.
I don't know if this is a small bug, a user error, or something bigger.
It almost feels like it's running 87 minibatches at the starting learning rate, and then thinking it's done already when it's barely started.
I mean worst case scenario, I can use hyperopt to check a few learning rates over night. But it would be really cool if this method could be made to work. I need a little guidance to help debug this.
Thanks so much.
The text was updated successfully, but these errors were encountered: