Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FEATURE] Warm restarts - potential means of working around ESR plateauing (?) #490

Closed
2-dor opened this issue Oct 11, 2024 · 1 comment
Closed
Labels
enhancement New feature or request priority:low Low-priority issues

Comments

@2-dor
Copy link

2-dor commented Oct 11, 2024

Some higher gain reamps with a high amount if high-end have proved trickier to nail.

While I can always up the channels, dilations or kernel in one (or both) of the layers, I have been trying to find out if there are ways to avoid slow (or no) progress in the ESR after about the 700 epoch threshold using the STANDARD architecture in the trainer.

Having spent a bit of time poking ChatGPT's brains and experimenting with various ways (RAdam, AdamW etc.), making use of warm restarts seems to help a little bit.
Over the course of a 700 epochs, it seems to help take the ESR down from 0.009798 (default scheduling) to 0.008789 (with CosineAnnealingWarmRestarts).

That said, there is a caveat and fine-tuning that needs to go with it - I'll post back if I find a good combination. Right now, first attempts look promising but after the restarts the ESR still takes a while until it manages to go under its previous plateau:

image

@2-dor 2-dor added enhancement New feature or request priority:low Low-priority issues unread This issue is new and hasn't been seen by the maintainers yet labels Oct 11, 2024
@2-dor
Copy link
Author

2-dor commented Oct 13, 2024

Doesn't seem to perform any better than the Exponential decay from all the tests I've run so far.

@2-dor 2-dor closed this as completed Oct 13, 2024
@sdatkinson sdatkinson removed the unread This issue is new and hasn't been seen by the maintainers yet label Oct 14, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request priority:low Low-priority issues
Projects
None yet
Development

No branches or pull requests

2 participants