Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Augmentation requests #5

Closed
LifeIsStrange opened this issue May 8, 2021 · 2 comments
Closed

Augmentation requests #5

LifeIsStrange opened this issue May 8, 2021 · 2 comments

Comments

@LifeIsStrange
Copy link

LifeIsStrange commented May 8, 2021

Those are apparently the most promising optimizers, would be very useful to see how it compare to RAdam/madgrad!

Adabelief
lessw2020/Ranger-Deep-Learning-Optimizer#44

Stochastic weight averaging
https://pytorch.org/blog/pytorch-1.6-now-includes-stochastic-weight-averaging/

Adas
https://paperswithcode.com/paper/adas-adaptive-scheduling-of-stochastic

@lessw2020
Copy link
Owner

Hi @LifeIsStrange - thanks for the feedback.
1 - Have already done some unpublished testing and Ranger21 outperforms Ranger. Thus recommend upgrading to Ranger21 if Ranger worked well already for a given dataset.
2 - Adabelief - have plans to test out the adabelief mechanism (adapting step size based on agreement with expected gradient) inside Ranger21 as an additional enhancement.
3 - SWA isn't an optimizer per se, it can be done with any optimizer. But, would be interesting to test that in addition to running with Ranger21 and see if it yields another way to improve training results.
4 - adas - looks interesting. Will work on testing it.
Thanks again for the ideas here. I'm going to move it over to discussion forum to track progress on it.

@lessw2020
Copy link
Owner

It's ported over to discussion now via this topic:
#7

Thanks again for the ideas! I'm closing this on the issue side and will track in discussions.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants