-
Notifications
You must be signed in to change notification settings - Fork 117
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Question] How do I calculate max_tokens max value? #34
Comments
batch_duration (s) = max_tokens / 16000 |
So my question is .. how many seconds can I have inside a batch? If I set max_tokens to a big number ( 1 200 000 ) I get an error: 2021-03-03 11:40:50 | WARNING | fairseq.trainer | OOM: Ran out of memory with exception: CUDA out of memory. Tried to allocate 210.00 MiB (G Does the number of virtual GPUs count? |
|
Given that I'm using for training 5 GPUs GeForce GTX 1080 Ti with 10.917GB memory each, how can I calculate the max_tokens so that no memory error occurs?
The text was updated successfully, but these errors were encountered: