Templates for training #335
elismasilva
started this conversation in
General
Replies: 1 comment 3 replies
-
|
likely the most you can 'get away with' is:
even then, with all of the optimisations, SDXL training will consume more than 22G of VRAM without DeepSpeed. SDXL is a large model. if you aren't re-parameterising the thing (eg. going from epsilon to x-prediction) you are likely better off training a LoRA. try USE_DORA=true for LoRA training. if you're really needing to optimise full u-net training at the expense of correctness guarantees, OneTrainer claims to have support for full u-net training on a 3060 Ti. that's not something i'm interested in implementing here, as i honestly believe that's just not a good idea. |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi, do you have some template or sample configuration file for training SDXL ? i am starting training and testing hyperparameters, but i am trying to find one configuration that combine low vram consumption and velocity on training. I changed my bucket to 768x768 put grandient to 4 and batch size to 1 with 16 images. My scheduler is 'sine', what params impact more in memory consumption?
Thank you!
Beta Was this translation helpful? Give feedback.
All reactions