Replies: 1 comment 1 reply
-
It is definitely possible to do the training on one or more 24 GB GPUs. I would recommend trying a finetuning toolkit like SimpleTuner as a starting point, but if you'd like to dive deep into how low memory training could be done generally, I would recommend trying to read up on and use the following techniques/suggestions:
Typically, DeepSpeed + Gradient checkpointing is more than enough to finetune billion parameter models under 24 GB at low batch sizes (1-4). DeepSpeed should be quite easy to enable - you just have to follow the questions and create appropriate config using |
Beta Was this translation helpful? Give feedback.
-
Hi, anyone successfully ran this demo script:
https://github.com/huggingface/diffusers/blob/main/examples/controlnet/README_sdxl.md
on something like RTX 3090 or RTX 4090?
I have not been successful and always ran into OOM..
tried xformes but probably way more optimization things would be necessary..
I can use 2x 3090 or 4090 but only pcie connected so not even sure if that would be any help, probably would be very slow but if some sharding can be implemented to at least fit into VRAM?
Any help would be greatly appreciated :) Thanks
Beta Was this translation helpful? Give feedback.
All reactions