Replies: 1 comment 1 reply
-
Hey @adivoj, can you tell us a bit about the environment you're trying to run in? What model you're trying to train, how many nodes, how many GPUs per node, and what kind of GPUs? I ask, because the right distributed training approach will depend on the above factors. For example, if you want to train on a single node with multiple GPUs, and the model is too large to fit on a single GPU, that is supported in Ludwig today. But if you want to do multi-node, or data-parallel, then it can be done with DeepSpeed integration in Ludwig, but only up to 8 bit quantization today. Happy to give you more details when I better understand the use case and can offer more precise advice. |
Beta Was this translation helpful? Give feedback.
-
When will distributed training on quantized models (llm with lora) be available? Do you know any other simple framework i can try?
Beta Was this translation helpful? Give feedback.
All reactions