You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In the scripts script/MLLMSD_7b.sh and script/SmartEdit_7b.sh, you have specified --bf16 True, yet it seems that the corresponding deepspeed configuration in scripts/zero_mixed.json is missing the line "bf16": {"enabled": "auto"}. As a result, the --bf16 True flag does not appear to be taking effect. I would like to confirm whether this is a mistake or intentional.
Additionally, when training the MLLMSD 7b model, the logs indicate the following data types, which indicates that you have set some parts of the model to use float32 and other parts to use bfloat16.
It is acceptable to me that the LLM model uses torch.bfloat16. However, I am curious as to why the VAE model, which has a relatively small number of parameters, is also set to use torch.bfloat16. Could there be a specific reason for this choice?
The text was updated successfully, but these errors were encountered:
In the scripts
script/MLLMSD_7b.sh
andscript/SmartEdit_7b.sh
, you have specified--bf16 True
, yet it seems that the corresponding deepspeed configuration inscripts/zero_mixed.json
is missing the line"bf16": {"enabled": "auto"}.
As a result, the--bf16 True
flag does not appear to be taking effect. I would like to confirm whether this is a mistake or intentional.Additionally, when training the MLLMSD 7b model, the logs indicate the following data types, which indicates that you have set some parts of the model to use
float32
and other parts to usebfloat16
.It is acceptable to me that the LLM model uses
torch.bfloat16
. However, I am curious as to why the VAE model, which has a relatively small number of parameters, is also set to usetorch.bfloat16
. Could there be a specific reason for this choice?The text was updated successfully, but these errors were encountered: