-
Notifications
You must be signed in to change notification settings - Fork 48
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Setting default values in training job config #104
Setting default values in training job config #104
Conversation
Signed-off-by: Thara Palanivel <[email protected]>
Signed-off-by: Thara Palanivel <[email protected]>
Signed-off-by: Thara Palanivel <[email protected]>
Signed-off-by: Thara Palanivel <[email protected]>
Signed-off-by: Thara Palanivel <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
ok with code changes, suggestions around refactoring files and functions.
we can keep all build related code inside build folder and tests can be inside tests/build as well
Signed-off-by: Thara Palanivel <[email protected]>
Signed-off-by: Thara Palanivel <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
some questions - also @anhuong have you reviewed the refactor to launch_training as well?
Signed-off-by: Thara Palanivel <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
approved from my end, just a remaining question on default for peft_method^. - I want to ensure we are testing fine tuning correctly, peft_method should be None for fine tuning.
- Does peft_method need to be passed explicitly as None then for fine tuning?
- or should we keep default to None, and pass "pt"/"Lora" explictly for prompt tuning/LoRA?
I think second one makes more sense.
When @anhuong approves, we can merge
Signed-off-by: Thara Palanivel <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM!
with open(json_path, "r", encoding="utf-8") as f: | ||
contents = json.load(f) | ||
peft_method_parsed = contents.get("peft_method") | ||
logging.debug("Input params parsed: %s", contents) | ||
job_config_dict = json.load(f) | ||
elif json_env_var: | ||
job_config_dict = txt_to_obj(json_env_var) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Note to self: perhaps we should refactor this into build/utils as well since accelerate_launch and launch_training use the same method to parse the JSON
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Good point! I'll add that in the follow-up tests PR that is in the queue next
Signed-off-by: Thara Palanivel <[email protected]>
Signed-off-by: Thara Palanivel <[email protected]>
…#104) * Allow for default params to be set Signed-off-by: Thara Palanivel <[email protected]> * Add tests Signed-off-by: Thara Palanivel <[email protected]> * Simplifying default params logic Signed-off-by: Thara Palanivel <[email protected]> * Setting use_flash_attn default Signed-off-by: Thara Palanivel <[email protected]> * Formatting Signed-off-by: Thara Palanivel <[email protected]> * Address review comments Signed-off-by: Thara Palanivel <[email protected]> * Moving tests Signed-off-by: Thara Palanivel <[email protected]> * Address review comments Signed-off-by: Thara Palanivel <[email protected]> * Address review comment Signed-off-by: Thara Palanivel <[email protected]> * Fix merge conflicts Signed-off-by: Thara Palanivel <[email protected]> --------- Signed-off-by: Thara Palanivel <[email protected]> Signed-off-by: Anh-Uong <[email protected]>
…#104) * Allow for default params to be set Signed-off-by: Thara Palanivel <[email protected]> * Add tests Signed-off-by: Thara Palanivel <[email protected]> * Simplifying default params logic Signed-off-by: Thara Palanivel <[email protected]> * Setting use_flash_attn default Signed-off-by: Thara Palanivel <[email protected]> * Formatting Signed-off-by: Thara Palanivel <[email protected]> * Address review comments Signed-off-by: Thara Palanivel <[email protected]> * Moving tests Signed-off-by: Thara Palanivel <[email protected]> * Address review comments Signed-off-by: Thara Palanivel <[email protected]> * Address review comment Signed-off-by: Thara Palanivel <[email protected]> * Fix merge conflicts Signed-off-by: Thara Palanivel <[email protected]> --------- Signed-off-by: Thara Palanivel <[email protected]>
Description of the change
Setting the following defaults for training jobs, will get overwritten if user explicitly passes in a value
save_strategy
=epoch
logging_strategy
=epoch
Related issue number
How to verify the PR
Was the PR tested