-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ESM2 Infer partial batches using predict method #304
base: main
Are you sure you want to change the base?
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes! Doesn't this also depend on the latest nemo though? Should you bump nemo to top of tree?
Also see #302 which bumps the nemo version, is it new enough for your needs? |
/build-ci |
The NVIDIA/NeMo#10934 has overridden the changes to expose CC @jstjohn |
/build-ci |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM, but where does this use the new drop_last
argument?
|
/build-ci |
…IA/bionemo-framework into farhadr/infer_partial_batch
/build-ci |
/build-ci |
/build-ci |
1 similar comment
/build-ci |
/build-ci |
1 similar comment
/build-ci |
/build-ci |
This depends on changes in NVIDIA/NeMo#10837 to expose
drop_last
inMegatronDataSampler
and allow inference of partial batches.Note:
The NVIDIA/NeMo#10934 has overridden the changes to expose drop_last. We should now wrap dataloaders with nemo.lightning.data.WrappedDataLoader that can store the mode attribute when creating the dataloader in datamodules. Then drop_last=False if dataloader is in test or predict mode.