Skip to content

Commit

Permalink
clean up the_cauldron_transform args
Browse files Browse the repository at this point in the history
  • Loading branch information
andrewkho committed Nov 26, 2024
1 parent 7057aa9 commit cb67f43
Showing 1 changed file with 6 additions and 6 deletions.
12 changes: 6 additions & 6 deletions recipes/lora_finetune_distributed_multi_dataset.py
Original file line number Diff line number Diff line change
Expand Up @@ -948,12 +948,12 @@ def train(self) -> None:
# will include multiple forward / backward passes if gradient accumulation > 1
self._profiler.step()

if self._is_rank_zero:
total_twfb_pct = round(epoch_total_dl_time / epoch_step_time, 5)
log.info(
f"End of epoch {self.epochs_run}! "
f"{total_twfb_pct=}, {epoch_step_time=}, {epoch_dl_time=}, {epoch_total_dl_time=}",
)
if self._is_rank_zero:
total_twfb_pct = round(epoch_total_dl_time / epoch_step_time, 5)
log.info(
f"End of epoch {self.epochs_run}! "
f"{total_twfb_pct=}, {epoch_step_time=}, {epoch_dl_time=}, {epoch_total_dl_time=}",
)
self.epochs_run += 1
self.save_checkpoint(epoch=curr_epoch)

Expand Down

0 comments on commit cb67f43

Please sign in to comment.