Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Rename configs for consistency #640

Closed
wants to merge 12 commits into from
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
30 changes: 11 additions & 19 deletions recipes/configs/llama2/13B_full.yaml
Original file line number Diff line number Diff line change
@@ -1,27 +1,19 @@
# Config for multi-device full finetuning in full_finetune_distributed.py
# using a Llama2 13B model
# Config for multi-device with full_finetune_distributed.py using a Llama2 13B model
#
# This config assumes that you've run the following command before launching
# this run:
# tune download --repo-id meta-llama/Llama-2-13b-hf \
# --hf-token <HF_TOKEN> \
# --output-dir /tmp/llama2-13b-hf
# This config assumes that you've run the following command before launching:
# $ tune download meta-llama/Llama-2-13b \
# --hf-token <HF_TOKEN> \
# --output-dir /tmp/llama2
#
# To launch on 4 devices, run the following command from root:
# tune run --nproc_per_node 4 full_finetune_distributed \
# --config llama2/13B_full \
# $ tune run --nproc_per_node 4 full_finetune_distributed \
# --config llama2/13B_full
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Do we want to split out with an example single device command here too? (I guess for 13B this is maybe less likely anyways)

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nvm I think these are not registered for the single-device recipes. Then maybe update the comment in 13B_lora.yaml?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Update what comment? The topline comment says it's just for multi-device

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sorry it's in the other yaml, commented at the specific lines

#
# You can add specific overrides through the command line. For example
# to override the checkpointer directory while launching training
# you can run:
# tune --nnodes 1 --nproc_per_node 4 full_finetune_distributed \
# --config llama2/13B_full \
# checkpointer.checkpoint_dir=<YOUR_CHECKPOINT_DIR>
#
# This config should be used with 2+ GPUs. Single device full fine-tuning
# requires several memory optimizations which are exposed through
# 7B_full_single_device.yaml. Please update the model and checkpoints to 13B
# in that config.
# to override the checkpointer directory while launching training:
# $ tune run --nproc_per_node 4 full_finetune_distributed \
# --config llama2/13B_full \
# checkpointer.checkpoint_dir=<YOUR_CHECKPOINT_DIR>


# Tokenizer
Expand Down
31 changes: 12 additions & 19 deletions recipes/configs/llama2/13B_lora.yaml
Original file line number Diff line number Diff line change
@@ -1,28 +1,21 @@
# Config for multi-device LoRA in lora_finetune_distributed.py
# using a Llama2 13B model
# Config for multi-device with lora_finetune_distributed.py using a Llama2 13B model
#
# This config assumes that you've run the following command before launching
# this run:
# tune download --repo-id meta-llama/Llama-2-13b-hf \
# --hf-token <HF_TOKEN> \
# --output-dir /tmp/llama2-13b-hf
# This config assumes that you've run the following command before launching:
# $ tune download meta-llama/Llama-2-13b \
# --hf-token <HF_TOKEN> \
# --output-dir /tmp/llama2
#
# To launch on 4 devices, run the following command from root:
# tune run --nproc_per_node 4 lora_finetune_distributed \
# --config llama2/13B_lora \
# $ tune run --nproc_per_node 4 lora_finetune_distributed \
# --config llama2/13B_lora
#
# You can add specific overrides through the command line. For example
# to override the checkpointer directory while launching training
# you can run:
# tune --nnodes 1 --nproc_per_node 4 lora_finetune_distributed \
# --config llama2/13B_lora \
# checkpointer.checkpoint_dir=<YOUR_CHECKPOINT_DIR>
# to override the checkpointer directory while launching training:
# $ tune run --nproc_per_node 4 lora_finetune_distributed \
# --config llama2/13B_lora \
# checkpointer.checkpoint_dir=<YOUR_CHECKPOINT_DIR>
#
# This config works best when the model is being fine-tuned on 2+ GPUs.
# For single device lora finetuning please use 7B_lora_single_device.yaml
# or 7B_qlora_single_device.yaml and update the model and checkpoints to
# the 13B model.

# This config is only tested on a multi-device setup

# Model Arguments
model:
Expand Down
35 changes: 18 additions & 17 deletions recipes/configs/llama2/7B_full.yaml
Original file line number Diff line number Diff line change
@@ -1,26 +1,27 @@
# Config for multi-device full finetuning in full_finetune_distributed.py
# using a Llama2 7B model
# Config settings for multi-device using full_finetune_distributed.py or single-device
# with full_finetune_single_device.py, specifically for a Llama2 7B model
#
# This config assumes that you've run the following command before launching
# this run:
# tune download --repo-id meta-llama/Llama-2-7b \
# --hf-token <HF_TOKEN> \
# --output-dir /tmp/llama2
# This config assumes that you've run the following command before launching:
# $ tune download meta-llama/Llama-2-7b \
# --hf-token <HF_TOKEN> \
# --output-dir /tmp/llama2
#
# To launch on 4 devices, run the following command from root:
# tune run --nproc_per_node 4 full_finetune_distributed \
# --config llama2/7B_full \
# $ tune run --nproc_per_node 4 full_finetune_distributed \
# --config llama2/7B_full
#
# To launch on a single device, run the following command:
# $ tune run full_finetune_single_device \
# --config llama2/7B_full
#
# You can add specific overrides through the command line. For example
# to override the checkpointer directory while launching training
# you can run:
# tune --nnodes 1 --nproc_per_node 4 full_finetune_distributed \
# --config llama2/7B_full \
# checkpointer.checkpoint_dir=<YOUR_CHECKPOINT_DIR>
# to override the checkpointer directory while launching training:
# $ tune run --nproc_per_node 4 full_finetune_distributed \
# --config llama2/7B_full \
# checkpointer.checkpoint_dir=<YOUR_CHECKPOINT_DIR>
#
# This config works best when the model is being fine-tuned on 2+ GPUs.
# Single device full finetuning requires more memory optimizations. It's
# best to use 7B_full_single_device.yaml for those cases
# For more memory optimizations, such as those needed when running on a single GPU,
# use llama2/7B_full_low_memory_example


# Tokenizer
Expand Down
Original file line number Diff line number Diff line change
@@ -1,25 +1,22 @@
# Config for single device full finetuning in full_finetune_single_device.py
# Config for single-device full finetuning with full_finetune_single_device.py
# using a Llama2 7B model
#
# This config assumes that you've run the following command before launching
# this run:
# tune download --repo-id meta-llama/Llama-2-7b \
# --hf-token <HF_TOKEN> \
# --output-dir /tmp/llama2
# This config assumes that you've run the following command before launching:
# $ tune download meta-llama/Llama-2-7b \
# --hf-token <HF_TOKEN> \
# --output-dir /tmp/llama2
#
# To launch on a single device, run the following command from root:
# tune run full_finetune_single_device \
# --config llama2/7B_full_single_device_low_memory \
# To launch, run the following command:
# $ tune run full_finetune_single_device \
# --config llama2/7B_full_low_memory
#
# You can add specific overrides through the command line. For example
# to override the checkpointer directory while launching training
# you can run:
# tune --nnodes 1 --nproc_per_node 1 full_finetune_single_device \
# --config llama2/7B_full_single_device_low_memory \
# checkpointer.checkpoint_dir=<YOUR_CHECKPOINT_DIR>
# to override the checkpointer directory while launching training:
# $ tune run full_finetune_single_device \
# --config llama2/7B_full_low_memory \
# checkpointer.checkpoint_dir=<YOUR_CHECKPOINT_DIR>
#
# This config works only for training on single device.

# This config specifies components from `bitsandbytes`, make sure you have it installed

# Tokenizer
tokenizer:
Expand Down
76 changes: 0 additions & 76 deletions recipes/configs/llama2/7B_full_single_device.yaml

This file was deleted.

35 changes: 18 additions & 17 deletions recipes/configs/llama2/7B_lora.yaml
Original file line number Diff line number Diff line change
@@ -1,26 +1,27 @@
# Config for multi-device LoRA finetuning in lora_finetune_distributed.py
# using a Llama2 7B model
# Config for multi-device with lora_finetune_distributed.py or single-device LoRA
# finetuning with lora_finetune_single_device.py using a Llama2 7B model
#
# This config assumes that you've run the following command before launching
# this run:
# tune download --repo-id meta-llama/Llama-2-7b \
# --hf-token <HF_TOKEN> \
# --output-dir /tmp/llama2
# This config assumes that you've run the following command before launching:
# $ tune download meta-llama/Llama-2-7b \
# --hf-token <HF_TOKEN> \
# --output-dir /tmp/llama2
#
# To launch on 4 devices, run the following command from root:
# tune run --nproc_per_node 4 lora_finetune_distributed \
# --config llama2/7B_lora \
# $ tune run --nproc_per_node 4 lora_finetune_distributed \
# --config llama2/7B_lora
#
# To launch on a single device, run the following command:
# $ tune run lora_finetune_single_device \
# --config llama2/7B_lora
#
# You can add specific overrides through the command line. For example
# to override the checkpointer directory while launching training
# you can run:
# tune --nnodes 1 --nproc_per_node 4 lora_finetune_distributed \
# --config llama2/7B_lora \
# checkpointer.checkpoint_dir=<YOUR_CHECKPOINT_DIR>
# to override the checkpointer directory while launching training:
# $ tune run --nproc_per_node 4 lora_finetune_distributed \
# --config llama2/7B_lora \
# checkpointer.checkpoint_dir=<YOUR_CHECKPOINT_DIR>
#
# This config works best when the model is being fine-tuned on 2+ GPUs.
# For single device lora finetuning please use 7B_lora_single_device.yaml
# or 7B_qlora_single_device.yaml
# For more memory optimizations, such as those needed when running on a single GPU,
# use llama2/7B_qlora


# Model Arguments
Expand Down
83 changes: 0 additions & 83 deletions recipes/configs/llama2/7B_lora_single_device.yaml

This file was deleted.

Loading
Loading