You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, I have been trying to train images using LLaMA-3.2. However, I encountered an issue with the current requirements.txt file, specifically with transformers==4.42.4. The error I faced is as follows:
ValueError: rope_scaling must be a dictionary with two fields, type and factor, got {'factor': 32.0, 'high_freq_factor': 4.0, 'low_freq_factor': 1.0, 'original_max_position_embeddings': 8192, 'rope_type': 'llama3'}
Upon inspecting the Transformers code (configuration_llama.py), I noticed that this case is not handled in version 4.42.4. Updating transformers to 4.43.1 resolves this specific problem.
Additionally:
According to Hugging Face's release notes, LLaMA 3.2 support begins with transformers==4.45.0.
The config.json file provided with the model weights indicates "transformers_version": "4.44.2".
However, updating to newer versions of transformers such as 4.45.0 introduces issues with methods like save_checkpoint, as the number of arguments has changed.
In light of this, could you please share the exact requirements.txt you used for image training with LLaMA-3.2 (specifically for the 3B and 1B models)? This would greatly help ensure compatibility and avoid any unnoticed errors during training.
Steps to Reproduce
Use transformers==4.42.4 and attempt to train LLaMA 32.
Encounter the rope_scalingValueError.
Thank you for your assistance! Looking forward to your response.
The text was updated successfully, but these errors were encountered:
tsw123tsw
changed the title
# Request for requirements.txt for LLaMA-3.2 Image Training
Request for requirements.txt for LLaMA-3.2 Image Training
Jan 8, 2025
Description
Hi, I have been trying to train images using LLaMA-3.2. However, I encountered an issue with the current
requirements.txt
file, specifically withtransformers==4.42.4
. The error I faced is as follows:Upon inspecting the Transformers code (
configuration_llama.py
), I noticed that this case is not handled in version4.42.4
. Updatingtransformers
to4.43.1
resolves this specific problem.Additionally:
transformers==4.45.0
.config.json
file provided with the model weights indicates"transformers_version": "4.44.2"
.transformers
such as 4.45.0 introduces issues with methods likesave_checkpoint
, as the number of arguments has changed.In light of this, could you please share the exact
requirements.txt
you used for image training with LLaMA-3.2 (specifically for the 3B and 1B models)? This would greatly help ensure compatibility and avoid any unnoticed errors during training.Steps to Reproduce
transformers==4.42.4
and attempt to train LLaMA 32.rope_scaling
ValueError
.Thank you for your assistance! Looking forward to your response.
The text was updated successfully, but these errors were encountered: