-
Notifications
You must be signed in to change notification settings - Fork 825
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bug Fix: 443 Bytes adapter_model.bin
files
#44
Conversation
I haven't yet tested the output of the adapters trained after this. There seems to be a debate on this issue in the linked @artidoro do let us know if this is the right approach or folks should wait for fix on the peft end |
Thank you for your contribution! Your code has been merged into the main branch. |
Your code style is very consistent and easy to read, thanks for that! |
Did a bit of verification on
I also checked out the |
Thank you @KKcorps! I also just replicated your fix and it seems to properly store the adapter checkpoints. |
Bug Fix: 443 Bytes `adapter_model.bin` files
Aims to fix #38 and #41
Currently, we get extremely small adapter files on checkpoint.
This seems to be due to some issue in the PEFT library.
One of the working solution is to return to an older version which is not possible (since it doesn't contain QLoRA changes)
The following solution works on my setup with 4080 card as well as on Colab notebook.
It has been borrowed from alpaca-lora