We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hello,
Could you please advise me on how to disable the KV cache? I would also appreciate any guidance on how to implement this change in code.
Thank you for your assistance.
The text was updated successfully, but these errors were encountered:
You can refer to this: https://github.com/JAYANDJEAN/From_Transformer_to_GPTs/blob/main/04_llama2/llama.py I use use_cache to control whether to use the cache, because we don't need to use the cache during training.
Sorry, something went wrong.
@JAYANDJEAN Thanks It means I can turn off caching by modifying codes?
No branches or pull requests
Hello,
Could you please advise me on how to disable the KV cache?
I would also appreciate any guidance on how to implement this change in code.
Thank you for your assistance.
The text was updated successfully, but these errors were encountered: