Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

8-bit precision training #1788

Open
rasbt opened this issue Oct 10, 2024 · 2 comments
Open

8-bit precision training #1788

rasbt opened this issue Oct 10, 2024 · 2 comments
Labels
enhancement New feature or request

Comments

@rasbt
Copy link
Collaborator

rasbt commented Oct 10, 2024

Might be worthile to look into

from torchao.float8 import convert_to_float8_training
convert_to_float8_training(model)

some time

@rasbt rasbt added the enhancement New feature or request label Oct 10, 2024
@mtasic85
Copy link
Contributor

I watched this week this video: https://www.youtube.com/watch?v=Br07GsnnvWc

This performance is expected:

image

About Embedding layer, it should be investigated if they use appropriate class like this one, or we need to take care of it: https://github.com/pytorch/ao/blob/e7b33bc91c831d10249c1222c8b4b667f18f28b7/torchao/quantization/prototype/qat/embedding.py#L27

@rasbt
Copy link
Collaborator Author

rasbt commented Oct 22, 2024

Thanks for sharing this. Need to check out the video some time. It's long (but good explanations of tricky things are naturally long) but looks really informative.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants