Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Resize token embedding. #1145

Open
hungphongtrn opened this issue Jul 7, 2024 · 1 comment
Open

Resize token embedding. #1145

hungphongtrn opened this issue Jul 7, 2024 · 1 comment
Assignees
Labels
discussion Start a discussion

Comments

@hungphongtrn
Copy link

Hi,

Are there any methods to resize a pretrained models after I add in new vocabulary?

Thanks in advance!

@ebsmothers
Copy link
Contributor

Hi @hungphongtrn thanks for creating the issue. We don't currently have any such methods in the library, but I'm curious to know more about what you have in mind here. Is the idea to append additional indices to the embedding for an existing vocabulary? If so, how would the new indices be initialized -- randomly, or from some other set of weights that can be passed in? Similar question for the final projection back to the vocab dimension since this would need to be updated as well.

@joecummings joecummings added discussion Start a discussion and removed help wanted labels Dec 13, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
discussion Start a discussion
Projects
None yet
Development

No branches or pull requests

4 participants