Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Checkpoint for ViT-L/14 Pretrained from Scratch (with SwiGLU) #473

Open
dfan opened this issue Oct 23, 2024 · 0 comments
Open

Checkpoint for ViT-L/14 Pretrained from Scratch (with SwiGLU) #473

dfan opened this issue Oct 23, 2024 · 0 comments

Comments

@dfan
Copy link
Member

dfan commented Oct 23, 2024

In Table 17 of the paper, it is mentioned that SwiGLU is used as the FFN layer for ViT-L/14 when training from scratch. In the model card, only the ViT-G model trained from scratch with SwiGLU is available (the ViT-L uses MLP FFN and seems to be the distilled version). Is it possible to add the checkpoint for the ViT-L/14 trained from scratch?

I.e. matching this config: https://github.com/facebookresearch/dinov2/blob/main/dinov2/configs/train/vitl14.yaml

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant