Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Not Comparable Results #42

Open
Ferenas opened this issue Sep 30, 2022 · 5 comments
Open

Not Comparable Results #42

Ferenas opened this issue Sep 30, 2022 · 5 comments

Comments

@Ferenas
Copy link

Ferenas commented Sep 30, 2022

Hi, I have trained GroupVIT under your default settings with RedCAP+GCC3M+GCC12M, and I ran the experiments 2 times, only achieving 26.8% mIoU and 27.9% mIoU. According to the previous discussions in the Closed Issue, I found that it may be caused by the training data size. So would you plz tell me the specific data size of RedCap, GCC3M, GCC12M, and YFC100M?

@xvjiarui
Copy link
Contributor

Hi @Ferenas

May I have your global batch size? It is crucial to have large enough batch size to make contrastive learning work. In GroupViT, we use 4096.

The detailed size of each dataset could be found here

https://github.com/xvjiarui/GroupViT/blob/dbaf9e85a6903c948df05f292d0b2091e9852f94/configs/default.yml#L11-L30

@Ferenas
Copy link
Author

Ferenas commented Sep 30, 2022

Thanks for your feedback! I set the batch size as the default and use 8 GPUs to train the network, so the global batch size is 2048, which I think may not be the problem. But I found that the number of shards is different, even smaller, from yours. For instance, in GCC3M you have 436 shards but I only have 331 shards. Does it work fine? I actually want to know what the specific data shards storage in your device. For instance, GCC12M is reported to have 331GB but you say "GCC12M is 1.8T" [https://github.com//issues/16#issuecomment-1122995696].

@xvjiarui
Copy link
Contributor

Hi @Ferenas

It's due to saved with different image sized. But we are resizing to 224 during training, so it won't affect the result.

But it's a common issue that GCC12M is not complete since many urls become invalid. You may need to check how many images you downloaded.

@Ferenas
Copy link
Author

Ferenas commented Oct 8, 2022

Thanks for your reply, and I have downloaded the datasets again, found this is mainly due to the size of the datasets. BTW, I have one question about the convert_coco_object.py, could you please explain a little about the clsID_to_trID? Why do the label ids here need to be remapped?

@xvjiarui
Copy link
Contributor

xvjiarui commented Oct 8, 2022

Hi @Ferenas

It's to map the segmentation label into continuous integers.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants