-
Notifications
You must be signed in to change notification settings - Fork 15
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
finetuning problem with evfsam2 #26
Comments
The main differences between sam1 and sam2 lie in:
Might these differences affect your training? |
@vvvvvjdy can you share your fine-tuning script? many thanks. |
Sorry for the late reply.
During my experiment,I found that the data augmentation caused the
overfitting probelm. Once I used the same augmentation as used in sam2
pretrain (only random hflip for image) ,the problem was solved. I assume
that stronger data augmentation for sam2 than pre-training may cause the
model to learn some unnatural shortcut features.
The fine-tuneing scripe is simple,just like
many sam1 fine-tuneing works. Only difference is the backbone. (I just
train my model on images)
Regards.
yi-ming-qian ***@***.***> 于 2024年11月30日周六 上午8:09写道:
… @vvvvvjdy <https://github.com/vvvvvjdy> can you share your fine-tuning
script? many thanks.
—
Reply to this email directly, view it on GitHub
<#26 (comment)>, or
unsubscribe
<https://github.com/notifications/unsubscribe-auth/BACXUVSRN6EEIJO7MEBG35L2DD6ZZAVCNFSM6AAAAABOWGI2AWVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDKMBYG42DKNJXGY>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
The augmentations influence the model performance in another way. In referring segmentation tasks, text prompts contain geometric words like "on the left". Once flipping or cropping or some other augmentations are applied, the prompts would fail. So only non-geometric augmentations are recommended. |
@CoderZhangYx Quite agree with this statement. But even without such prompts in my finetuning data, the stronger data augmentation than the pretraining may cause this problem (some works have demonstrate the small-size model slike resnet-18 are especially sensitive to aug),I m shocked that such a large foundation model sam2 has such characteristics. |
That's amazing. What aug did you use? Could it be the reason that the aug didn't applied to the source fed to multi_model_extractor? Curious about this bug, honestly. |
@CoderZhangYx I originally use large scale jittering(strong) for the input image(both beit and sam1 or 2) and gtmask, and found it works well on evfsam1 but not on evfsam2. Did you use the same aug for evfsam1 and 2, and which aug did you use ? ( it not mentioned in the paper)? |
In fact we use no aug when training our model. It is so strange that scale jittering affect performance of sam2. Inform me if you find out any other reasons, thanks! |
Nice work of extending sams ability to text guidied!
We have used your evfsam1 as our baseline to a new area and showd a signifcant performance. However, when we finetuned your evfsam2, it was easy to overfiiting(didnt see in evfsam1).
Did you met the same prob when your fintune sam2 ? or some hyperpramters different from sam1?
Hope to recieve your suggestion!
The text was updated successfully, but these errors were encountered: