Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about HQ-Output Token and weight updates in the frozen Mask Decoder #145

Open
Linn0910 opened this issue Sep 30, 2024 · 0 comments

Comments

@Linn0910
Copy link

Hello,

Thank you for your great work on the HQ-SAM model! I have a question regarding the role of the HQ-Output Token in the model when interacting with the frozen Mask Decoder.
From the architecture diagram, I understand that the HQ-Output Token is integrated into the frozen Mask Decoder to improve segmentation accuracy. However, I am curious about how the HQ-Output Token's weights are updated during training, given that the Mask Decoder itself is frozen and its weights are not updated.
Here are my specific questions:

1.Since the Mask Decoder is frozen, how are the HQ-Output Token's weights updated during training?
2.Does the HQ-Output Token rely solely on the Global-local Fusion and MLP layers for weight updates, or does it interact with the Mask Decoder in a different way for updates?
3.How does the error correction mechanism contribute to the HQ-Output Token’s learning in this setup?
I would greatly appreciate it if you could clarify these points. Thank you again for your time and for sharing your amazing research!

Best regards,
Lin

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant