Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about ACFM #6

Open
Qiublack opened this issue Feb 15, 2023 · 2 comments
Open

Question about ACFM #6

Qiublack opened this issue Feb 15, 2023 · 2 comments

Comments

@Qiublack
Copy link

Hi, thanks for your excellent work. In this article, I have a little confusion about ACFM. Why do you use 1-M as a weighting factor to multiply Fb in the formula of the article, instead of multiplying it by M as Fa.
image

@thograce
Copy link
Owner

Hi, thanks for your excellent work. In this article, I have a little confusion about ACFM. Why do you use 1-M as a weighting factor to multiply Fb in the formula of the article, instead of multiplying it by M as Fa. image

This is a good question. My assumption is: 1. ACFM is a cross-layer fusion module, and the information is transmitted forward layer by layer, that is, the high-level information (Fb) should keep close to the low-level information (Fa). Therefore, after the initial fusion and calculation of MSCA, M should be used as the Fa attention weight. 2. If the same M is used to multiply Fa and Fb at the same time, the flexibility is lost, so 1-M is used to let the module allocate the weights of Fa and Fb during model learning. I'm sorry I didn't explain this point in the paper. Thank you for your attention.

@Qiublack
Copy link
Author

Hi, thanks for your excellent work. In this article, I have a little confusion about ACFM. Why do you use 1-M as a weighting factor to multiply Fb in the formula of the article, instead of multiplying it by M as Fa. image

This is a good question. My assumption is: 1. ACFM is a cross-layer fusion module, and the information is transmitted forward layer by layer, that is, the high-level information (Fb) should keep close to the low-level information (Fa). Therefore, after the initial fusion and calculation of MSCA, M should be used as the Fa attention weight. 2. If the same M is used to multiply Fa and Fb at the same time, the flexibility is lost, so 1-M is used to let the module allocate the weights of Fa and Fb during model learning. I'm sorry I didn't explain this point in the paper. Thank you for your attention.

I see, thanks a lot.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants