Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Shifted Window and Shared Attention Patterns in Consecutive GMSAs #1

Open
JLiu-Edinburgh opened this issue Apr 7, 2022 · 1 comment

Comments

@JLiu-Edinburgh
Copy link

Hello,

The paper is very interesting to me, since SwinIR suffers from high memory consumption and slow convergence. I recently have two questions about the proposed framework.

Firstly, two consecutive GMSAs can share the attention maps, while the shifted window is applied to partition neighboring pixels together, which should derive different attention patterns. How is it addressed or is interleaved sharing mechanism adopted?

Secondly, the results in Table 3 show the reduction of FLOPs and Latency by using the shifted mechanism. How could this method reduce the computational footprint? Is it solely due to the removal of the masking and relative positional encoding used in SwinIR?

Finally, could you present the convergence of ELAN, compared with SwinIR and other CNN-based models? It can provide a more comprehensive comparison and better show the advantages of ELAN.

Thanks a lot.

BTW, the neat model architecture is definitely appealing.

@blackcow
Copy link

blackcow commented Sep 9, 2022

CVPR

Excellent questions! looking forward to the author's response

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants