Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Implementation details of view transformer #12

Open
Zhentao-Liu opened this issue Feb 4, 2023 · 3 comments
Open

Implementation details of view transformer #12

Zhentao-Liu opened this issue Feb 4, 2023 · 3 comments

Comments

@Zhentao-Liu
Copy link

In the provided code, attn = k - q[:,:,None,:] + pos, attn = self.attn_fc(attn). However, in Fig. 2.a and alg.1, there should not be self.attn_fc component. Could you give an explanation?

@Zhentao-Liu
Copy link
Author

This part code is in transformer_network.py class Attention2D

@Zhentao-Liu
Copy link
Author

In Eq 9, what do you mean by applying diag(.)

@MukundVarmaT
Copy link
Collaborator

Hi @Zhentao-Liu!

Thank you for pointing it out! Yes, there is an error in our pseudo-code in algorithm. 1 (although fa(.) was defined we never used it). However, our implementation details (in text) do discuss the same (Appendix. B - Memory-efficient Cross-View Attention).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants