Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about hierarchical Structure #65

Open
Titolasanta opened this issue Jan 11, 2024 · 1 comment
Open

Question about hierarchical Structure #65

Titolasanta opened this issue Jan 11, 2024 · 1 comment

Comments

@Titolasanta
Copy link

Hi, researching your paper I read that the used architecture utilizes a hierarchical Structure in a similar way as in the Hourglass transformer ( https://arxiv.org/abs/2110.13711 ). But for your case you are utilizing a encoder-decoder transformer, so how do you reshape what would usually be the input for your decoders which consists of both the output of the encoder which I would consider to be downsampled and the original input, which would be in the original input shape?

Thank you

@gaozhihan
Copy link
Contributor

Thank you for your interest and your question. It may help to refer to the following piece of our implementation:

if self.num_global_vectors > 0:
dec_out = self.decoder(initial_z, mem_l, mem_global_vector_l)
else:
dec_out = self.decoder(initial_z, mem_l)

The decoder typically takes as input the outputs from each layer of the encoder, as well as a dummy input named initial_z that matches the shape of the encoder's final layer output.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants