Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Reproduce results on SST-2 / SST-5 #2

Open
villmow opened this issue Apr 20, 2020 · 3 comments
Open

Reproduce results on SST-2 / SST-5 #2

villmow opened this issue Apr 20, 2020 · 3 comments

Comments

@villmow
Copy link

villmow commented Apr 20, 2020

Hi,

could you please provide an instruction on how to reproduce your results on SST.

It's impossible for me to navigate in your impressive codebase (its huge). E.g. 8 different attention implementations:

  1. default_multihead_attention.py
     2. dptree_individual_multihead_attention.py
     3. dptree_multihead_attention.py
     4. dptree_onseq_multihead_attention.py
     5. dptree_sep_multihead_attention.py
     6. nstack_merge_tree_attention.py
     7. nstack_tree_attention.py
     8. nstack_tree_attention_eff.py

Thanks!

@walt676
Copy link

walt676 commented Mar 25, 2021

did you solve this problem? this codebase is too big for me to read.
thank you.

@nxphi47
Copy link
Owner

nxphi47 commented Mar 26, 2021

Hi, very sorry we did not have time to clean up the codes. As in shown in the instruction, please follow the configuration dwnstack_merge2seq_node_iwslt_onvalue_base_upmean_mean_mlesubenc_allcross_hier to find its implementation in the files nstack_archs.py and nstack_transformer.py

@walt676
Copy link

walt676 commented Apr 1, 2021 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants