Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Some personal questions #15

Open
719304040 opened this issue Nov 17, 2024 · 0 comments
Open

Some personal questions #15

719304040 opened this issue Nov 17, 2024 · 0 comments

Comments

@719304040
Copy link

I am eager to integrate this work into my own project. However, I have a question regarding certain parts of the code and would greatly appreciate the author's assistance.

When calculating the logits of the Q-values (in def get_q_values), why is torch.roll used to shift embedding_values? I am struggling to understand why this step is necessary, as it seems to only occur during backpropagation.

action_bin_embeddings = self.action_bin_embeddings[:num_actions] action_bin_embeddings = torch.roll(action_bin_embeddings, shifts = -1, dims = 1) logits = einsum('b n d, n a d -> b n a', embed, action_bin_embeddings)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant