You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is this strictly necessary? Since we are shifting right, the shifted tokens should be valid right?
Or is this accounting for items masked on the left? In which case you might be shifting and adding with an invalid token?
I noticed that RecurrentMemoryTransformer didn't do this:
In token shifting, you explicitly zero out masked items:
x-transformers/x_transformers/x_transformers.py
Lines 554 to 555 in 5ce82c9
Is this strictly necessary? Since we are shifting right, the shifted tokens should be valid right?
Or is this accounting for items masked on the left? In which case you might be shifting and adding with an invalid token?
I noticed that
RecurrentMemoryTransformer
didn't do this:https://github.com/lucidrains/recurrent-memory-transformer-pytorch/blob/d45ef72a40324c6224ffacb890d5593a69db73de/recurrent_memory_transformer_pytorch/recurrent_memory_transformer.py#L65-L70
Hence why I'm asking if it's strictly necessary.
The text was updated successfully, but these errors were encountered: