You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Oct 31, 2023. It is now read-only.
I am wondering if there are more explanations about the vectors concatenation. In the paper, u, v, uv, and |u-v| are concat and then it is used for the Softmax loss. |u-v| seems to be intuitive since when u and v are close, this goes to near zero (semantically similar). But the meaning of u, v and uv are not clear, would you elaborate more? Also, I see some work took square of max(u, v), and it seems to be working well. Overall, I am curious why those are working and how they are designed.
The text was updated successfully, but these errors were encountered:
Sign up for freeto subscribe to this conversation on GitHub.
Already have an account?
Sign in.
I am wondering if there are more explanations about the vectors concatenation. In the paper, u, v, uv, and |u-v| are concat and then it is used for the Softmax loss. |u-v| seems to be intuitive since when u and v are close, this goes to near zero (semantically similar). But the meaning of u, v and uv are not clear, would you elaborate more? Also, I see some work took square of max(u, v), and it seems to be working well. Overall, I am curious why those are working and how they are designed.
The text was updated successfully, but these errors were encountered: