You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Oct 31, 2023. It is now read-only.
Hi,
In the github documentation under DocSpace, it says the following:
Model: Each document is represented by a bag-of-words of the document. Each user is represented as a (bag of) the documents that they liked/clicked in the past. At training time, at each step one random document is selected as the label and the rest of the bag of documents are selected as input.
I would like to know if the same parameter p is used for both aggregating words up to doc level and aggregating docs up to user level.
That is to say, are the following representations correct?:
doc_1 = (word_vec_1 + word_vec_2 + ... + word_vec_n)/n^p
and
user_1 = (doc_1 + doc_2 + ... + doc_m)/m^p
In the above example, the same value for p is used to aggregate words to doc-level and docs to user-level. Is this what is going on under the hood in DocSpace?
Thank you
The text was updated successfully, but these errors were encountered:
Sign up for freeto subscribe to this conversation on GitHub.
Already have an account?
Sign in.
Hi,
In the github documentation under DocSpace, it says the following:
I would like to know if the same parameter p is used for both aggregating words up to doc level and aggregating docs up to user level.
That is to say, are the following representations correct?:
doc_1 = (word_vec_1 + word_vec_2 + ... + word_vec_n)/n^p
and
user_1 = (doc_1 + doc_2 + ... + doc_m)/m^p
In the above example, the same value for p is used to aggregate words to doc-level and docs to user-level. Is this what is going on under the hood in DocSpace?
Thank you
The text was updated successfully, but these errors were encountered: