- Movie: We use the widely-adopted MovieLens-1M dataset that contains 1M user ratings for movies. Download
- Book: We use the "Books" subset of Amazon dataset that contains user ratings for books. Download
- Music: We use the "CDs & Vinyl" subset of Amazon to conduct experiments on the music domain. Download
- News: We use the MIND-small dataset as the benchmark for news domain. Download
Dataset | # User | # Item | # Interaction | # Sparsity |
---|---|---|---|---|
Movie | 6,040 | 3,706 | 1,000,209 | 95.53% |
Book | 8,026,324 | 2,330,066 | 22,507,155 | 99.99% |
Music | 1,578,597 | 486,360 | 3,749,004 | 99.99% |
News | 50,000 | 42,416 | 73,152 | 99.99% |
For each dataset, we remove items that do not have a provided title, as we require the title of items as input for the LLMs.
Next, we sort the interactions between users and items by their timestamps and remove users who have only a minimal amount of interactions.
Finally, we construct sets of candidate items and gather the sequences of items that the users had previously interacted with.
Specifically, for each test record, we gather the n_history
latest interacted items with positive feedback as user history.
To construct the candidate set for testing, we consider the item actually interacted with as the positive sample, and randomly sample N-1 items as negative samples for each selected interaction.
Note that we shuffle the candidate list because that LLMs may be sensitive to the indexes of candidates.
n_history
and n_candidate
are parameters, which can be customized according to your needs. In our experiments, we set both of them as 5.
Each dataset contains around 10,000 records for the final evaluation. If you want to remain more records, you can change our data-preprocessing codes in data_process
folder.
Dataset | Movie | Book | Music | News |
---|---|---|---|---|
# Records | 99,38 | 10,269 | 90,90 | 10,000 |
In our experiments, we remain the first 5 samples as the few-shot prompt examples and test on the rest samples.
You can directly download our pre-processed data from here.