--- task_categories: - other configs: - config_name: AmazonReviews2014-Beauty data_files: - split: val path: AmazonReviews2014-Beauty/val.jsonl - split: test path: AmazonReviews2014-Beauty/test.jsonl - config_name: AmazonReviews2014-Sports_and_Outdoors data_files: - split: val path: AmazonReviews2014-Sports_and_Outdoors/val.jsonl - split: test path: AmazonReviews2014-Sports_and_Outdoors/test.jsonl - config_name: AmazonReviews2023-Industrial_and_Scientific data_files: - split: val path: AmazonReviews2023-Industrial_and_Scientific/val.jsonl - split: test path: AmazonReviews2023-Industrial_and_Scientific/test.jsonl - config_name: AmazonReviews2023-Musical_Instruments data_files: - split: val path: AmazonReviews2023-Musical_Instruments/val.jsonl - split: test path: AmazonReviews2023-Musical_Instruments/test.jsonl - config_name: AmazonReviews2023-Office_Products data_files: - split: val path: AmazonReviews2023-Office_Products/val.jsonl - split: test path: AmazonReviews2023-Office_Products/test.jsonl - config_name: Steam data_files: - split: val path: Steam/val.jsonl - split: test path: Steam/test.jsonl - config_name: Yelp-Yelp_2020 data_files: - split: val path: Yelp-Yelp_2020/val.jsonl - split: test path: Yelp-Yelp_2020/test.jsonl --- # MemGen Annotations This is the annotation dataset for the paper **[How Well Does Generative Recommendation Generalize?](https://huggingface.co/papers/2603.19809)**. The annotations categorize evaluation instances under the leave-one-out protocol: - **test** split uses the **last** item in the user history sequence as target, - **val** split uses the **second-to-last** item as target. ## Columns - `sample_id`: row index within the split in the original dataset. - `user_id`: raw user identifier (join key). - `master`: one of `memorization`, `generalization`, `uncategorized`. - `subcategories`: list of `{rule, hop}` for fine-grained generalization types. - `all_labels`: all string labels (e.g., `["generalization", "symmetry_3"]`). ## Load in M&G annotations ```python from datasets import load_dataset labels = load_dataset( "jamesding0302/memgen-annotations", "AmazonReviews2014-Beauty", split="test", ) print(labels[0]) ``` ## Merge with processed dataset ```python # 1) Load your processed dataset split (must be aligned with labels by row order) ds = pipeline.split_datasets["test"] # 2) Append label columns to the original dataset ds = (ds .add_column("master", labels["master"]) .add_column("subcategories", labels["subcategories"]) .add_column("all_labels", labels["all_labels"])) ```