Datasets:
Tasks:
Feature Extraction
Formats:
csv
Sub-tasks:
language-modeling
Languages:
English
Size:
1K - 10K
License:
| pretty_name: Dataset Featurization | |
| language: | |
| - en | |
| license: | |
| - mit | |
| task_categories: | |
| - feature-extraction | |
| task_ids: | |
| - language-modeling | |
| configs: | |
| - config_name: nyt | |
| data_files: | |
| - split: train | |
| path: data/nyt/samples.csv | |
| - config_name: nyt-evaluation-0 | |
| data_files: | |
| - split: train | |
| path: data/nyt/evaluation/evaluation_df_group_0.csv | |
| - config_name: nyt-evaluation-1 | |
| data_files: | |
| - split: train | |
| path: data/nyt/evaluation/evaluation_df_group_1.csv | |
| - config_name: nyt-evaluation-2 | |
| data_files: | |
| - split: train | |
| path: data/nyt/evaluation/evaluation_df_group_2.csv | |
| - config_name: amazon | |
| data_files: | |
| - split: train | |
| path: data/amazon/samples.csv | |
| - config_name: amazon-evaluation-0 | |
| data_files: | |
| - split: train | |
| path: data/amazon/evaluation/evaluation_df_group_0.csv | |
| - config_name: amazon-evaluation-1 | |
| data_files: | |
| - split: train | |
| path: data/amazon/evaluation/evaluation_df_group_1.csv | |
| - config_name: amazon-evaluation-2 | |
| data_files: | |
| - split: train | |
| path: data/amazon/evaluation/evaluation_df_group_2.csv | |
| - config_name: dbpedia | |
| data_files: | |
| - split: train | |
| path: data/dbpedia/samples.csv | |
| - config_name: dbpedia-evaluation-0 | |
| data_files: | |
| - split: train | |
| path: data/dbpedia/evaluation/evaluation_df_group_0.csv | |
| - config_name: dbpedia-evaluation-1 | |
| data_files: | |
| - split: train | |
| path: data/dbpedia/evaluation/evaluation_df_group_1.csv | |
| - config_name: dbpedia-evaluation-2 | |
| data_files: | |
| - split: train | |
| path: data/dbpedia/evaluation/evaluation_df_group_2.csv | |
| # Dataset Featurization: Experiments | |
| This repository contains datasets used in evaluating **Dataset Featurization** against the prompting baseline. For datasets used in the case studies, please refer to [Compositional Preference Modeling](https://huggingface.co/datasets/Bravansky/compositional-preference-modeling) and [Compact Jailbreaks](https://huggingface.co/datasets/Bravansky/compact-jailbreaks). | |
| The evaluation focuses on three datasets: The [New York Times Annotated Corpus (NYT)](https://catalog.ldc.upenn.edu/docs/LDC2008T19/new_york_times_annotated_corpus.pdf), [Amazon Reviews (Amazon)](https://amazon-reviews-2023.github.io/), and [DBPEDIA](https://huggingface.co/datasets/DeveloperOats/DBPedia_Classes). For each dataset, we sample 15 different categories and construct three separate subsets, each containing 5 categories with 100 samples per category. We evaluate the featurization method's performance on each subset. | |
| ### NYT | |
| From the NYT corpus, we utilize manually reviewed tags from the NYT taxonomy classifier, specifically focusing on articles under "Features" and "News" categories, to construct a dataset of texts with their assigned categories. Below is how to access the input datasets and the proposed features with their assignments from the evaluation stage: | |
| ```python | |
| import datasets | |
| text_df = load_dataset("Bravansky/dataset-featurization", "nyt", split="train").to_pandas() | |
| evaluation_df_0 = load_dataset("Bravansky/dataset-featurization", "nyt-evaluation-0", split="train").to_pandas() | |
| evaluation_df_1 = load_dataset("Bravansky/dataset-featurization", "nyt-evaluation-1", split="train").to_pandas() | |
| evaluation_df_2 = load_dataset("Bravansky/dataset-featurization", "nyt-evaluation-2", split="train").to_pandas() | |
| ``` | |
| ### Amazon | |
| Using a dataset of half a million customer reviews, we focus on identifying high-level item categories (e.g., Books, Fashion, Beauty), excluding reviews labeled "Unknown". The input datasets and the proposed features with their assignments from the evaluation stage can be accessed as follows: | |
| ```python | |
| import datasets | |
| text_df = load_dataset("Bravansky/dataset-featurization", "amazon", split="train").to_pandas() | |
| evaluation_df_0 = load_dataset("Bravansky/dataset-featurization", "amazon-evaluation-0", split="train").to_pandas() | |
| evaluation_df_1 = load_dataset("Bravansky/dataset-featurization", "amazon-evaluation-1", split="train").to_pandas() | |
| evaluation_df_2 = load_dataset("Bravansky/dataset-featurization", "amazon-evaluation-2", split="train").to_pandas() | |
| ``` | |
| ### DBPEDIA | |
| Using the pre-processed DBPEDIA dataset, we focus on reconstructing categories labeled as level `l2`: | |
| ```python | |
| import datasets | |
| text_df = load_dataset("Bravansky/dataset-featurization", "dbpedia", split="train").to_pandas() | |
| evaluation_df_0 = load_dataset("Bravansky/dataset-featurization", "dbpedia-evaluation-0", split="train").to_pandas() | |
| evaluation_df_1 = load_dataset("Bravansky/dataset-featurization", "dbpedia-evaluation-1", split="train").to_pandas() | |
| evaluation_df_2 = load_dataset("Bravansky/dataset-featurization", "dbpedia-evaluation-2", split="train").to_pandas() | |
| ``` |