| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| |
| """AG News topic classification dataset.""" |
|
|
|
|
| import csv |
|
|
| import datasets |
| from datasets.tasks import TextClassification |
|
|
|
|
| _DESCRIPTION = """\ |
| AG is a collection of more than 1 million news articles. News articles have been |
| gathered from more than 2000 news sources by ComeToMyHead in more than 1 year of |
| activity. ComeToMyHead is an academic news search engine which has been running |
| since July, 2004. The dataset is provided by the academic comunity for research |
| purposes in data mining (clustering, classification, etc), information retrieval |
| (ranking, search, etc), xml, data compression, data streaming, and any other |
| non-commercial activity. For more information, please refer to the link |
| http://www.di.unipi.it/~gulli/AG_corpus_of_news_articles.html . |
| The AG's news topic classification dataset is constructed by Xiang Zhang |
| (xiang.zhang@nyu.edu) from the dataset above. It is used as a text |
| classification benchmark in the following paper: Xiang Zhang, Junbo Zhao, Yann |
| LeCun. Character-level Convolutional Networks for Text Classification. Advances |
| in Neural Information Processing Systems 28 (NIPS 2015). |
| """ |
|
|
| _CITATION = """\ |
| @inproceedings{Zhang2015CharacterlevelCN, |
| title={Character-level Convolutional Networks for Text Classification}, |
| author={Xiang Zhang and Junbo Jake Zhao and Yann LeCun}, |
| booktitle={NIPS}, |
| year={2015} |
| } |
| """ |
|
|
| _TRAIN_DOWNLOAD_URL = "https://github.com/edubrigham/data/blob/541db4cdbb566aa5909e3eb4904d64f9683e5d4a/yahoo_answers_csv/train.csv" |
| _TEST_DOWNLOAD_URL = "https://github.com/edubrigham/data/blob/541db4cdbb566aa5909e3eb4904d64f9683e5d4a/yahoo_answers_csv/test.csv" |
|
|
|
|
| class AGNews(datasets.GeneratorBasedBuilder): |
| """AG News topic classification dataset.""" |
|
|
| def _info(self): |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=datasets.Features( |
| { |
| "text": datasets.Value("string"), |
| "labels": datasets.features.ClassLabel(names=["Society & Culture", "Science & Mathematics", "Health","Education & Reference", "Computers & Internet", "Sports", "Business & Finance", "Entertainment & Music", "Family & Relationships", "Politics & Government"]), |
| } |
| ), |
| homepage="http://groups.di.unipi.it/~gulli/AG_corpus_of_news_articles.html", |
| citation=_CITATION, |
| task_templates=[TextClassification(text_column="text", label_column="labels")], |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| train_path = dl_manager.download_and_extract(_TRAIN_DOWNLOAD_URL) |
| test_path = dl_manager.download_and_extract(_TEST_DOWNLOAD_URL) |
| return [ |
| datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": train_path}), |
| datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"filepath": test_path}), |
| ] |
|
|
| def _generate_examples(self, filepath): |
| """Generate AG News examples.""" |
| with open(filepath, encoding="utf-8") as csv_file: |
| csv_reader = csv.reader( |
| csv_file, quotechar='"', delimiter=",", quoting=csv.QUOTE_ALL, skipinitialspace=True |
| ) |
| for id_, row in enumerate(csv_reader): |
| label, title, description = row |
| |
| |
| |
| label = int(label) - 1 |
| text = " ".join((title, description)) |
| yield id_, {"text": text, "labels": label} |
|
|
|
|