| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | |
| | """The CC-News dataset is based on Common Crawl News Dataset by Sebastian Nagel""" |
| |
|
| | import json |
| | import os |
| | import tarfile |
| | from fnmatch import fnmatch |
| |
|
| | import datasets |
| |
|
| | def custom_iter_archive(path_or_buf, _filter=lambda x: True): |
| | def _iter_archive(f): |
| | stream = tarfile.open(fileobj=f, mode="r|*") |
| | for i, tarinfo in enumerate(stream): |
| | if not _filter(i): |
| | continue |
| | file_path = tarinfo.name |
| | if not tarinfo.isreg(): |
| | continue |
| | if file_path is None: |
| | continue |
| | if os.path.basename(file_path).startswith(".") or os.path.basename(file_path).startswith("__"): |
| | |
| | continue |
| | file_obj = stream.extractfile(tarinfo) |
| | yield file_path, file_obj |
| | stream.members = [] |
| | del stream |
| |
|
| | if hasattr(path_or_buf, "read"): |
| | yield from _iter_archive(path_or_buf) |
| | else: |
| | with open(path_or_buf, "rb") as f: |
| | yield from _iter_archive(f) |
| |
|
| | logger = datasets.logging.get_logger(__name__) |
| |
|
| |
|
| | _DESCRIPTION = """\ |
| | CC-News containing news articles from news sites all over the world \ |
| | The data is available on AWS S3 in the Common Crawl bucket at /crawl-data/CC-NEWS/. \ |
| | This version of the dataset has 708241 articles. It represents a small portion of English \ |
| | language subset of the CC-News dataset created using news-please(Hamborg et al.,2017) to \ |
| | collect and extract English language portion of CC-News. |
| | """ |
| |
|
| | _CITATION = """\ |
| | @InProceedings{Hamborg2017, |
| | author = {Hamborg, Felix and Meuschke, Norman and Breitinger, Corinna and Gipp, Bela}, |
| | title = {news-please: A Generic News Crawler and Extractor}, |
| | year = {2017}, |
| | booktitle = {Proceedings of the 15th International Symposium of Information Science}, |
| | location = {Berlin}, |
| | doi = {10.5281/zenodo.4120316}, |
| | pages = {218--223}, |
| | month = {March} |
| | } |
| | """ |
| | _PROJECT_URL = "https://commoncrawl.org/2016/10/news-dataset-available/" |
| | _DOWNLOAD_URL = "https://storage.googleapis.com/huggingface-nlp/datasets/cc_news/cc_news.tar.gz" |
| |
|
| |
|
| | class CCNewsConfig(datasets.BuilderConfig): |
| | """BuilderConfig for CCNews.""" |
| |
|
| | def __init__(self, **kwargs): |
| | """BuilderConfig for CCNews. |
| | Args: |
| | **kwargs: keyword arguments forwarded to super. |
| | """ |
| | super(CCNewsConfig, self).__init__(version=datasets.Version("1.0.0", ""), **kwargs) |
| |
|
| |
|
| | class CCNews(datasets.GeneratorBasedBuilder): |
| | """CC-News dataset.""" |
| |
|
| | BUILDER_CONFIGS = [ |
| | CCNewsConfig( |
| | name="plain_text", |
| | description="Plain text", |
| | ), |
| | CCNewsConfig( |
| | name="plain_text_sentences", |
| | description="Plain text (sentence level)", |
| | ) |
| | ] |
| |
|
| | def _info(self): |
| | return datasets.DatasetInfo( |
| | description=_DESCRIPTION, |
| | features=datasets.Features( |
| | { |
| | "text": datasets.Value("string"), |
| | } |
| | ), |
| | supervised_keys=None, |
| | homepage=_PROJECT_URL, |
| | citation=_CITATION, |
| | ) |
| |
|
| | def _split_generators(self, dl_manager): |
| | archive = dl_manager.download(_DOWNLOAD_URL) |
| | |
| | train_filter = lambda x : (x%10) < 8 |
| | val_filter = lambda x: (x%10) == 8 |
| | test_filter = lambda x: (x%10) == 9 |
| | |
| | level = "doc" if self.config.name == "plain_text" else "sentence" |
| |
|
| | return [ |
| | datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"files": custom_iter_archive(archive, train_filter), "level": level}), |
| | datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={"files": custom_iter_archive(archive, val_filter), "level": level}), |
| | datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"files": custom_iter_archive(archive, test_filter), "level": level}), |
| | ] |
| |
|
| | def _generate_examples(self, files, level): |
| | id_ = 0 |
| | for article_file_path, f in files: |
| | if fnmatch(os.path.basename(article_file_path), "*.json"): |
| | article = json.load(f) |
| | if level == "sentence": |
| | full_article = article["maintext"].strip() if article["maintext"] is not None else "" |
| | doc_dict = {} |
| | for sent in full_article.split("\n"): |
| | doc_dict["text"] = sent |
| | yield id_, doc_dict |
| | id_ += 1 |
| | else: |
| | yield id_, { |
| | "text": article["maintext"].strip() if article["maintext"] is not None else "", |
| | } |
| | id_ += 1 |
| |
|
| |
|