| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| |
| """RedPajama V2: Quality annotated Web Text Documents.""" |
|
|
| import gzip |
| import json |
| import traceback |
| from typing import List |
|
|
| import datasets |
| import pyarrow.parquet as pq |
|
|
| logger = datasets.logging.get_logger(__name__) |
|
|
| _DESCRIPTION = """\ |
| RedPajama V2: an Open Dataset for Training Large Language Models |
| """ |
|
|
| _URL_BASE = "https://data.together.xyz/redpajama-data-v2/v1.0.0" |
| _LANGUAGES = ("en", "de", "fr", "es", "it") |
| _MISSING_FILES_PATTERN = "urls/missing-{component}.txt" |
| _NUM_SHARDS = 5000 |
| _SUBSAMPLE_FILE_COUNTS = {"sample-10B": 1, "sample-100B": 10, "sample-1T": 100} |
|
|
| _CC_SNAPSHOT_IDS = ( |
| "2014-15", |
| "2014-23", |
| "2014-35", |
| "2014-41", |
| "2014-42", |
| "2014-49", |
| "2014-52", |
| "2015-14", |
| "2015-22", |
| "2015-27", |
| "2015-32", |
| "2015-35", |
| "2015-40", |
| "2015-48", |
| "2016-07", |
| "2016-18", |
| "2016-22", |
| "2016-26", |
| "2016-30", |
| "2016-36", |
| "2016-40", |
| "2016-44", |
| "2016-50", |
| "2017-04", |
| "2017-09", |
| "2017-17", |
| "2017-22", |
| "2017-26", |
| "2017-30", |
| "2017-34", |
| "2017-39", |
| "2017-43", |
| "2017-47", |
| "2017-51", |
| "2018-05", |
| "2018-09", |
| "2018-13", |
| "2018-17", |
| "2018-22", |
| "2018-26", |
| "2018-30", |
| "2018-34", |
| "2018-39", |
| "2018-43", |
| "2018-47", |
| "2018-51", |
| "2019-04", |
| "2019-09", |
| "2019-13", |
| "2019-18", |
| "2019-22", |
| "2019-26", |
| "2019-30", |
| "2019-35", |
| "2019-39", |
| "2019-43", |
| "2019-47", |
| "2019-51", |
| "2020-05", |
| "2020-10", |
| "2020-16", |
| "2020-24", |
| "2020-29", |
| "2020-34", |
| "2020-40", |
| "2020-45", |
| "2020-50", |
| "2021-04", |
| "2021-10", |
| "2021-17", |
| "2021-21", |
| "2021-25", |
| "2021-31", |
| "2021-39", |
| "2021-43", |
| "2021-49", |
| "2022-05", |
| "2022-21", |
| "2022-27", |
| "2022-33", |
| "2022-40", |
| "2022-49", |
| "2023-06", |
| "2023-14", |
| ) |
|
|
|
|
| class RedPajamaDataV2Config(datasets.BuilderConfig): |
| """BuilderConfig for RedPajama.""" |
|
|
| def __init__(self, *args, **kwargs): |
| """BuilderConfig for RedPajama. |
| Args: |
| **kwargs: keyword arguments forwarded to super. |
| """ |
| super(RedPajamaDataV2Config, self).__init__(**kwargs) |
| self.partition: str = kwargs.pop("partition", "all") |
| self.snapshots: List[str] = kwargs.pop("snapshots", _CC_SNAPSHOT_IDS) |
| self.languages: List[str] = kwargs.pop("languages", _LANGUAGES) |
|
|
|
|
| class RedPajamaV2(datasets.GeneratorBasedBuilder): |
| """RedPajama V2: Quality annotated Web Text Documents.""" |
|
|
| BUILDER_CONFIGS = [ |
| RedPajamaDataV2Config( |
| name="sample", |
| version=datasets.Version("1.0.0", ""), |
| description=f"RedPajamaV2 Sample", |
| ), |
| RedPajamaDataV2Config( |
| name="sample-10B", |
| version=datasets.Version("1.0.0", ""), |
| description=f"RedPajamaV2 Sample with 10B tokens", |
| ), |
| RedPajamaDataV2Config( |
| name="sample-100B", |
| version=datasets.Version("1.0.0", ""), |
| description=f"RedPajamaV2 Sample with 100B tokens", |
| ), |
| RedPajamaDataV2Config( |
| name="sample-1T", |
| version=datasets.Version("1.0.0", ""), |
| description=f"RedPajamaV2 Sample with 1T tokens", |
| ), |
| RedPajamaDataV2Config( |
| name="default", |
| version=datasets.Version("1.0.0", ""), |
| description=f"RedPajamaV2", |
| ), |
| ] |
|
|
| def _info(self): |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=datasets.Features( |
| { |
| "raw_content": datasets.Value("string"), |
| "doc_id": datasets.Value("string"), |
| "meta": datasets.Value("string"), |
| "quality_signals": datasets.Value("string"), |
| } |
| ), |
| supervised_keys=None, |
| ) |
|
|
| def _split_generators_sample(self, dl_manager): |
| |
| sample_base_tags_fp = dl_manager.download_and_extract( |
| "sample/sample_listings.txt" |
| ) |
| with open(sample_base_tags_fp, "r") as fd: |
| sample_base_tags = [line.strip() for line in fd] |
|
|
| |
| logger.info(f"Downloading {len(sample_base_tags)} documents files.") |
| documents_files = dl_manager.download( |
| { |
| base_tag: f"sample/documents/{base_tag}.json.gz" |
| for base_tag in sample_base_tags |
| } |
| ) |
|
|
| |
| logger.info(f"Downloading {len(sample_base_tags)} quality signals files.") |
| quality_signals_files = dl_manager.download( |
| { |
| base_tag: f"sample/quality_signals/{base_tag}.signals.json.gz" |
| for base_tag in sample_base_tags |
| } |
| ) |
|
|
| |
| logger.info(f"Downloading {len(sample_base_tags)} duplicates ids files.") |
| duplicates_ids_files = dl_manager.download( |
| { |
| base_tag: f"sample/duplicates/{base_tag}.duplicates.parquet" |
| for base_tag in sample_base_tags |
| } |
| ) |
|
|
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, |
| gen_kwargs={ |
| "base_tags": sample_base_tags, |
| "documents_files": documents_files, |
| "quality_signals_files": quality_signals_files, |
| "duplicates_ids_files": duplicates_ids_files, |
| }, |
| ) |
| ] |
|
|
| def _split_generators_full(self, dl_manager): |
| snapshots = getattr(self.config, "snapshots", _CC_SNAPSHOT_IDS) |
| languages = getattr(self.config, "languages", _LANGUAGES) |
| partition = getattr(self.config, "partition", "all") |
|
|
| if self.config.name in ("sample-10B", "sample-100B", "sample-1T"): |
| partition = "head_middle" |
| languages = _LANGUAGES |
| snapshots = _CC_SNAPSHOT_IDS |
| num_shards = _SUBSAMPLE_FILE_COUNTS[self.config.name] |
| else: |
| num_shards = _NUM_SHARDS |
|
|
| if partition == "all": |
| partitions = ["head", "middle", "tail"] |
| elif partition == "head_middle": |
| partitions = ["head", "middle"] |
| elif partition == "tail": |
| partitions = [partition] |
| else: |
| raise ValueError(f"invalid partition: {partition}") |
|
|
| |
| |
| missing_files_paths = dl_manager.download_and_extract( |
| { |
| component: _MISSING_FILES_PATTERN.format(component=component) |
| for component in ("documents", "signals", "duplicates") |
| } |
| ) |
|
|
| missing_files = {} |
| for component, missing_file in missing_files_paths.items(): |
| with open(missing_file, "r", encoding="utf-8") as f: |
| missing_files[component] = set(line.strip() for line in f) |
|
|
| |
| documents_urls = {} |
| quality_signals_urls = {} |
| duplicates_ids_urls = {} |
| base_tags = [] |
|
|
| for lang in languages: |
| for snapshot in snapshots: |
| for part in partitions: |
| for n in range(num_shards): |
| base_tag = f"{snapshot}/{n:04d}/{lang}_{part}" |
| base_tags.append(base_tag) |
|
|
| |
| url = f"{_URL_BASE}/documents/{base_tag}.json.gz" |
| if url not in missing_files["documents"]: |
| documents_urls[base_tag] = url |
|
|
| |
| url = f"{_URL_BASE}/quality_signals/{base_tag}.signals.json.gz" |
| if url not in missing_files["signals"]: |
| quality_signals_urls[base_tag] = url |
|
|
| |
| url = f"{_URL_BASE}/duplicates/{base_tag}.duplicates.parquet" |
| if url not in missing_files["duplicates"]: |
| duplicates_ids_urls[base_tag] = url |
|
|
| |
| logger.info(f"Downloading {len(documents_urls)} documents files.") |
| documents_files = dl_manager.download(documents_urls) |
|
|
| |
| logger.info(f"Downloading {len(quality_signals_urls)} quality signals files.") |
| quality_signals_files = dl_manager.download(quality_signals_urls) |
|
|
| |
| logger.info(f"Downloading {len(duplicates_ids_urls)} duplicates ids files.") |
| duplicates_ids_files = dl_manager.download(duplicates_ids_urls) |
|
|
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, |
| gen_kwargs={ |
| "base_tags": base_tags, |
| "documents_files": documents_files, |
| "quality_signals_files": quality_signals_files, |
| "duplicates_ids_files": duplicates_ids_files, |
| }, |
| ) |
| ] |
|
|
| def _split_generators(self, dl_manager): |
| if self.config.name == "sample": |
| return self._split_generators_sample(dl_manager) |
|
|
| return self._split_generators_full(dl_manager) |
|
|
| def _generate_examples( |
| self, base_tags, documents_files, quality_signals_files, duplicates_ids_files |
| ): |
| key = 0 |
| for base_tag in base_tags: |
| doc_file = documents_files.get(base_tag) |
| qs_file = quality_signals_files.get(base_tag) |
| dupe_file = duplicates_ids_files.get(base_tag) |
|
|
| if doc_file is None: |
| continue |
|
|
| for sample in self.__get_generator(base_tag, doc_file, qs_file, dupe_file): |
| yield key, sample |
| key += 1 |
|
|
| def __get_generator(self, base_tag, doc_file, qs_file, dupe_file): |
| if "_tail" in base_tag: |
| yield from self._handle_tail(base_tag, doc_file, qs_file, dupe_file) |
| else: |
| yield from self._handle_head_middle(base_tag, doc_file, qs_file, dupe_file) |
|
|
| def _handle_tail(self, base_tag, doc_file, qs_file, dupe_file): |
| try: |
| with gzip.open(doc_file, "rt", encoding="utf-8") as df: |
| for row, doc in enumerate(df): |
| doc_id = f"{base_tag}.json.gz/{row}" |
| try: |
| yield self.handle_record("tail", doc_id, doc, None, None) |
| except Exception as e: |
| logger.warning(f"failed handling row {row} in {doc_file}") |
| traceback.print_exc() |
| continue |
|
|
| except gzip.BadGzipFile as e: |
| |
| print(f"BadGzipFile: {doc_file, qs_file}") |
| traceback.print_exc() |
| return |
|
|
| def _handle_head_middle(self, base_tag, doc_file, qs_file, dupe_file): |
| if qs_file is None: |
| yield from self._handle_tail(base_tag, doc_file, None, None) |
| return |
|
|
| |
| try: |
| with open(dupe_file, "rb") as df: |
| duplicates = set( |
| pq.read_table(df, columns=["doc_id"], use_pandas_metadata=False)[ |
| "doc_id" |
| ].to_pylist() |
| ) |
| except Exception as e: |
| logger.warning(f"no duplicate ids found for {base_tag}") |
| duplicates = set() |
|
|
| try: |
| with gzip.open(doc_file, "rt", encoding="utf-8") as df: |
| with gzip.open(qs_file, "rt", encoding="utf-8") as qf: |
| for row, (doc, qs) in enumerate(zip(df, qf)): |
| doc_id = f"{base_tag}.json.gz/{row}" |
|
|
| try: |
| yield self.handle_record( |
| part="head_middle", |
| doc_id=doc_id, |
| doc=doc, |
| qs=qs, |
| is_duplicate=doc_id in duplicates, |
| ) |
| except Exception as e: |
| logger.warning( |
| f"failed handling row {row} in {doc_file} ({qs_file})" |
| ) |
| traceback.print_exc() |
| continue |
|
|
| except gzip.BadGzipFile as e: |
| |
| print(f"BadGzipFile: {doc_file, qs_file}") |
| traceback.print_exc() |
| return |
|
|
| @staticmethod |
| def handle_record(part, doc_id, doc, qs, is_duplicate=None): |
| doc = json.loads(doc) |
| qs = json.loads(qs) if qs is not None else {} |
|
|
| meta = { |
| "url": doc["url"], |
| "partition": part, |
| "language": doc["language"], |
| "source_domain": doc["source_domain"], |
| "date_download": doc["date_download"], |
| "digest": doc["digest"], |
| } |
|
|
| quality_signals = qs.get("quality_signals", {}) |
| quality_signals["is_duplicate"] = is_duplicate |
|
|
| return { |
| "raw_content": doc["raw_content"], |
| "doc_id": doc_id, |
| "meta": json.dumps(meta), |
| "quality_signals": json.dumps(quality_signals), |
| } |
|
|