| from typing import List, Tuple |
| from datasets import DatasetDict, Dataset, load_dataset |
|
|
|
|
| class BaseDataFormatter: |
| def get_nested(self) -> Tuple[List[List[str]], List[List[str]]]: |
| raise NotImplementedError |
|
|
| def get_flattened(self) -> Tuple[List[str], List[str]]: |
| raise NotImplementedError |
|
|
| def get_queries(self) -> Tuple[List[str], List[str]]: |
| raise NotImplementedError |
|
|
|
|
| class DataFormatter(BaseDataFormatter): |
| def __init__(self, dataset_path, split, query_key="queries", doc_key="documents"): |
| self.doc_dataset = None |
| self.queries_dataset = None |
| self._load_from_path(dataset_path, split, query_key, doc_key) |
| self.doc_dataset = self.doc_dataset.map(self.parse_id) |
|
|
| def _load_from_path(self, path, split, query_key, doc_key): |
| self.doc_dataset = load_dataset(path, doc_key, split=split) |
| self.queries_dataset = load_dataset(path, query_key, split=split) |
| |
|
|
| @staticmethod |
| def parse_id(sample): |
| doc_id, internal_id = sample["chunk_id"].split("_") |
| return {"doc_id": doc_id, "internal_id": int(internal_id)} |
|
|
| def get_nested(self) -> Tuple[List[List[str]], List[List[str]]]: |
| |
| return list( |
| self.doc_dataset.to_pandas().groupby("doc_id")["chunk"].apply(list) |
| ), list(self.doc_dataset.to_pandas().groupby("doc_id")["chunk_id"].apply(list)) |
|
|
| def get_flattened(self) -> Tuple[List[str], List[str]]: |
| |
| return self.doc_dataset["chunk"], self.doc_dataset["chunk_id"] |
|
|
| def get_queries(self) -> Tuple[List[str], List[str]]: |
| return self.queries_dataset["query"], self.queries_dataset["chunk_id"] |
|
|
|
|
| class BEIRDataFormatter(BaseDataFormatter): |
| def __init__( |
| self, |
| dataset_path, |
| split, |
| query_key="queries", |
| doc_key="corpus", |
| concat_num_docs=2, |
| ): |
| self.doc_dataset = None |
| self.queries_dataset = None |
| self.mapping = None |
| self._load_from_path(dataset_path, split, query_key, doc_key) |
| self.concat_num_docs = concat_num_docs |
|
|
| def _load_from_path(self, path, split, query_key, doc_key): |
| self.doc_dataset = load_dataset(path, doc_key, split=split) |
| self.queries_dataset = load_dataset(path, query_key, split=split) |
| mapping_dataset = load_dataset(path, "qrels", split=split) |
| self.mapping = { |
| query["query-id"]: query["corpus-id"] for query in mapping_dataset |
| } |
| |
|
|
| def get_nested(self) -> Tuple[List[List[str]], List[List[str]]]: |
|
|
| self.doc_dataset = self.doc_dataset.shuffle(seed=42) |
| idx = [] |
| for i in range(0, len(self.doc_dataset)): |
| idx.extend([i] * self.concat_num_docs) |
| idx = idx[: len(self.doc_dataset)] |
| self.doc_dataset = self.doc_dataset.add_column("doc_id", idx) |
|
|
| return list( |
| self.doc_dataset.to_pandas().groupby("doc_id")["text"].apply(list) |
| ), list(self.doc_dataset.to_pandas().groupby("doc_id")["_id"].apply(list)) |
|
|
| def get_flattened(self) -> Tuple[List[str], List[str]]: |
| |
| return self.doc_dataset["text"], self.doc_dataset["_id"] |
|
|
| def get_queries(self) -> Tuple[List[str], List[str]]: |
| gold_docs = [] |
| for query in self.queries_dataset: |
| gold_docs.append(self.mapping[query["_id"]]) |
| return self.queries_dataset["text"], gold_docs |
|
|