| | import pandas as pd |
| | from pathlib import Path |
| | from datasets import Features, Value, Sequence |
| | from datasets import DatasetInfo, GeneratorBasedBuilder, SplitGenerator, Split |
| | import pickle |
| | import pyarrow as pa |
| | import pyarrow.parquet as pq |
| |
|
| |
|
| | class FcsParquetIO: |
| | """A class including all methods to write and read parquets transformed from one fcs by FileTransformer.fcs_to_parquet. |
| | """ |
| | @staticmethod |
| | def read_parquets(read_dir: Path) -> tuple[dict, pd.DataFrame]: |
| | """A method to read a group of parquets, which are transformed and partitioned from one fcs by FileTransformer.fcs_to_parquet, from the given directory. |
| | |
| | Args: |
| | read_dir (Path): The directory including all parquets transformed and partitioned from one fcs. |
| | |
| | Returns: |
| | tuple[dict, pd.DataFrame]: The tuple includes two elements. The first one is the meta data from the original fcs. The second one is the event data from the original fcs. |
| | """ |
| | parquet_filepaths = list(read_dir.glob("*.parquet")) |
| | table = {"meta": None, "event": pd.DataFrame()} |
| | for parquet_filepath in parquet_filepaths: |
| | table = pq.read_table(parquet_filepath) |
| |
|
| | parquet_filename = str( |
| | parquet_filepath.name).replace(".parquet", '') |
| | if "part_001" in parquet_filename: |
| | raw_meta = table.schema.metadata or {} |
| | table["meta"] = {k.decode('utf8'): v.decode('utf8') |
| | for k, v in raw_meta.items()} |
| |
|
| | table["event"] = pd.concat( |
| | [table["event"], table.to_pandas()], axis=0, ignore_index=True) |
| |
|
| | return table["meta"], table["event"] |
| |
|
| |
|
| | feature_fields = { |
| | "flow_id": Value("string"), |
| | "original_id": Value("string"), |
| | "specimen": Value("string"), |
| | "purpose": Value("string"), |
| | "site": Value("string"), |
| | "machine": Value("string"), |
| | "color_counts": Value("int32"), |
| | "sampling_date": Value("date32"), |
| | "measuring_date": Value("date32"), |
| | "panel": Value("string"), |
| | "label": Value("string"), |
| | "sublabel": Sequence(Value("string")), |
| | "RDP": Value("float32"), |
| |
|
| | "tube_counts": Value("int32"), |
| | "tube_event_counts": Value("string"), |
| | "tube_channels": Value("string"), |
| |
|
| | "data": Value("binary") |
| | } |
| | features = Features(feature_fields) |
| |
|
| |
|
| | class SCMH_Malignancy(GeneratorBasedBuilder): |
| | def _info(self): |
| | return DatasetInfo( |
| | features=features, |
| | description="This dataset is a project of SCMH from datalake with various training conditions, contained in different branches.", |
| | homepage="", |
| | citation="", |
| | license="" |
| | ) |
| |
|
| | def _split_generators(self, dl_manager): |
| | data_dir = Path(self.base_path) / "fcs" |
| | return [ |
| | SplitGenerator( |
| | name=Split.TRAIN, |
| | gen_kwargs={"data_dir": data_dir}, |
| | ) |
| | ] |
| |
|
| | def _generate_examples(self, data_dir: Path): |
| | sample_metas = pd.read_csv(Path(self.base_path) / "sample_info.csv") |
| |
|
| | for column in sample_metas.columns.to_list(): |
| | if column in feature_fields.keys(): |
| | feature_field = column |
| | feature_type = feature_fields[feature_field] |
| | if feature_type is Value("int32"): |
| | sample_metas[feature_field] = sample_metas[feature_field].astype( |
| | "Int32") |
| | elif feature_type is Value("float32"): |
| | sample_metas[feature_field] = sample_metas[feature_field].astype( |
| | "float32") |
| | elif feature_type is Value("date32"): |
| | sample_metas[feature_field] = pd.to_datetime( |
| | sample_metas[feature_field], format="%m/%d/%Y").dt.date |
| | elif feature_type is Value("string"): |
| | sample_metas[feature_field] = sample_metas[feature_field].astype( |
| | "string") |
| |
|
| | for column in feature_fields.keys(): |
| | if column not in sample_metas.columns.to_list(): |
| | sample_metas[column] = None |
| |
|
| | sublabel_columns = [ |
| | column for column in sample_metas.columns.to_list() if "sublabel" in column] |
| | if len(sublabel_columns) > 0: |
| | sample_metas["sublabel"] = sample_metas.apply( |
| | lambda row: [str(row[col]) for col in sublabel_columns], axis=1) |
| | sample_metas = sample_metas.drop(columns=sublabel_columns) |
| |
|
| | for sample_index, sample_dir in enumerate(sorted(list(data_dir.glob('*')))): |
| | if sample_dir.is_dir(): |
| | sample_meta = sample_metas[sample_metas["flow_id"] |
| | == str(sample_dir.name)].to_dict(orient="records")[0] |
| | sample_tubes = {} |
| | for tube_dir in sorted(list(sample_dir.glob('*'))): |
| | if tube_dir.is_dir(): |
| | meta, event = FcsParquetIO.read_parquets(tube_dir) |
| | sample_tubes[tube_dir.name] = { |
| | "meta": meta, "event": event} |
| | sample_meta["data"] = pickle.dumps(sample_tubes) |
| |
|
| | yield sample_index, sample_meta |
| |
|