import pandas as pd from pathlib import Path from datasets import Features, Value, Sequence from datasets import DatasetInfo, GeneratorBasedBuilder import pickle import pyarrow as pa import pyarrow.parquet as pq class fcsParquetIO: def __init__(self): pass def write_parquets(self, tables: dict[str, pa.Table], dump_dir: Path): if not dump_dir.exists(): dump_dir.mkdir(parents=True, exist_ok=True) for parquet_filename, table in tables.items(): pq.write_table(table, dump_dir / f"{parquet_filename}.parquet") def read_parquets(self, read_dir: Path) -> tuple[dict, pd.DataFrame]: parquet_filepaths = list(read_dir.glob("*.parquet")) table = {"meta": None, "event": pd.DataFrame()} for parquet_filepath in parquet_filepaths: table = pq.read_table(parquet_filepath) parquet_filename = str( parquet_filepath.name).replace(".parquet", '') if "part_001" in parquet_filename: raw_meta = table.schema.metadata or {} table["meta"] = {k.decode('utf8'): v.decode('utf8') for k, v in raw_meta.items()} table["event"] = pd.concat( [table["event"], table.to_pandas()], axis=0, ignore_index=True) return table["meta"], table["event"] features = Features({ "flow_id": Value("string"), "original_id": Value("string"), "specimen": Value("string"), "purpose": Value("string"), "site": Value("string"), "machine": Value("string"), "color_counts": Value("int16"), "sampling_date": Value("date32"), "measuring_date": Value("date32"), "panel": Value("string"), "label": Value("string"), "sublabel": Sequence(Value("string")), "RDP": Value("float64"), "tube_counts": Value("int16"), "tube_event_counts": Value("string"), "tube_channels": Value("string"), "data": Value("binary"), }) class BD_LST(GeneratorBasedBuilder): def _info(self): return DatasetInfo( features=features, description="This dataset is a project of BD LST data from datalake with various training conditions, contained in different branches.", homepage="", citation="", license="" ) def _generate_examples(self, dataset_dir: Path): sample_metas = pd.read_csv(dataset_dir / "sample_metas.csv") data_dir = dataset_dir / "fcs" for sample_dir in sorted(list(data_dir.glob('*'))): if sample_dir.is_dir(): sample_meta = sample_metas[sample_metas["flow_id"] == sample_dir.name].to_dict() sample_tubes = {} for tube_dir in sorted(list(sample_dir.glob('*'))): if tube_dir.is_dir(): meta, event = fcsParquetIO().read_parquets(tube_dir) sample_tubes[tube_dir.name] = { "meta": meta, "event": event} sample_data = pickle.dumps(sample_tubes) yield sample_meta | sample_data