BCCasebook_binary / BCCasebook_binary.py
RobinChu's picture
upload data
7868a18 verified
import pandas as pd
from pathlib import Path
from datasets import Features, Value, Sequence
from datasets import DatasetInfo, GeneratorBasedBuilder, SplitGenerator, Split
import pickle
import pyarrow as pa
import pyarrow.parquet as pq
class FcsParquetIO:
"""A class including all methods to write and read parquets transformed from one fcs by FileTransformer.fcs_to_parquet.
"""
@staticmethod
def read_parquets(read_dir: Path) -> tuple[dict, pd.DataFrame]:
"""A method to read a group of parquets, which are transformed and partitioned from one fcs by FileTransformer.fcs_to_parquet, from the given directory.
Args:
read_dir (Path): The directory including all parquets transformed and partitioned from one fcs.
Returns:
tuple[dict, pd.DataFrame]: The tuple includes two elements. The first one is the meta data from the original fcs. The second one is the event data from the original fcs.
"""
parquet_filepaths = list(read_dir.glob("*.parquet"))
table = {"meta": None, "event": pd.DataFrame()}
for parquet_filepath in parquet_filepaths:
table = pq.read_table(parquet_filepath)
parquet_filename = str(
parquet_filepath.name).replace(".parquet", '')
if "part_001" in parquet_filename:
raw_meta = table.schema.metadata or {}
table["meta"] = {k.decode('utf8'): v.decode('utf8')
for k, v in raw_meta.items()}
table["event"] = pd.concat(
[table["event"], table.to_pandas()], axis=0, ignore_index=True)
return table["meta"], table["event"]
feature_fields = {
"flow_id": Value("string"),
"original_id": Value("string"),
"specimen": Value("string"),
"purpose": Value("string"),
"site": Value("string"),
"machine": Value("string"),
"color_counts": Value("int32"),
"sampling_date": Value("date32"),
"measuring_date": Value("date32"),
"panel": Value("string"),
"label": Value("string"),
"sublabel": Sequence(Value("string")),
"RDP": Value("float32"),
"tube_counts": Value("int32"),
"tube_event_counts": Value("string"),
"tube_channels": Value("string"),
"data": Value("binary")
}
features = Features(feature_fields)
class BCCasebook_binary(GeneratorBasedBuilder):
def _info(self):
return DatasetInfo(
features=features,
description="This dataset is a project of BC Casebook data labeled for Malignant and Non-malignant.",
homepage="",
citation="",
license=""
)
def _split_generators(self, dl_manager):
data_dir = Path(self.base_path) / "fcs"
return [
SplitGenerator(
name=Split.TRAIN,
gen_kwargs={"data_dir": data_dir},
)
]
def _generate_examples(self, data_dir: Path):
sample_metas = pd.read_csv(Path(self.base_path) / "sample_metas.csv")
for column in sample_metas.columns.to_list():
if column in feature_fields.keys():
feature_field = column
feature_type = feature_fields[feature_field]
if feature_type is Value("int32"):
sample_metas[feature_field] = sample_metas[feature_field].astype(
"Int32")
elif feature_type is Value("float32"):
sample_metas[feature_field] = sample_metas[feature_field].astype(
"float32")
elif feature_type is Value("date32"):
sample_metas[feature_field] = pd.to_datetime(
sample_metas[feature_field], format="%m/%d/%Y").dt.date
elif feature_type is Value("string"):
sample_metas[feature_field] = sample_metas[feature_field].astype(
"string")
for column in feature_fields.keys():
if column not in sample_metas.columns.to_list():
sample_metas[column] = None
sublabel_columns = [
column for column in sample_metas.columns.to_list() if "sublabel" in column]
if len(sublabel_columns) > 0:
sample_metas["sublabel"] = sample_metas.apply(
lambda row: [str(row[col]) for col in sublabel_columns], axis=1)
sample_metas = sample_metas.drop(columns=sublabel_columns)
for sample_index, sample_dir in enumerate(sorted(list(data_dir.glob('*')))):
if sample_dir.is_dir():
sample_meta = sample_metas[sample_metas["flow_id"]
== str(sample_dir.name)].to_dict(orient="records")[0]
sample_tubes = {}
for tube_dir in sorted(list(sample_dir.glob('*'))):
if tube_dir.is_dir():
meta, event = FcsParquetIO.read_parquets(tube_dir)
sample_tubes[tube_dir.name] = {
"meta": meta, "event": event}
sample_meta["data"] = pickle.dumps(sample_tubes)
yield sample_index, sample_meta