# build large-scale data in scBank format from a group of AnnData objects # %% import gc import json from pathlib import Path import argparse import shutil import traceback from typing import Dict, List, Optional import warnings import numpy as np import os import scanpy as sc import sys sys.path.insert(0, "../../") import scgpt as scg from scgpt import scbank # %% parser = argparse.ArgumentParser( description="Build large-scale data in scBank format from a group of AnnData objects" ) parser.add_argument( "--input-dir", type=str, required=True, help="Directory containing AnnData objects", ) parser.add_argument( "--output-dir", type=str, default="./data.scb", help="Directory to save scBank data, by default will make a directory named " "data.scb in the current directory", ) parser.add_argument( "--include-files", type=str, nargs="*", help="Space separated file names to include, default to all files in input_dir", ) parser.add_argument( "--metainfo", type=str, default=None, help="Json file containing meta information for each dataset, default to None.", ) # vocabulary parser.add_argument( "--vocab-file", type=str, default=None, help="File containing the gene vocabulary, default to None. If None, will " "use the default gene vocabulary from scGPT, which use HGNC gene symbols.", ) parser.add_argument( "--N", type=int, default=10000, help="Hyperparam for filtering genes, default to 10000.", ) # if scg.utils.isnotebook(): # args = parser.parse_args( # [ # "--input-dir", # "./datasets/", # "--output-dir", # "./databanks/", # "--include-files", # "f72958f5-7f42-4ebb-98da-445b0c6de516.h5ad", # "--metainfo", # "./metainfo.json", # "--vocab-file", # "../../scgpt/tokenizer/default_cellxgene_vocab.json", # ] # ) # else: args = parser.parse_args() """command line example python build_large_scale_data.py \ --input-dir ./datasets/ \ --output-dir ./databanks/ \ --metainfo ./metainfo.json \ --vocab-file ../../scgpt/tokenizer/default_cellxgene_vocab.json """ # %% print(args) input_dir = Path(args.input_dir) output_dir = Path(args.output_dir) files = [f for f in input_dir.glob("*.h5ad")] print(f"Found {len(files)} files in {input_dir}") if args.include_files is not None: files = [f for f in files if f.name in args.include_files] if args.metainfo is not None: metainfo = json.load(open(args.metainfo)) files = [f for f in files if f.stem in metainfo] include_obs = { f.stem: {"disease": metainfo[f.stem]["include_disease"]} for f in files if "include_disease" in metainfo[f.stem] } if args.vocab_file is None: vocab = scg.tokenizer.get_default_gene_vocab() else: vocab = scg.tokenizer.GeneVocab.from_file(args.vocab_file) # %% [markdown] # # preprocessing data def preprocess( adata: sc.AnnData, main_table_key: str = "counts", include_obs: Optional[Dict[str, List[str]]] = None, N=10000, ) -> sc.AnnData: """ Preprocess the data for scBank. This function will modify the AnnData object in place. Args: adata: AnnData object to preprocess main_table_key: key in adata.layers to store the main table include_obs: dict of column names and values to include in the main table Returns: The preprocessed AnnData object """ if include_obs is not None: # include only cells that have the specified values in the specified columns for col, values in include_obs.items(): adata = adata[adata.obs[col].isin(values)] # filter genes sc.pp.filter_genes(adata, min_counts=(3 / 10000) * N) # TODO: add binning in sparse matrix and save in separate datatable # preprocessor = Preprocessor( # use_key="X", # the key in adata.layers to use as raw data # filter_gene_by_counts=False, # step 1 # filter_cell_by_counts=False, # step 2 # normalize_total=False, # 3. whether to normalize the raw data and to what sum # log1p=False, # 4. whether to log1p the normalized data # binning=51, # 6. whether to bin the raw data and to what number of bins # result_binned_key="X_binned", # the key in adata.layers to store the binned data # ) # preprocessor(adata) adata.layers[main_table_key] = adata.X.copy() # preserve counts # sc.pp.normalize_total(adata, target_sum=1e4) # sc.pp.log1p(adata) # adata.raw = adata # freeze the state in `.raw` # apply a hard clip to the data for now print( f"original mean and max of counts: {adata.layers[main_table_key].mean():.2f}, " f"{adata.layers[main_table_key].max():.2f}" ) # if isinstance(adata.layers[main_table_key], np.ndarray): # adata.layers[main_table_key] = adata.layers[main_table_key].clip(0, 30) # else: # assume it is a sparse matrix # adata.layers[main_table_key].data = adata.layers[main_table_key].data.clip(0, 30) return adata # %% main_table_key = "counts" token_col = "feature_name" for f in files: try: adata = sc.read(f, cache=True) adata = preprocess(adata, main_table_key, N=args.N) print(f"read {adata.shape} valid data from {f.name}") # TODO: CHECK AND EXPAND VOCABULARY IF NEEDED # NOTE: do not simply expand, need to check whether to use the same style of gene names # BUILD SCBANK DATA db = scbank.DataBank.from_anndata( adata, vocab=vocab, to=output_dir / f"{f.stem}.scb", main_table_key=main_table_key, token_col=token_col, immediate_save=False, ) db.meta_info.on_disk_format = "parquet" # sync all to disk db.sync() # clean up del adata del db gc.collect() except Exception as e: traceback.print_exc() warnings.warn(f"failed to process {f.name}: {e}") shutil.rmtree(output_dir / f"{f.stem}.scb", ignore_errors=True) # or run scbank.DataBank.batch_from_anndata(files, to=args.output_dir) # %% # test loading from disk # db = scbank.DataBank.from_path(args.output_dir) # %% run this to copy all parquet datatables to a single directory target_dir = output_dir / f"all_{main_table_key}" target_dir.mkdir(exist_ok=True) for f in files: output_parquet_dt = ( output_dir / f"{f.stem}.scb" / f"{main_table_key}.datatable.parquet" ) if output_parquet_dt.exists(): os.symlink(output_parquet_dt, target_dir / f"{f.stem}.datatable.parquet")