File size: 7,113 Bytes
56598a1 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 | # Path Configuration
from tools.preprocess import *
# Processing context
trait = "Bladder_Cancer"
cohort = "GSE253531"
# Input paths
in_trait_dir = "../DATA/GEO/Bladder_Cancer"
in_cohort_dir = "../DATA/GEO/Bladder_Cancer/GSE253531"
# Output paths
out_data_file = "./output/z1/preprocess/Bladder_Cancer/GSE253531.csv"
out_gene_data_file = "./output/z1/preprocess/Bladder_Cancer/gene_data/GSE253531.csv"
out_clinical_data_file = "./output/z1/preprocess/Bladder_Cancer/clinical_data/GSE253531.csv"
json_path = "./output/z1/preprocess/Bladder_Cancer/cohort_info.json"
# Step 1: Initial Data Loading
from tools.preprocess import *
# 1. Identify the paths to the SOFT file and the matrix file
soft_file, matrix_file = geo_get_relevant_filepaths(in_cohort_dir)
# 2. Read the matrix file to obtain background information and sample characteristics data
background_prefixes = ['!Series_title', '!Series_summary', '!Series_overall_design']
clinical_prefixes = ['!Sample_geo_accession', '!Sample_characteristics_ch1']
background_info, clinical_data = get_background_and_clinical_data(matrix_file, background_prefixes, clinical_prefixes)
# 3. Obtain the sample characteristics dictionary from the clinical dataframe
sample_characteristics_dict = get_unique_values_by_row(clinical_data)
# 4. Explicitly print out all the background information and the sample characteristics dictionary
print("Background Information:")
print(background_info)
print("Sample Characteristics Dictionary:")
print(sample_characteristics_dict)
# Step 2: Dataset Analysis and Clinical Feature Extraction
import re
# 1. Gene Expression Data Availability
is_gene_available = True # Gene expression microarrays mentioned in background info
# 2. Variable Availability and Data Type Conversion
# From sample characteristics, only 'lab' and 'tcga_molecular_subtype' are available.
# Trait is Bladder_Cancer; all samples are bladder cancer (constant), so not available for association.
trait_row = None
age_row = None
gender_row = None
def _after_colon(value):
if value is None:
return None
s = str(value)
parts = s.split(":", 1)
val = parts[1] if len(parts) > 1 else parts[0]
return val.strip()
def convert_trait(value):
v = _after_colon(value)
if v is None:
return None
v_low = v.lower()
# Map controls/normal to 0; cancer/tumor-related to 1
control_keys = ["normal", "healthy", "control", "benign", "non-cancer", "no cancer", "adjacent normal"]
case_keys = ["cancer", "tumor", "carcinoma", "malignant", "mibc", "bladder cancer", "urothelial"]
if any(k in v_low for k in control_keys):
return 0
if any(k in v_low for k in case_keys):
return 1
return None
def convert_age(value):
v = _after_colon(value)
if v is None:
return None
# Extract first number that looks like an age
match = re.search(r"(\d+(\.\d+)?)", v)
if not match:
return None
try:
age = float(match.group(1))
# Basic sanity check for human age
if 0 <= age <= 120:
return age
except:
pass
return None
def convert_gender(value):
v = _after_colon(value)
if v is None:
return None
v_low = v.lower()
if v_low in ["female", "f", "woman", "women"]:
return 0
if v_low in ["male", "m", "man", "men"]:
return 1
return None
# 3. Save Metadata (initial filtering)
is_trait_available = trait_row is not None
_ = validate_and_save_cohort_info(
is_final=False,
cohort=cohort,
info_path=json_path,
is_gene_available=is_gene_available,
is_trait_available=is_trait_available
)
# 4. Clinical Feature Extraction: skipped because trait_row is None
# Step 3: Gene Data Extraction
# 1. Use the get_genetic_data function from the library to get the gene_data from the matrix_file previously defined.
gene_data = get_genetic_data(matrix_file)
# 2. Print the first 20 row IDs (gene or probe identifiers) for future observation.
print(gene_data.index[:20])
# Step 4: Gene Identifier Review
print("requires_gene_mapping = True")
# Step 5: Gene Annotation
# 1. Use the 'get_gene_annotation' function from the library to get gene annotation data from the SOFT file.
gene_annotation = get_gene_annotation(soft_file)
# 2. Use the 'preview_df' function from the library to preview the data and print out the results.
print("Gene annotation preview:")
print(preview_df(gene_annotation))
# Step 6: Gene Identifier Mapping
# Determine the appropriate columns for probe ID and gene symbols based on the annotation preview
probe_col = 'ID' # Matches probe IDs in gene_data (e.g., '2315554')
gene_symbol_col = 'gene_assignment' # Contains gene symbol info within mixed text
# 2. Build mapping dataframe
mapping_df = get_gene_mapping(gene_annotation, prob_col=probe_col, gene_col=gene_symbol_col)
# 3. Apply mapping to convert probe-level data to gene-level expression
gene_data = apply_gene_mapping(expression_df=gene_data, mapping_df=mapping_df)
# Step 7: Data Normalization and Linking
import os
# 1. Normalize the obtained gene data and save to disk
normalized_gene_data = normalize_gene_symbols_in_index(gene_data)
os.makedirs(os.path.dirname(out_gene_data_file), exist_ok=True)
normalized_gene_data.to_csv(out_gene_data_file)
# Guard for missing clinical data (trait not available in this cohort per Step 2)
trait_row_val = globals().get('trait_row', None)
selected_clinical_data_obj = globals().get('selected_clinical_data', None)
linked_data = None # Ensure the variable exists for downstream compatibility
if (trait_row_val is None) or (selected_clinical_data_obj is None):
# Record unavailability and skip linking and downstream processing
_ = validate_and_save_cohort_info(
is_final=False,
cohort=cohort,
info_path=json_path,
is_gene_available=True,
is_trait_available=False
)
else:
# 2. Link the clinical and genetic data
linked_data = geo_link_clinical_genetic_data(selected_clinical_data_obj, normalized_gene_data)
# If trait column is missing after linking, treat as unavailable and skip
if trait not in linked_data.columns:
_ = validate_and_save_cohort_info(
is_final=False,
cohort=cohort,
info_path=json_path,
is_gene_available=True,
is_trait_available=False
)
else:
# 3. Handle missing values in the linked data
linked_data = handle_missing_values(linked_data, trait)
# 4. Determine whether the trait and demographic features are severely biased
is_trait_biased, unbiased_linked_data = judge_and_remove_biased_features(linked_data, trait)
# 5. Conduct quality check and save the cohort information
is_usable = validate_and_save_cohort_info(
True, cohort, json_path, True, True, is_trait_biased, unbiased_linked_data
)
# 6. If usable, save the linked data
if is_usable:
os.makedirs(os.path.dirname(out_data_file), exist_ok=True)
unbiased_linked_data.to_csv(out_data_file) |