Datasets:
Formats:
parquet
Languages:
English
Size:
10M - 100M
Tags:
biology
chemistry
drug-discovery
clinical-trials
protein-protein-interaction
gene-essentiality
License:
| #!/usr/bin/env python | |
| """Download DepMap data files for GE domain. | |
| Downloads from DepMap portal API and Figshare: | |
| - CRISPRGeneEffect.csv (Chronos scores) | |
| - CRISPRGeneDependency.csv (dependency probability) | |
| - Model.csv (cell line metadata) | |
| - AchillesCommonEssentialControls.csv | |
| - AchillesNonessentialControls.csv | |
| - D2_combined_gene_dep_scores.csv (DEMETER2 RNAi) | |
| - Omics files (expression, CN, mutations) | |
| Uses the DepMap download API to resolve file URLs dynamically. | |
| """ | |
| import argparse | |
| import csv | |
| import io | |
| import sys | |
| from pathlib import Path | |
| import requests | |
| _PROJECT_ROOT = Path(__file__).resolve().parent.parent | |
| _DEFAULT_OUTPUT = _PROJECT_ROOT / "data" / "depmap_raw" | |
| DEPMAP_FILES_API = "https://depmap.org/portal/api/download/files" | |
| # Files to download from DepMap portal | |
| DEPMAP_TARGET_FILES = [ | |
| "CRISPRGeneEffect.csv", | |
| "CRISPRGeneDependency.csv", | |
| "Model.csv", | |
| "AchillesCommonEssentialControls.csv", | |
| "AchillesNonessentialControls.csv", | |
| "OmicsExpressionProteinCodingGenesTPMLogp1.csv", | |
| "OmicsCNGene.csv", | |
| "OmicsSomaticMutationsMatrixDamaging.csv", | |
| ] | |
| DEMETER2_URL = "https://ndownloader.figshare.com/files/13515395" # D2_combined_gene_dep_scores.csv | |
| def resolve_depmap_urls(target_files: list[str]) -> dict[str, str]: | |
| """Fetch signed download URLs from the DepMap portal API. | |
| The API returns a CSV with columns: release, release_date, filename, url, md5_hash. | |
| We take the first (most recent) URL for each target file. | |
| """ | |
| print("Fetching download URLs from DepMap portal API...") | |
| resp = requests.get(DEPMAP_FILES_API, timeout=60) | |
| resp.raise_for_status() | |
| targets_remaining = set(target_files) | |
| urls: dict[str, str] = {} | |
| reader = csv.DictReader(io.StringIO(resp.text)) | |
| for row in reader: | |
| fname = row.get("filename", "") | |
| if fname in targets_remaining: | |
| urls[fname] = row["url"] | |
| targets_remaining.discard(fname) | |
| if not targets_remaining: | |
| break | |
| if targets_remaining: | |
| print(f" WARNING: Could not find URLs for: {targets_remaining}") | |
| print(f" Resolved {len(urls)}/{len(target_files)} file URLs\n") | |
| return urls | |
| def download_file(url: str, output_path: Path, force: bool = False) -> bool: | |
| """Download a file with progress reporting.""" | |
| if output_path.exists() and not force: | |
| print(f" SKIP (exists): {output_path.name}") | |
| return False | |
| print(f" Downloading: {output_path.name}...") | |
| resp = requests.get(url, stream=True, timeout=600) | |
| resp.raise_for_status() | |
| total = int(resp.headers.get("content-length", 0)) | |
| downloaded = 0 | |
| with open(output_path, "wb") as f: | |
| for chunk in resp.iter_content(chunk_size=1024 * 1024): | |
| f.write(chunk) | |
| downloaded += len(chunk) | |
| if total: | |
| pct = downloaded / total * 100 | |
| print(f"\r {downloaded / 1e6:.1f} MB / {total / 1e6:.1f} MB ({pct:.0f}%)", end="") | |
| print(f"\n Done: {output_path.name} ({output_path.stat().st_size / 1e6:.1f} MB)") | |
| return True | |
| def main(): | |
| parser = argparse.ArgumentParser(description="Download DepMap data files") | |
| parser.add_argument("--output-dir", type=str, default=str(_DEFAULT_OUTPUT)) | |
| parser.add_argument("--force", action="store_true", help="Re-download existing files") | |
| parser.add_argument("--skip-omics", action="store_true", help="Skip large omics files") | |
| parser.add_argument("--skip-rnai", action="store_true", help="Skip DEMETER2 RNAi file") | |
| args = parser.parse_args() | |
| output_dir = Path(args.output_dir) | |
| output_dir.mkdir(parents=True, exist_ok=True) | |
| print(f"Downloading DepMap files to {output_dir}\n") | |
| # Resolve URLs from DepMap API | |
| targets = [f for f in DEPMAP_TARGET_FILES | |
| if not (args.skip_omics and f.startswith("Omics"))] | |
| url_map = resolve_depmap_urls(targets) | |
| for filename in targets: | |
| if filename not in url_map: | |
| print(f" SKIP (no URL): {filename}") | |
| continue | |
| try: | |
| download_file(url_map[filename], output_dir / filename, force=args.force) | |
| except Exception as e: | |
| print(f" ERROR: {filename}: {e}") | |
| if not args.skip_rnai: | |
| try: | |
| download_file(DEMETER2_URL, output_dir / "D2_combined_gene_dep_scores.csv", force=args.force) | |
| except Exception as e: | |
| print(f" ERROR: DEMETER2: {e}") | |
| print("\nDownload complete.") | |
| if __name__ == "__main__": | |
| main() | |