|
|
import pandas as pd |
|
|
import numpy as np |
|
|
import networkx as nx |
|
|
from typing import List, Dict, Tuple, Set, Optional |
|
|
import json |
|
|
import random |
|
|
from collections import defaultdict, Counter |
|
|
import multiprocessing as mp |
|
|
from functools import partial |
|
|
import time |
|
|
import requests |
|
|
from urllib.parse import quote |
|
|
import xml.etree.ElementTree as ET |
|
|
from time import sleep |
|
|
import re |
|
|
import os |
|
|
import fcntl |
|
|
import tempfile |
|
|
import shutil |
|
|
|
|
|
|
|
|
MAX_PROTEINS_TO_PROCESS = 1000 |
|
|
TRUNCATE_MIDDLE_ABOVE_THIS = 120 |
|
|
|
|
|
class ProteinDataEnricher: |
|
|
""" |
|
|
Handles fetching and caching protein sequence and structural data with thread-safe persistent cache |
|
|
""" |
|
|
def __init__(self, cache_dir: str = "protein_cache"): |
|
|
self.cache_dir = cache_dir |
|
|
self.uniprot_cache = {} |
|
|
self.alphafold_cache = {} |
|
|
self.session = requests.Session() |
|
|
self.session.headers.update({'User-Agent': 'ProteinNetworkDataset/1.0'}) |
|
|
|
|
|
|
|
|
os.makedirs(cache_dir, exist_ok=True) |
|
|
|
|
|
|
|
|
self._load_caches() |
|
|
|
|
|
def _get_cache_files(self): |
|
|
"""Get cache file paths""" |
|
|
uniprot_cache_file = os.path.join(self.cache_dir, "uniprot_cache.json") |
|
|
alphafold_cache_file = os.path.join(self.cache_dir, "alphafold_cache.json") |
|
|
return uniprot_cache_file, alphafold_cache_file |
|
|
|
|
|
def _load_caches(self): |
|
|
"""Load existing caches from disk""" |
|
|
uniprot_cache_file, alphafold_cache_file = self._get_cache_files() |
|
|
|
|
|
|
|
|
if os.path.exists(uniprot_cache_file): |
|
|
try: |
|
|
with open(uniprot_cache_file, 'r') as f: |
|
|
self.uniprot_cache = json.load(f) |
|
|
print(f"Loaded UniProt cache with {len(self.uniprot_cache)} entries") |
|
|
except Exception as e: |
|
|
print(f"Error loading UniProt cache: {e}") |
|
|
self.uniprot_cache = {} |
|
|
|
|
|
|
|
|
if os.path.exists(alphafold_cache_file): |
|
|
try: |
|
|
with open(alphafold_cache_file, 'r') as f: |
|
|
self.alphafold_cache = json.load(f) |
|
|
print(f"Loaded AlphaFold cache with {len(self.alphafold_cache)} entries") |
|
|
except Exception as e: |
|
|
print(f"Error loading AlphaFold cache: {e}") |
|
|
self.alphafold_cache = {} |
|
|
|
|
|
def _safe_merge_and_save_cache(self, cache_type: str, new_data: Dict): |
|
|
""" |
|
|
Safely merge new cache data with existing cache using file locking |
|
|
""" |
|
|
if cache_type == "uniprot": |
|
|
cache_file = os.path.join(self.cache_dir, "uniprot_cache.json") |
|
|
elif cache_type == "alphafold": |
|
|
cache_file = os.path.join(self.cache_dir, "alphafold_cache.json") |
|
|
else: |
|
|
raise ValueError(f"Unknown cache type: {cache_type}") |
|
|
|
|
|
|
|
|
lock_file = cache_file + ".lock" |
|
|
|
|
|
try: |
|
|
|
|
|
with open(lock_file, 'w') as lock: |
|
|
fcntl.flock(lock.fileno(), fcntl.LOCK_EX) |
|
|
|
|
|
|
|
|
current_cache = {} |
|
|
if os.path.exists(cache_file): |
|
|
try: |
|
|
with open(cache_file, 'r') as f: |
|
|
current_cache = json.load(f) |
|
|
except Exception as e: |
|
|
print(f"Error loading {cache_type} cache for merge: {e}") |
|
|
current_cache = {} |
|
|
|
|
|
|
|
|
current_cache.update(new_data) |
|
|
|
|
|
|
|
|
temp_file = cache_file + ".tmp" |
|
|
with open(temp_file, 'w') as f: |
|
|
json.dump(current_cache, f, indent=2) |
|
|
|
|
|
|
|
|
shutil.move(temp_file, cache_file) |
|
|
|
|
|
print(f"Merged and saved {cache_type} cache: {len(current_cache)} total entries") |
|
|
|
|
|
except Exception as e: |
|
|
print(f"Error saving {cache_type} cache: {e}") |
|
|
finally: |
|
|
|
|
|
try: |
|
|
os.remove(lock_file) |
|
|
except: |
|
|
pass |
|
|
|
|
|
def get_uniprot_info(self, protein_name: str) -> Optional[Dict]: |
|
|
""" |
|
|
Fetch UniProt information for a protein |
|
|
""" |
|
|
if protein_name in self.uniprot_cache: |
|
|
return self.uniprot_cache[protein_name] |
|
|
|
|
|
try: |
|
|
|
|
|
search_url = f"https://rest.uniprot.org/uniprotkb/search" |
|
|
params = { |
|
|
'query': f'gene_exact:{protein_name} OR protein_name:{protein_name}', |
|
|
'format': 'json', |
|
|
'size': 1 |
|
|
} |
|
|
|
|
|
response = self.session.get(search_url, params=params, timeout=10) |
|
|
if response.status_code == 200: |
|
|
data = response.json() |
|
|
if data.get('results'): |
|
|
entry = data['results'][0] |
|
|
uniprot_info = { |
|
|
'uniprot_id': entry.get('primaryAccession'), |
|
|
'protein_name': entry.get('proteinDescription', {}).get('recommendedName', {}).get('fullName', {}).get('value'), |
|
|
'gene_names': [gn.get('geneName', {}).get('value') for gn in entry.get('genes', []) if gn.get('geneName')], |
|
|
'organism': entry.get('organism', {}).get('scientificName'), |
|
|
'sequence': entry.get('sequence', {}).get('value'), |
|
|
'sequence_length': entry.get('sequence', {}).get('length'), |
|
|
'function': entry.get('comments', [{}])[0].get('texts', [{}])[0].get('value') if entry.get('comments') else None |
|
|
} |
|
|
self.uniprot_cache[protein_name] = uniprot_info |
|
|
return uniprot_info |
|
|
|
|
|
except Exception as e: |
|
|
print(f"Error fetching UniProt data for {protein_name}: {e}") |
|
|
|
|
|
self.uniprot_cache[protein_name] = None |
|
|
return None |
|
|
|
|
|
def get_alphafold_info(self, uniprot_id: str) -> Optional[Dict]: |
|
|
""" |
|
|
Fetch AlphaFold structural information for a UniProt ID |
|
|
""" |
|
|
if not uniprot_id: |
|
|
return None |
|
|
|
|
|
cache_key = uniprot_id |
|
|
if cache_key in self.alphafold_cache: |
|
|
return self.alphafold_cache[cache_key] |
|
|
|
|
|
try: |
|
|
|
|
|
af_url = f"https://alphafold.ebi.ac.uk/api/prediction/{uniprot_id}" |
|
|
response = self.session.get(af_url, timeout=10) |
|
|
|
|
|
if response.status_code == 200: |
|
|
af_data = response.json() |
|
|
if af_data: |
|
|
entry = af_data[0] if isinstance(af_data, list) else af_data |
|
|
|
|
|
alphafold_info = { |
|
|
'alphafold_id': entry.get('entryId'), |
|
|
'model_confidence': entry.get('modelConfidence'), |
|
|
'model_url': entry.get('pdbUrl'), |
|
|
'confidence_score': entry.get('modelConfidence'), |
|
|
'structure_coverage': f"{entry.get('uniprotStart', 1)}-{entry.get('uniprotEnd', 'end')}", |
|
|
'has_structure': True |
|
|
} |
|
|
self.alphafold_cache[cache_key] = alphafold_info |
|
|
return alphafold_info |
|
|
|
|
|
except Exception as e: |
|
|
print(f"Error fetching AlphaFold data for {uniprot_id}: {e}") |
|
|
|
|
|
self.alphafold_cache[cache_key] = {'has_structure': False} |
|
|
return {'has_structure': False} |
|
|
|
|
|
def get_protein_enriched_data(self, protein_name: str) -> Dict: |
|
|
""" |
|
|
Get combined UniProt and AlphaFold data for a protein |
|
|
""" |
|
|
enriched_data = {'protein_name': protein_name} |
|
|
|
|
|
|
|
|
uniprot_info = self.get_uniprot_info(protein_name) |
|
|
if uniprot_info: |
|
|
enriched_data.update(uniprot_info) |
|
|
|
|
|
|
|
|
if uniprot_info.get('uniprot_id'): |
|
|
alphafold_info = self.get_alphafold_info(uniprot_info['uniprot_id']) |
|
|
if alphafold_info: |
|
|
enriched_data.update(alphafold_info) |
|
|
|
|
|
|
|
|
sleep(0.1) |
|
|
return enriched_data |
|
|
|
|
|
def save_cache(self): |
|
|
"""Save caches using safe merge strategy""" |
|
|
|
|
|
uniprot_to_save = {k: v for k, v in self.uniprot_cache.items() if v is not None} |
|
|
alphafold_to_save = {k: v for k, v in self.alphafold_cache.items() if v is not None} |
|
|
|
|
|
if uniprot_to_save: |
|
|
self._safe_merge_and_save_cache("uniprot", uniprot_to_save) |
|
|
|
|
|
if alphafold_to_save: |
|
|
self._safe_merge_and_save_cache("alphafold", alphafold_to_save) |
|
|
|
|
|
def enrich_proteins_worker(args): |
|
|
""" |
|
|
Worker function to enrich a batch of proteins with sequence/structure data |
|
|
""" |
|
|
protein_names_batch, cache_dir, worker_id = args |
|
|
enricher = ProteinDataEnricher(cache_dir) |
|
|
enriched_proteins = {} |
|
|
|
|
|
print(f"Worker {worker_id}: Processing {len(protein_names_batch)} proteins") |
|
|
|
|
|
for i, protein_name in enumerate(protein_names_batch): |
|
|
enriched_proteins[protein_name] = enricher.get_protein_enriched_data(protein_name) |
|
|
|
|
|
|
|
|
if (i + 1) % 25 == 0: |
|
|
enricher.save_cache() |
|
|
print(f"Worker {worker_id}: Saved cache after {i + 1} proteins") |
|
|
|
|
|
|
|
|
enricher.save_cache() |
|
|
print(f"Worker {worker_id}: Completed batch, final cache save") |
|
|
|
|
|
return enriched_proteins |
|
|
|
|
|
def extract_neighborhood_worker(args): |
|
|
""" |
|
|
Worker function for parallel neighborhood extraction |
|
|
""" |
|
|
center_protein, interactions_by_protein, all_interactions, max_size = args |
|
|
|
|
|
|
|
|
neighbors = set() |
|
|
relevant_interactions = [] |
|
|
|
|
|
for interaction in interactions_by_protein[center_protein]: |
|
|
other_protein = (interaction['protein_b'] if interaction['protein_a'] == center_protein |
|
|
else interaction['protein_a']) |
|
|
neighbors.add(other_protein) |
|
|
relevant_interactions.append(interaction) |
|
|
|
|
|
|
|
|
if len(neighbors) > max_size - 1: |
|
|
neighbors = set(random.sample(list(neighbors), max_size - 1)) |
|
|
|
|
|
|
|
|
neighborhood_proteins = {center_protein} | neighbors |
|
|
neighborhood_interactions = [] |
|
|
|
|
|
for interaction in all_interactions: |
|
|
if (interaction['protein_a'] in neighborhood_proteins and |
|
|
interaction['protein_b'] in neighborhood_proteins): |
|
|
neighborhood_interactions.append(interaction) |
|
|
|
|
|
return { |
|
|
'center_protein': center_protein, |
|
|
'proteins': sorted(list(neighborhood_proteins)), |
|
|
'interactions': neighborhood_interactions |
|
|
} |
|
|
|
|
|
def create_conversation_examples_worker(args): |
|
|
""" |
|
|
Worker function for parallel conversation creation |
|
|
""" |
|
|
neighborhood, enriched_proteins = args |
|
|
creator = ProteinNetworkConversationDataset("") |
|
|
creator.enriched_proteins = enriched_proteins |
|
|
conversations = [] |
|
|
|
|
|
|
|
|
conversations.extend(creator.create_protein_list_to_network_examples(neighborhood)) |
|
|
|
|
|
|
|
|
conversations.extend(creator.create_new_protein_prediction_examples(neighborhood)) |
|
|
|
|
|
|
|
|
conversations.extend(creator.create_partial_network_completion_examples(neighborhood)) |
|
|
|
|
|
|
|
|
conversations.extend(creator.create_network_property_examples(neighborhood)) |
|
|
|
|
|
return conversations |
|
|
|
|
|
class ProteinNetworkConversationDataset: |
|
|
def __init__(self, filename: str, cache_dir: str = "protein_cache"): |
|
|
""" |
|
|
Create conversational dataset for protein network prediction using diffusion models |
|
|
""" |
|
|
self.filename = filename |
|
|
self.cache_dir = cache_dir |
|
|
self.df = None |
|
|
self.graph = nx.Graph() |
|
|
self.protein_to_id = {} |
|
|
self.id_to_protein = {} |
|
|
self.interactions_by_protein = defaultdict(list) |
|
|
self.enriched_proteins = {} |
|
|
self.enricher = ProteinDataEnricher(cache_dir) |
|
|
|
|
|
def load_and_parse_biogrid(self): |
|
|
"""Load and parse BioGRID data""" |
|
|
print("Loading BioGRID data...") |
|
|
self.df = pd.read_csv( |
|
|
self.filename, |
|
|
sep='\t', |
|
|
comment='#', |
|
|
low_memory=False, |
|
|
dtype=str |
|
|
) |
|
|
|
|
|
|
|
|
protein_a_col = 7 |
|
|
protein_b_col = 8 |
|
|
interaction_type_col = 11 |
|
|
|
|
|
interactions = [] |
|
|
protein_set = set() |
|
|
|
|
|
seen_interactions = set() |
|
|
|
|
|
for idx, row in self.df.iterrows(): |
|
|
try: |
|
|
protein_a = str(row.iloc[protein_a_col]).strip() |
|
|
protein_b = str(row.iloc[protein_b_col]).strip() |
|
|
interaction_type = str(row.iloc[interaction_type_col]).strip() |
|
|
|
|
|
if protein_a in ['-', 'nan', ''] or protein_b in ['-', 'nan', '']: |
|
|
continue |
|
|
|
|
|
|
|
|
if protein_a == protein_b: |
|
|
continue |
|
|
|
|
|
|
|
|
interaction_key = tuple(sorted([protein_a, protein_b]) + [interaction_type]) |
|
|
|
|
|
|
|
|
if interaction_key in seen_interactions: |
|
|
continue |
|
|
|
|
|
seen_interactions.add(interaction_key) |
|
|
protein_set.add(protein_a) |
|
|
protein_set.add(protein_b) |
|
|
|
|
|
interaction = { |
|
|
'protein_a': protein_a, |
|
|
'protein_b': protein_b, |
|
|
'interaction_type': interaction_type |
|
|
} |
|
|
interactions.append(interaction) |
|
|
|
|
|
|
|
|
self.interactions_by_protein[protein_a].append(interaction) |
|
|
self.interactions_by_protein[protein_b].append(interaction) |
|
|
|
|
|
except Exception: |
|
|
continue |
|
|
|
|
|
print(f"Extracted {len(interactions)} valid unique interactions") |
|
|
print(f"Found {len(protein_set)} unique proteins") |
|
|
|
|
|
return interactions, sorted(list(protein_set)) |
|
|
|
|
|
def build_network_neighborhoods(self, interactions, proteins, min_connections=3, max_connections=15): |
|
|
""" |
|
|
Build subnetworks around high-degree proteins for training examples |
|
|
""" |
|
|
|
|
|
protein_degrees = Counter() |
|
|
for interaction in interactions: |
|
|
protein_degrees[interaction['protein_a']] += 1 |
|
|
protein_degrees[interaction['protein_b']] += 1 |
|
|
|
|
|
|
|
|
candidate_proteins = [ |
|
|
protein for protein, degree in protein_degrees.items() |
|
|
if min_connections <= degree <= max_connections |
|
|
] |
|
|
|
|
|
print(f"Found {len(candidate_proteins)} proteins with degree {min_connections}-{max_connections}") |
|
|
|
|
|
|
|
|
limited_proteins = candidate_proteins[:MAX_PROTEINS_TO_PROCESS] |
|
|
print(f"Processing {len(limited_proteins)} proteins with multiprocessing...") |
|
|
|
|
|
|
|
|
worker_args = [ |
|
|
(protein, self.interactions_by_protein, interactions, 10) |
|
|
for protein in limited_proteins |
|
|
] |
|
|
|
|
|
|
|
|
num_processes = min(int(mp.cpu_count()/2), len(limited_proteins)) |
|
|
print(f"Using {num_processes} processes") |
|
|
|
|
|
neighborhoods = [] |
|
|
with mp.Pool(processes=num_processes) as pool: |
|
|
results = pool.map(extract_neighborhood_worker, worker_args) |
|
|
|
|
|
|
|
|
neighborhoods = [ |
|
|
neighborhood for neighborhood in results |
|
|
if len(neighborhood['proteins']) >= 3 |
|
|
] |
|
|
|
|
|
return neighborhoods |
|
|
|
|
|
def enrich_proteins_with_data(self, proteins: List[str]): |
|
|
""" |
|
|
Enrich proteins with UniProt and AlphaFold data using multiprocessing with thread-safe persistent cache |
|
|
""" |
|
|
print(f"Enriching {len(proteins)} proteins with sequence and structural data...") |
|
|
|
|
|
|
|
|
cached_count = 0 |
|
|
proteins_to_fetch = [] |
|
|
|
|
|
for protein in proteins: |
|
|
|
|
|
if (protein in self.enricher.uniprot_cache): |
|
|
cached_count += 1 |
|
|
|
|
|
self.enriched_proteins[protein] = self.enricher.get_protein_enriched_data(protein) |
|
|
else: |
|
|
proteins_to_fetch.append(protein) |
|
|
|
|
|
print(f"Found {cached_count} proteins in cache, need to fetch {len(proteins_to_fetch)} new proteins") |
|
|
|
|
|
if not proteins_to_fetch: |
|
|
print("All proteins found in cache!") |
|
|
return self.enriched_proteins |
|
|
|
|
|
|
|
|
batch_size = 25 |
|
|
protein_batches = [proteins_to_fetch[i:i+batch_size] for i in range(0, len(proteins_to_fetch), batch_size)] |
|
|
|
|
|
|
|
|
worker_args = [(batch, self.cache_dir, i) for i, batch in enumerate(protein_batches)] |
|
|
|
|
|
|
|
|
num_processes = min(int(mp.cpu_count()/2), len(protein_batches)) |
|
|
print(f"Using {num_processes} processes for protein enrichment with thread-safe caching") |
|
|
|
|
|
enrichment_start = time.time() |
|
|
with mp.Pool(processes=num_processes) as pool: |
|
|
results = pool.map(enrich_proteins_worker, worker_args) |
|
|
|
|
|
|
|
|
for batch_result in results: |
|
|
self.enriched_proteins.update(batch_result) |
|
|
|
|
|
|
|
|
print("Reloading cache to get all merged data...") |
|
|
self.enricher._load_caches() |
|
|
|
|
|
|
|
|
for protein in proteins: |
|
|
if protein not in self.enriched_proteins and protein in self.enricher.uniprot_cache: |
|
|
self.enriched_proteins[protein] = self.enricher.get_protein_enriched_data(protein) |
|
|
|
|
|
enrichment_time = time.time() - enrichment_start |
|
|
successful_enrichments = sum(1 for data in self.enriched_proteins.values() |
|
|
if data.get('uniprot_id') is not None) |
|
|
|
|
|
print(f"Protein enrichment completed in {enrichment_time:.2f} seconds") |
|
|
print(f"Successfully enriched {successful_enrichments}/{len(proteins)} proteins") |
|
|
print(f"Final cache sizes - UniProt: {len(self.enricher.uniprot_cache)}, AlphaFold: {len(self.enricher.alphafold_cache)}") |
|
|
|
|
|
return self.enriched_proteins |
|
|
|
|
|
def extract_neighborhood(self, center_protein, interactions, max_size=10): |
|
|
""" |
|
|
Extract neighborhood around a protein |
|
|
""" |
|
|
|
|
|
neighbors = set() |
|
|
relevant_interactions = [] |
|
|
|
|
|
for interaction in self.interactions_by_protein[center_protein]: |
|
|
other_protein = (interaction['protein_b'] if interaction['protein_a'] == center_protein |
|
|
else interaction['protein_a']) |
|
|
neighbors.add(other_protein) |
|
|
relevant_interactions.append(interaction) |
|
|
|
|
|
|
|
|
if len(neighbors) > max_size - 1: |
|
|
neighbors = set(random.sample(list(neighbors), max_size - 1)) |
|
|
|
|
|
|
|
|
neighborhood_proteins = {center_protein} | neighbors |
|
|
neighborhood_interactions = [] |
|
|
|
|
|
for interaction in interactions: |
|
|
if (interaction['protein_a'] in neighborhood_proteins and |
|
|
interaction['protein_b'] in neighborhood_proteins): |
|
|
neighborhood_interactions.append(interaction) |
|
|
|
|
|
return { |
|
|
'center_protein': center_protein, |
|
|
'proteins': sorted(list(neighborhood_proteins)), |
|
|
'interactions': neighborhood_interactions |
|
|
} |
|
|
|
|
|
def create_conversation_examples(self, neighborhoods): |
|
|
""" |
|
|
Create different types of conversation examples for diffusion training (parallelized) |
|
|
""" |
|
|
print(f"Creating conversation examples for {len(neighborhoods)} neighborhoods using multiprocessing...") |
|
|
|
|
|
|
|
|
worker_args = [(neighborhood, self.enriched_proteins) for neighborhood in neighborhoods] |
|
|
|
|
|
|
|
|
num_processes = min(int(mp.cpu_count()/2), len(neighborhoods)) |
|
|
print(f"Using {num_processes} processes for conversation creation") |
|
|
|
|
|
conversations = [] |
|
|
with mp.Pool(processes=num_processes) as pool: |
|
|
results = pool.map(create_conversation_examples_worker, worker_args) |
|
|
|
|
|
|
|
|
for result in results: |
|
|
conversations.extend(result) |
|
|
|
|
|
return conversations |
|
|
|
|
|
def create_protein_list_to_network_examples(self, neighborhood): |
|
|
""" |
|
|
Context: List of proteins with sequence/structure info |
|
|
Generation: Complete interaction network |
|
|
""" |
|
|
examples = [] |
|
|
proteins = neighborhood['proteins'] |
|
|
interactions = neighborhood['interactions'] |
|
|
|
|
|
|
|
|
network_text = self.format_enriched_network_as_text(proteins, interactions) |
|
|
|
|
|
system_msg = { |
|
|
"role": "system", |
|
|
"content": "You are a protein interaction prediction system. Given a list of proteins with their sequence and structural information, predict all likely interactions between them based on biological knowledge, sequence similarity, and structural compatibility." |
|
|
} |
|
|
|
|
|
|
|
|
protein_context = self.format_proteins_with_context(proteins) |
|
|
|
|
|
user_msg = { |
|
|
"role": "user", |
|
|
"content": f"Predict the protein interaction network for these proteins:\n\n{protein_context}" |
|
|
} |
|
|
|
|
|
assistant_msg = { |
|
|
"role": "assistant", |
|
|
"content": network_text |
|
|
} |
|
|
|
|
|
conversation = [system_msg, user_msg, assistant_msg] |
|
|
examples.append({"updated": conversation}) |
|
|
|
|
|
return examples |
|
|
|
|
|
def create_new_protein_prediction_examples(self, neighborhood): |
|
|
""" |
|
|
Context: Known network + new protein with sequence/structure info |
|
|
Generation: Interactions for the new protein |
|
|
""" |
|
|
examples = [] |
|
|
if len(neighborhood['proteins']) < 4: |
|
|
return examples |
|
|
|
|
|
proteins = neighborhood['proteins'] |
|
|
interactions = neighborhood['interactions'] |
|
|
|
|
|
|
|
|
target_protein = random.choice(proteins) |
|
|
remaining_proteins = [p for p in proteins if p != target_protein] |
|
|
|
|
|
|
|
|
known_interactions = [ |
|
|
i for i in interactions |
|
|
if target_protein not in [i['protein_a'], i['protein_b']] |
|
|
] |
|
|
|
|
|
|
|
|
target_interactions = [ |
|
|
i for i in interactions |
|
|
if target_protein in [i['protein_a'], i['protein_b']] |
|
|
] |
|
|
|
|
|
if not target_interactions: |
|
|
return examples |
|
|
|
|
|
known_network_text = self.format_enriched_network_as_text(remaining_proteins, known_interactions) |
|
|
target_network_text = self.format_interactions_as_text(target_interactions) |
|
|
|
|
|
|
|
|
target_enriched_data = self.enriched_proteins.get(target_protein, {}) |
|
|
target_context = self.format_proteins_with_context([target_protein]) |
|
|
|
|
|
system_msg = { |
|
|
"role": "system", |
|
|
"content": "You are a protein interaction prediction system. Given a known protein network and a new protein with sequence and structural information, predict which proteins in the network the new protein will interact with based on sequence similarity, structural compatibility, and functional relationships." |
|
|
} |
|
|
|
|
|
user_msg = { |
|
|
"role": "user", |
|
|
"content": f"Known protein network:\n{known_network_text}\n\nNew protein to integrate:\n{target_context}\n\nPredict the interactions for {target_protein} based on its sequence, structure, and function:" |
|
|
} |
|
|
|
|
|
assistant_msg = { |
|
|
"role": "assistant", |
|
|
"content": target_network_text |
|
|
} |
|
|
|
|
|
conversation = [system_msg, user_msg, assistant_msg] |
|
|
examples.append({"updated": conversation}) |
|
|
|
|
|
return examples |
|
|
|
|
|
def create_partial_network_completion_examples(self, neighborhood): |
|
|
""" |
|
|
Context: Partial network with some missing interactions |
|
|
Generation: Complete network |
|
|
""" |
|
|
examples = [] |
|
|
proteins = neighborhood['proteins'] |
|
|
interactions = neighborhood['interactions'] |
|
|
|
|
|
if len(interactions) < 3: |
|
|
return examples |
|
|
|
|
|
|
|
|
n_hidden = max(1, len(interactions) // 3) |
|
|
hidden_interactions = random.sample(interactions, n_hidden) |
|
|
visible_interactions = [i for i in interactions if i not in hidden_interactions] |
|
|
|
|
|
partial_network_text = self.format_network_as_text(proteins, visible_interactions) |
|
|
complete_network_text = self.format_network_as_text(proteins, interactions) |
|
|
|
|
|
system_msg = { |
|
|
"role": "system", |
|
|
"content": "You are a protein interaction prediction system. Given a partial protein network, predict the complete network including missing interactions." |
|
|
} |
|
|
|
|
|
user_msg = { |
|
|
"role": "user", |
|
|
"content": f"Complete this partial protein network:\n{partial_network_text}" |
|
|
} |
|
|
|
|
|
assistant_msg = { |
|
|
"role": "assistant", |
|
|
"content": complete_network_text |
|
|
} |
|
|
|
|
|
conversation = [system_msg, user_msg, assistant_msg] |
|
|
examples.append({"updated": conversation}) |
|
|
|
|
|
return examples |
|
|
|
|
|
def create_network_property_examples(self, neighborhood): |
|
|
""" |
|
|
Context: Network properties and constraints |
|
|
Generation: Network that satisfies those properties |
|
|
""" |
|
|
examples = [] |
|
|
proteins = neighborhood['proteins'] |
|
|
interactions = neighborhood['interactions'] |
|
|
|
|
|
|
|
|
n_proteins = len(proteins) |
|
|
n_interactions = len(interactions) |
|
|
density = (2 * n_interactions) / (n_proteins * (n_proteins - 1)) if n_proteins > 1 else 0 |
|
|
|
|
|
|
|
|
protein_degrees = Counter() |
|
|
for interaction in interactions: |
|
|
protein_degrees[interaction['protein_a']] += 1 |
|
|
protein_degrees[interaction['protein_b']] += 1 |
|
|
|
|
|
hub_proteins = [p for p, degree in protein_degrees.most_common(2)] |
|
|
|
|
|
network_text = self.format_network_as_text(proteins, interactions) |
|
|
|
|
|
system_msg = { |
|
|
"role": "system", |
|
|
"content": "You are a protein interaction network generator. Given network constraints and properties, generate a biologically plausible protein network." |
|
|
} |
|
|
|
|
|
properties_text = (f"Generate a protein network with the following properties:\n" |
|
|
f"- Proteins: {', '.join(proteins)}\n" |
|
|
f"- Network density: approximately {density:.2f}\n" |
|
|
f"- Hub proteins (highly connected): {', '.join(hub_proteins)}\n" |
|
|
f"- Total interactions: approximately {n_interactions}") |
|
|
|
|
|
user_msg = { |
|
|
"role": "user", |
|
|
"content": properties_text |
|
|
} |
|
|
|
|
|
assistant_msg = { |
|
|
"role": "assistant", |
|
|
"content": network_text |
|
|
} |
|
|
|
|
|
conversation = [system_msg, user_msg, assistant_msg] |
|
|
examples.append({"updated": conversation}) |
|
|
|
|
|
return examples |
|
|
|
|
|
def format_network_as_text(self, proteins, interactions): |
|
|
""" |
|
|
Format network as structured text for the model to predict |
|
|
""" |
|
|
|
|
|
proteins = sorted(proteins) |
|
|
|
|
|
|
|
|
interactions_by_type = defaultdict(set) |
|
|
for interaction in interactions: |
|
|
|
|
|
if interaction['protein_a'] == interaction['protein_b']: |
|
|
continue |
|
|
|
|
|
int_type = interaction.get('interaction_type', 'physical') |
|
|
|
|
|
p1, p2 = sorted([interaction['protein_a'], interaction['protein_b']]) |
|
|
interactions_by_type[int_type].add(f"{p1}--{p2}") |
|
|
|
|
|
result = f"PROTEINS: {', '.join(proteins)}\n\n" |
|
|
|
|
|
for int_type, edges in interactions_by_type.items(): |
|
|
if edges: |
|
|
result += f"{int_type.upper()} INTERACTIONS:\n" |
|
|
for edge in sorted(edges): |
|
|
result += f" {edge}\n" |
|
|
result += "\n" |
|
|
|
|
|
total_interactions = sum(len(edges) for edges in interactions_by_type.values()) |
|
|
result += f"NETWORK SUMMARY: {len(proteins)} proteins, {total_interactions} unique interactions" |
|
|
return result.strip() |
|
|
|
|
|
def format_proteins_with_context(self, proteins: List[str]) -> str: |
|
|
""" |
|
|
Format proteins with their enriched sequence and structural context |
|
|
""" |
|
|
protein_contexts = [] |
|
|
|
|
|
for protein in sorted(proteins): |
|
|
enriched_data = self.enriched_proteins.get(protein, {}) |
|
|
|
|
|
context_parts = [f"PROTEIN: {protein}"] |
|
|
|
|
|
|
|
|
if enriched_data.get('uniprot_id'): |
|
|
context_parts.append(f" UniProt ID: {enriched_data['uniprot_id']}") |
|
|
|
|
|
if enriched_data.get('protein_name'): |
|
|
context_parts.append(f" Full Name: {enriched_data['protein_name']}") |
|
|
|
|
|
if enriched_data.get('organism'): |
|
|
context_parts.append(f" Organism: {enriched_data['organism']}") |
|
|
|
|
|
|
|
|
if enriched_data.get('sequence_length'): |
|
|
context_parts.append(f" Sequence Length: {enriched_data['sequence_length']} amino acids") |
|
|
|
|
|
if enriched_data.get('sequence'): |
|
|
|
|
|
seq = enriched_data['sequence'] |
|
|
if len(seq) > TRUNCATE_MIDDLE_ABOVE_THIS: |
|
|
seq_preview = f"{seq[:int(TRUNCATE_MIDDLE_ABOVE_THIS*0.5)]}...{seq[-int(TRUNCATE_MIDDLE_ABOVE_THIS*0.2):]}" |
|
|
else: |
|
|
seq_preview = seq |
|
|
context_parts.append(f" Sequence: {seq_preview}") |
|
|
|
|
|
|
|
|
if enriched_data.get('has_structure'): |
|
|
context_parts.append(f" AlphaFold Structure: Available") |
|
|
if enriched_data.get('model_confidence'): |
|
|
context_parts.append(f" Structure Confidence: {enriched_data['model_confidence']}") |
|
|
if enriched_data.get('structure_coverage'): |
|
|
context_parts.append(f" Structure Coverage: residues {enriched_data['structure_coverage']}") |
|
|
else: |
|
|
context_parts.append(f" AlphaFold Structure: Not available") |
|
|
|
|
|
|
|
|
if enriched_data.get('function'): |
|
|
func_text = enriched_data['function'][:200] + "..." if len(enriched_data['function']) > 200 else enriched_data['function'] |
|
|
context_parts.append(f" Function: {func_text}") |
|
|
|
|
|
protein_contexts.append("\n".join(context_parts)) |
|
|
|
|
|
return "\n\n".join(protein_contexts) |
|
|
|
|
|
def format_enriched_network_as_text(self, proteins, interactions): |
|
|
""" |
|
|
Format network with enriched protein information |
|
|
""" |
|
|
|
|
|
basic_network = self.format_network_as_text(proteins, interactions) |
|
|
|
|
|
|
|
|
enriched_summary = "\n\nPROTEIN DETAILS:\n" |
|
|
|
|
|
for protein in sorted(proteins): |
|
|
enriched_data = self.enriched_proteins.get(protein, {}) |
|
|
details = [f"{protein}"] |
|
|
|
|
|
if enriched_data.get('sequence_length'): |
|
|
details.append(f"{enriched_data['sequence_length']}aa") |
|
|
|
|
|
if enriched_data.get('has_structure'): |
|
|
confidence = enriched_data.get('model_confidence', 'unknown') |
|
|
details.append(f"AlphaFold({confidence})") |
|
|
|
|
|
if enriched_data.get('organism'): |
|
|
org = enriched_data['organism'].split()[0] if ' ' in enriched_data['organism'] else enriched_data['organism'] |
|
|
details.append(f"{org}") |
|
|
|
|
|
enriched_summary += f" {' | '.join(details)}\n" |
|
|
|
|
|
return basic_network + enriched_summary |
|
|
|
|
|
def format_interactions_as_text(self, interactions): |
|
|
""" |
|
|
Format just interactions as text |
|
|
""" |
|
|
if not interactions: |
|
|
return "No interactions predicted." |
|
|
|
|
|
|
|
|
interactions_by_type = defaultdict(set) |
|
|
for interaction in interactions: |
|
|
|
|
|
if interaction['protein_a'] == interaction['protein_b']: |
|
|
continue |
|
|
|
|
|
int_type = interaction.get('interaction_type', 'physical') |
|
|
p1, p2 = sorted([interaction['protein_a'], interaction['protein_b']]) |
|
|
interactions_by_type[int_type].add(f"{p1}--{p2}") |
|
|
|
|
|
result = "" |
|
|
for int_type, edges in interactions_by_type.items(): |
|
|
if edges: |
|
|
result += f"{int_type.upper()} INTERACTIONS:\n" |
|
|
for edge in sorted(edges): |
|
|
result += f" {edge}\n" |
|
|
result += "\n" |
|
|
|
|
|
return result.strip() |
|
|
|
|
|
def save_conversation_dataset(self, output_file="processed_dataset.json"): |
|
|
""" |
|
|
Create and save the full conversation dataset with enriched protein data |
|
|
""" |
|
|
start_time = time.time() |
|
|
|
|
|
|
|
|
print("Step 1: Loading and parsing data...") |
|
|
load_start = time.time() |
|
|
interactions, proteins = self.load_and_parse_biogrid() |
|
|
load_time = time.time() - load_start |
|
|
print(f"Data loading completed in {load_time:.2f} seconds") |
|
|
|
|
|
|
|
|
print("Step 2: Building protein neighborhoods...") |
|
|
neighborhood_start = time.time() |
|
|
neighborhoods = self.build_network_neighborhoods(interactions, proteins) |
|
|
neighborhood_time = time.time() - neighborhood_start |
|
|
print(f"Built {len(neighborhoods)} protein neighborhoods in {neighborhood_time:.2f} seconds") |
|
|
|
|
|
|
|
|
print("Step 3: Enriching proteins with sequence and structural data...") |
|
|
enrichment_start = time.time() |
|
|
|
|
|
unique_proteins = set() |
|
|
for neighborhood in neighborhoods: |
|
|
unique_proteins.update(neighborhood['proteins']) |
|
|
|
|
|
self.enrich_proteins_with_data(list(unique_proteins)) |
|
|
enrichment_time = time.time() - enrichment_start |
|
|
print(f"Protein enrichment completed in {enrichment_time:.2f} seconds") |
|
|
|
|
|
|
|
|
print("Step 4: Creating conversation examples...") |
|
|
conversation_start = time.time() |
|
|
conversations = self.create_conversation_examples(neighborhoods) |
|
|
conversation_time = time.time() - conversation_start |
|
|
print(f"Created {len(conversations)} conversation examples in {conversation_time:.2f} seconds") |
|
|
|
|
|
|
|
|
print("Step 5: Shuffling and saving dataset...") |
|
|
random.shuffle(conversations) |
|
|
|
|
|
|
|
|
with open(output_file, 'w') as f: |
|
|
json.dump(conversations, f, indent=2) |
|
|
|
|
|
|
|
|
enriched_data_file = output_file.replace('.json', '_protein_data.json') |
|
|
with open(enriched_data_file, 'w') as f: |
|
|
json.dump(self.enriched_proteins, f, indent=2) |
|
|
|
|
|
total_time = time.time() - start_time |
|
|
print(f"Saved dataset to {output_file}") |
|
|
print(f"Saved enriched protein data to {enriched_data_file}") |
|
|
print(f"Total processing time: {total_time:.2f} seconds") |
|
|
|
|
|
|
|
|
print("\n=== Example Conversations ===") |
|
|
for i, conv in enumerate(conversations[:2]): |
|
|
print(f"\n--- Example {i+1} ---") |
|
|
for msg in conv["updated"]: |
|
|
print(f"{msg['role'].upper()}: {msg['content'][:300]}...") |
|
|
|
|
|
return conversations |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
|
|
|
random.seed(42) |
|
|
np.random.seed(42) |
|
|
|
|
|
print(f"Configuration: Processing up to {MAX_PROTEINS_TO_PROCESS} proteins") |
|
|
print(f"Available CPU cores: {int(mp.cpu_count()/2)}") |
|
|
print(f"Cache directory: protein_cache/") |
|
|
|
|
|
creator = ProteinNetworkConversationDataset( |
|
|
"./unzipped/BIOGRID-ALL-4.4.246.tab3/BIOGRID-ALL-4.4.246.tab3.txt", |
|
|
cache_dir="protein_cache" |
|
|
) |
|
|
|
|
|
conversations = creator.save_conversation_dataset("processed_dataset.json") |
|
|
|
|
|
print(f"\n=== Dataset Summary ===") |
|
|
print(f"Total conversations: {len(conversations)}") |
|
|
|
|
|
|
|
|
task_types = Counter() |
|
|
for conv in conversations: |
|
|
system_content = conv["updated"][0]["content"] |
|
|
if "list of proteins" in system_content: |
|
|
task_types["protein_list_to_network"] += 1 |
|
|
elif "new protein" in system_content: |
|
|
task_types["new_protein_integration"] += 1 |
|
|
elif "partial" in system_content: |
|
|
task_types["partial_completion"] += 1 |
|
|
elif "properties" in system_content: |
|
|
task_types["property_based_generation"] += 1 |
|
|
|
|
|
print("\nTask distribution:") |
|
|
for task, count in task_types.items(): |
|
|
print(f" {task}: {count}") |
|
|
|
|
|
|
|
|
print(f"\n=== Protein Enrichment Summary ===") |
|
|
total_proteins = len(creator.enriched_proteins) |
|
|
proteins_with_uniprot = sum(1 for data in creator.enriched_proteins.values() |
|
|
if data.get('uniprot_id') is not None) |
|
|
proteins_with_sequence = sum(1 for data in creator.enriched_proteins.values() |
|
|
if data.get('sequence') is not None) |
|
|
proteins_with_structure = sum(1 for data in creator.enriched_proteins.values() |
|
|
if data.get('has_structure') == True) |
|
|
|
|
|
print(f"Total proteins processed: {total_proteins}") |
|
|
print(f"Proteins with UniProt data: {proteins_with_uniprot} ({proteins_with_uniprot/total_proteins*100:.1f}%)") |
|
|
print(f"Proteins with sequences: {proteins_with_sequence} ({proteins_with_sequence/total_proteins*100:.1f}%)") |
|
|
print(f"Proteins with AlphaFold structures: {proteins_with_structure} ({proteins_with_structure/total_proteins*100:.1f}%)") |
|
|
|
|
|
|
|
|
print(f"\n=== Cache Statistics ===") |
|
|
print(f"UniProt cache entries: {len(creator.enricher.uniprot_cache)}") |
|
|
print(f"AlphaFold cache entries: {len(creator.enricher.alphafold_cache)}") |
|
|
print(f"Cache files location: {creator.cache_dir}/") |