scl / upload.py
hazemessam's picture
Upload upload.py
2e90987 verified
import os
from Bio import SeqIO
from tqdm.auto import tqdm
import pandas as pd
from huggingface_hub import HfApi
aav_files = [
"des_mut",
"low_vs_high",
"mut_des",
"one_vs_many",
"sampled",
"seven_vs_many",
"two_vs_many",
]
meltome_files = [
"human",
"human_cell",
"mixed_split",
]
sav_files = [
"human",
"only_savs",
"mixed",
]
scl_files = [
"balanced",
"human_hard",
"human_soft",
"mixed_hard",
"mixed_soft",
]
gb1_files = [
"low_vs_high",
"one_vs_rest",
"sampled",
"three_vs_rest",
"two_vs_rest",
]
def download_wget(filename, repo):
if os.path.exists(f"{filename}.fasta"):
return
url = f"http://data.bioembeddings.com/public/FLIP/fasta/{repo}/{filename}.fasta"
os.system(f"wget {url}")
def upload_aav():
repo = "aav"
for filename in aav_files:
download_wget(filename, repo)
fasta_file = f"{filename}.fasta"
csv_file = f"{filename}.csv"
db = SeqIO.index(fasta_file, "fasta")
output = []
for _, record in tqdm(db.items()):
description = record.description
description = description.split()
sequence = str(record.seq)
seqid = description[0]
label = float(description[1].split("=")[1])
split = description[2].split("=")[1]
validation = description[3].split("=")[1]
output.append({
"seqid": seqid,
"label": label,
"sequence": sequence,
"split": split.lower(),
"validation": validation.lower(),
})
pd.DataFrame(output).to_csv(csv_file, index=False)
api = HfApi()
api.create_repo(
repo_id=f"hazemessam/{repo}",
repo_type="dataset",
exist_ok=True,
)
api.upload_file(
path_or_fileobj=csv_file,
path_in_repo=csv_file,
repo_id=f"hazemessam/{repo}",
repo_type="dataset",
)
os.system(f"rm -rf {fasta_file}")
os.system(f"rm -rf {csv_file}")
def upload_meltome():
repo = "meltome"
for filename in meltome_files:
download_wget(filename, repo)
fasta_file = f"{filename}.fasta"
csv_file = f"{filename}.csv"
db = SeqIO.index(fasta_file, "fasta")
output = []
for _, record in tqdm(db.items()):
description = record.description
description = description.split()
sequence = str(record.seq)
seqid = description[0]
label = float(description[1].split("=")[1])
split = description[2].split("=")[1]
validation = description[3].split("=")[1]
output.append({
"seqid": seqid,
"label": label,
"sequence": sequence,
"split": split.lower(),
"validation": validation.lower(),
})
pd.DataFrame(output).to_csv(csv_file, index=False)
api = HfApi()
api.create_repo(
repo_id=f"hazemessam/{repo}",
repo_type="dataset",
exist_ok=True,
)
api.upload_file(
path_or_fileobj=csv_file,
path_in_repo=csv_file,
repo_id=f"hazemessam/{repo}",
repo_type="dataset",
)
os.system(f"rm -rf {fasta_file}")
os.system(f"rm -rf {csv_file}")
def upload_sav():
repo = "sav"
for filename in sav_files:
download_wget(filename, repo)
fasta_file = f"{filename}.fasta"
csv_file = f"{filename}.csv"
db = SeqIO.index(fasta_file, "fasta")
output = []
for _, record in tqdm(db.items()):
description = record.description
description = description.split()
sequence = str(record.seq)
seqid = description[0]
# the label is a string in sav and scl datasets
label = description[1].split("=")[1]
split = description[2].split("=")[1]
validation = description[3].split("=")[1]
output.append({
"seqid": seqid,
"label": label,
"sequence": sequence,
"split": split.lower(),
"validation": validation.lower(),
})
pd.DataFrame(output).to_csv(csv_file, index=False)
api = HfApi()
api.create_repo(
repo_id=f"hazemessam/{repo}",
repo_type="dataset",
exist_ok=True,
)
api.upload_file(
path_or_fileobj=csv_file,
path_in_repo=csv_file,
repo_id=f"hazemessam/{repo}",
repo_type="dataset",
)
os.system(f"rm -rf {fasta_file}")
os.system(f"rm -rf {csv_file}")
def upload_scl():
repo = "scl"
for filename in scl_files:
download_wget(filename, repo)
fasta_file = f"{filename}.fasta"
csv_file = f"{filename}.csv"
db = SeqIO.index(fasta_file, "fasta")
output = []
for _, record in tqdm(db.items()):
description = record.description
description = description.split()
sequence = str(record.seq)
seqid = description[0]
# the label is a string in sav and scl datasets
label = description[1].split("=")[1]
split = description[2].split("=")[1]
validation = description[3].split("=")[1]
output.append({
"seqid": seqid,
"label": label,
"sequence": sequence,
"split": split.lower(),
"validation": validation.lower(),
})
pd.DataFrame(output).to_csv(csv_file, index=False)
api = HfApi()
api.create_repo(
repo_id=f"hazemessam/{repo}",
repo_type="dataset",
exist_ok=True,
)
api.upload_file(
path_or_fileobj=csv_file,
path_in_repo=csv_file,
repo_id=f"hazemessam/{repo}",
repo_type="dataset",
)
os.system(f"rm -rf {fasta_file}")
os.system(f"rm -rf {csv_file}")
def upload_gb1():
repo = "gb1"
for filename in gb1_files:
download_wget(filename, repo)
fasta_file = f"{filename}.fasta"
csv_file = f"{filename}.csv"
db = SeqIO.index(fasta_file, "fasta")
output = []
for _, record in tqdm(db.items()):
description = record.description
description = description.split()
sequence = str(record.seq)
seqid = description[0]
label = float(description[1].split("=")[1])
split = description[2].split("=")[1]
validation = description[3].split("=")[1]
output.append({
"seqid": seqid,
"label": label,
"sequence": sequence,
"split": split.lower(),
"validation": validation.lower(),
})
pd.DataFrame(output).to_csv(csv_file, index=False)
api = HfApi()
api.create_repo(
repo_id=f"hazemessam/{repo}",
repo_type="dataset",
exist_ok=True,
)
api.upload_file(
path_or_fileobj=csv_file,
path_in_repo=csv_file,
repo_id=f"hazemessam/{repo}",
repo_type="dataset",
)
os.system(f"rm -rf {fasta_file}")
os.system(f"rm -rf {csv_file}")
if __name__ == "__main__":
upload_gb1()
upload_meltome()
upload_sav()
upload_scl()
upload_aav()