| { | |
| "Description": "This is a pre-training configuration for codon optimization model based on BART language model.", | |
| "project_name": "COnTRA_SCPECBS3_wintest", | |
| "dataset_path": "/mount/COnTRA/data/datasets/processed_data_SCPECBS3/homologs/SCPECBS3_SCPECBS3_ExprRefined", | |
| "tokenizer_path": "/mount/COnTRA/tokenizers/contra_tokenizer_gen_exprrefined", | |
| "cai_refference_path": "/mount/COnTRA/data/datasets/processed_data_SCPECBS3/S_cerevisiae/S_cerevisiae.0.nt.fasta", | |
| "checkpoint_flag": true, | |
| "checkpoint_path": "/mount/COnTRA/models/Finetuned2Steps_homologs_75_ExR/best_model/", | |
| "special_token_th": 42, | |
| "mask_all": false, | |
| "sw_aa_size": 75, | |
| "eval_type": "model", | |
| "model_type": "COBaBExRi", | |
| "cai_query_species": "S_cerevisiae", | |
| "outdir": "/mount/COnTRA/models/Finetuned2Steps_homologs_75_ExR/best_model/", | |
| "outfile": "model_eval_mimic.csv", | |
| "out_dict": [ | |
| "prot_len", | |
| "num_of_correct_predicted_codons", | |
| "accuracy", | |
| "cross_entropy_loss", | |
| "entropy", | |
| "query_codons", | |
| "subject_codons", | |
| "pred_codons" | |
| ], | |
| "orig_dict": [ | |
| "qseqid", | |
| "sseqid", | |
| "query_species", | |
| "subject_species", | |
| "pident", | |
| "length", | |
| "mismatch", | |
| "gapopen", | |
| "qstart", | |
| "qend", | |
| "sstart", | |
| "send", | |
| "evalue", | |
| "bitscore" | |
| ], | |
| "debug": false | |
| } |