| | import argparse
|
| |
|
| | def update_data_args(parser):
|
| | parser.add_argument("-g", "--genome", type=str, required=True, help="Genome fasta")
|
| | parser.add_argument("-b", "--bigwig", type=str, required=False, help="Bigwig of tn5 insertions. Ensure it is +4/-4 shifted")
|
| | parser.add_argument("-p", "--peaks", type=str, default="None", help="10 column bed file of peaks. Sequences and labels will be extracted centered at start (2nd col) + summit (10th col).")
|
| | parser.add_argument("-n", "--nonpeaks", type=str, default="None" ,help="10 column bed file of non-peak regions, centered at summit (10th column)")
|
| | parser.add_argument("-o", "--output_prefix", type=str, required=True, help="Output prefix")
|
| | parser.add_argument("-fl", "--chr_fold_path", type=str, required=True, help="Fold information - see splits.py to set folds")
|
| | parser.add_argument("--trackables",nargs="*",default=['loss','val_loss'], help="list of things to track per batch, such as logcount_predictions_loss,loss,profile_predictions_loss,val_logcount_predictions_loss,val_loss,val_profile_predictions_loss")
|
| |
|
| |
|
| | def update_train_args(parser):
|
| | parser.add_argument("-e", "--epochs", type=int, default=50, help="Maximum epochs to train")
|
| | parser.add_argument("-es", "--early-stop", type=int, default=5, help="Early stop limit, corresponds to 'patience' in callback")
|
| | parser.add_argument("-bs", "--batch_size", type=int, default=64)
|
| | parser.add_argument("-l", "--learning-rate", type=float, default=0.001)
|
| | parser.add_argument("-pf", "--params", type=str, required=True, default=None)
|
| |
|
| | def update_model_args(parser):
|
| | parser.add_argument("-s", "--seed", type=int, default=1234, help="seed to use for model training")
|
| | parser.add_argument("-a","--architecture_from_file",type=str,required=True, default=None, help="Model to use for training")
|
| |
|
| | def fetch_train_args():
|
| | parser = argparse.ArgumentParser()
|
| | update_data_args(parser)
|
| | update_train_args(parser)
|
| | update_model_args(parser)
|
| | args = parser.parse_args()
|
| |
|
| | assert((args.peaks.lower() != "none") or (args.nonpeaks.lower() != "none"))
|
| |
|
| | return args
|
| |
|
| | def fetch_predict_args():
|
| | parser = argparse.ArgumentParser()
|
| | update_data_args(parser)
|
| | parser.add_argument("-m", "--model_h5", type=str, required=True, help="Path to model hdf5")
|
| | parser.add_argument("-bs", "--batch_size", type=int, default=512)
|
| | parser.add_argument("-s", "--seed", type=int, default=1234, help="seed to use for model training")
|
| | parser.add_argument("-il", "--inputlen", type=int, default=2114, help="Sequence input length")
|
| | parser.add_argument("-ol", "--outputlen", type=int, default=1000, help="Prediction output length")
|
| | args = parser.parse_args()
|
| |
|
| | assert((args.peaks.lower() != "none") or (args.nonpeaks.lower() != "none"))
|
| | assert(args.inputlen % 2 ==0)
|
| | assert(args.outputlen % 2 ==0)
|
| |
|
| | return args
|
| |
|
| |
|
| | def fetch_modisco_args():
|
| | parser = argparse.ArgumentParser()
|
| | parser.add_argument("-s", "--scores-prefix", type=str, required=True, help="Prefix to counts/profile h5 files. Will use prefix.{profile,counts}_scores.h5")
|
| | parser.add_argument("-p","--profile-or-counts", type=str, required=True, help="Scoring method to use, profile or counts scores")
|
| | parser.add_argument("-o", "--output-dir", type=str, required=True, help="Output directory")
|
| | parser.add_argument("-c", "--crop", type=int, default=500, help="Crop scores to this width from the center for each example")
|
| | parser.add_argument("-m", "--max-seqlets", type=int, default=50000, help="Max number of seqlets per metacluster for modisco")
|
| |
|
| | args = parser.parse_args()
|
| | return args
|
| |
|