File size: 1,897 Bytes
44a25da
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
#!/bin/bash

# # Define the base directory containing subdirectories
BASE_DIR="./lightning_logs_round_1"

# # Get a sorted list of subdirectories
subdirs=($(ls -d "$BASE_DIR"/*/ | sort))

# Generate sbatch scripts
for i in {0..9}; do
    # Extract the version number (subdirectory name)
    version_number=$(basename "${subdirs[$i]}")

    # Define the sbatch filename
    sbatch_filename="alignment_dataset_2_96_from_ESM3_${i}.sh"

    # Write the sbatch file
    cat <<EOL > "$sbatch_filename"
#!/bin/bash
#SBATCH -N 1
#SBATCH -G 4
#SBATCH -C gpu&hbm80g
#SBATCH -q regular
#SBATCH -t 12:00:00
#SBATCH -A m4235
#SBATCH -J alignment

source /global/u2/s/si264/mambaforge/etc/profile.d/conda.sh
conda activate proera

pera_train \\
    "train.lightning_model_args.eval_type=era" \\
    "train.lightning_model_args.beta=10.0" \\
    "train.lightning_model_args.gamma=0" \\
    "train.trainer_args.devices=4" \\
    "train.trainer_args.max_epochs=25" \\
    "train.trainer_args.log_every_n_steps=1" \\
    "train.trainer_args.enable_progress_bar=True" \\
    "train.lightning_model_args.interval=epoch" \\
    "train.lightning_model_args.monitor=val/ERALoss" \\
    "train.lightning_model_args.lr_scheduler=ReduceLROnPlateau" \\
    "++train.lightning_model_args.lr_scheduler_args.patience=5" \\
    "train.lightning_model_args.optimizer=AdamW" \\
    "train.lightning_model_args.optimizer_args.lr=1.0e-6" \\
    "++train.lightning_model_args.optimizer_args.betas=[0.9,0.99]" \\
    "++train.lightning_model_args.optimizer_args.weight_decay=0.01" \\
    "train.lightning_model_args.on_step=false" \\
    "global_args.dataset_filename=alignment_dataset_2_96_from_ESM3_${i}.hdf5" \\
    "nn.batch_size=4" \\
    "nn.load_model=$BASE_DIR/$version_number/checkpoints/best_model.ckpt" \\
    "++nn.model_args.unified_transformer_args.ida_layer_indices=[]" \\
EOL

    echo "Created $sbatch_filename"
done