Commit ·
6cb1a61
1
Parent(s): 8c39cb0
restructured job submission scripts
Browse files- root_gnn_dgl/jobs/cpu.sh +1 -1
- root_gnn_dgl/jobs/{inference.py → inference/run_inference.py} +0 -0
- root_gnn_dgl/jobs/interactive.sh +1 -1
- root_gnn_dgl/jobs/prep_data/prep_data.sh +20 -0
- root_gnn_dgl/jobs/{process_data.py → prep_data/run_processing.py} +6 -14
- root_gnn_dgl/jobs/salloc.sh +1 -1
- root_gnn_dgl/jobs/{multinode → training/multinode}/run_multinode_1.sh +0 -0
- root_gnn_dgl/jobs/{multinode → training/multinode}/run_multinode_2.sh +0 -0
- root_gnn_dgl/jobs/{multinode → training/multinode}/run_multinode_3.sh +0 -0
- root_gnn_dgl/jobs/{multinode/submit_multinode.sh → training/multinode/submit.sh} +0 -0
- root_gnn_dgl/jobs/{singlegpu → training/singlegpu}/run_job.sh +0 -0
- root_gnn_dgl/jobs/{singlegpu → training/singlegpu}/run_job_image.sh +0 -0
- root_gnn_dgl/jobs/{singlegpu/submit_job.sh → training/singlegpu/submit.sh} +0 -0
root_gnn_dgl/jobs/cpu.sh
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
salloc --nodes=1 --ntasks=64 --cpus-per-task=1 --qos=interactive --time=04:00:00 --constraint=cpu --account=
|
|
|
|
| 1 |
+
salloc --nodes=1 --ntasks=64 --cpus-per-task=1 --qos=interactive --time=04:00:00 --constraint=cpu --account=trn007
|
root_gnn_dgl/jobs/{inference.py → inference/run_inference.py}
RENAMED
|
File without changes
|
root_gnn_dgl/jobs/interactive.sh
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
salloc --nodes 1 --qos shared_interactive --time 04:00:00 --constraint gpu --account=
|
|
|
|
| 1 |
+
salloc --nodes 1 --qos shared_interactive --time 04:00:00 --constraint gpu --account=trn007 --gres=gpu:1
|
root_gnn_dgl/jobs/prep_data/prep_data.sh
ADDED
|
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#!/bin/bash
|
| 2 |
+
|
| 3 |
+
# Arguments
|
| 4 |
+
config=$1 # YAML configuration file
|
| 5 |
+
dataset=$2 # Dataset name
|
| 6 |
+
CHUNKS=$3 # Number of chunks
|
| 7 |
+
|
| 8 |
+
args=$4 # Additional arguments to pass to prep_data.py
|
| 9 |
+
|
| 10 |
+
# Prepare the initial command for chunk 0
|
| 11 |
+
command="scripts/prep_data.py --dataset $dataset --shuffle_mode --chunk 0 --config $config $args"
|
| 12 |
+
echo "$command"
|
| 13 |
+
python $command
|
| 14 |
+
|
| 15 |
+
# Loop through all chunks
|
| 16 |
+
for ((i=0; i<$CHUNKS; i++)); do
|
| 17 |
+
command="scripts/prep_data.py --dataset $dataset --shuffle_mode --chunk $i --config $config $args"
|
| 18 |
+
echo "$command"
|
| 19 |
+
python $command
|
| 20 |
+
done
|
root_gnn_dgl/jobs/{process_data.py → prep_data/run_processing.py}
RENAMED
|
@@ -71,26 +71,18 @@ def process_yaml_file(config, base_directory, bash_script):
|
|
| 71 |
|
| 72 |
def main():
|
| 73 |
# Base directory containing the YAML files
|
| 74 |
-
base_directory = "/
|
| 75 |
|
| 76 |
# List of YAML configuration files (relative to the base directory)
|
| 77 |
configs = [
|
| 78 |
-
|
| 79 |
-
|
| 80 |
-
|
| 81 |
-
|
| 82 |
-
# "run_3_ttH/v06/tHjb_CP_no_weights.yaml",
|
| 83 |
-
# "run_3_ttH/v06/tHjb_CP_weights.yaml",
|
| 84 |
-
"run_3_ttH/v06/ttH_CP_no_weights.yaml",
|
| 85 |
-
"run_3_ttH/v06/ttH_CP_weights.yaml",
|
| 86 |
-
"run_3_ttH/v06/tWH_CP_no_weights.yaml",
|
| 87 |
-
"run_3_ttH/v06/tWH_CP_weights.yaml",
|
| 88 |
-
"run_3_ttH/v06/ttH_tHjb_tWH_CP_no_weights.yaml",
|
| 89 |
-
"run_3_ttH/v06/ttH_tHjb_tWH_CP_weights.yaml",
|
| 90 |
]
|
| 91 |
|
| 92 |
# Path to the bash script to be called
|
| 93 |
-
bash_script = "/
|
| 94 |
|
| 95 |
# Use ThreadPoolExecutor to process YAML files concurrently
|
| 96 |
max_workers = os.cpu_count() # Use all available CPU cores
|
|
|
|
| 71 |
|
| 72 |
def main():
|
| 73 |
# Base directory containing the YAML files
|
| 74 |
+
base_directory = os.getcwd() + "/"
|
| 75 |
|
| 76 |
# List of YAML configuration files (relative to the base directory)
|
| 77 |
configs = [
|
| 78 |
+
"configs/stats_100K/pretraining_multiclass.yaml",
|
| 79 |
+
"configs/stats_100K/ttH_CP_even_vs_odd.yaml",
|
| 80 |
+
"configs/stats_all/pretraining_multiclass.yaml",
|
| 81 |
+
"configs/stats_all/ttH_CP_even_vs_odd.yaml",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 82 |
]
|
| 83 |
|
| 84 |
# Path to the bash script to be called
|
| 85 |
+
bash_script = base_directory + "jobs/prep_data.sh"
|
| 86 |
|
| 87 |
# Use ThreadPoolExecutor to process YAML files concurrently
|
| 88 |
max_workers = os.cpu_count() # Use all available CPU cores
|
root_gnn_dgl/jobs/salloc.sh
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
salloc --nodes 4 --qos interactive --time 04:00:00 --constraint gpu --account=
|
|
|
|
| 1 |
+
salloc --nodes 4 --qos interactive --time 04:00:00 --constraint gpu --account=trn007 --gres=gpu:4
|
root_gnn_dgl/jobs/{multinode → training/multinode}/run_multinode_1.sh
RENAMED
|
File without changes
|
root_gnn_dgl/jobs/{multinode → training/multinode}/run_multinode_2.sh
RENAMED
|
File without changes
|
root_gnn_dgl/jobs/{multinode → training/multinode}/run_multinode_3.sh
RENAMED
|
File without changes
|
root_gnn_dgl/jobs/{multinode/submit_multinode.sh → training/multinode/submit.sh}
RENAMED
|
File without changes
|
root_gnn_dgl/jobs/{singlegpu → training/singlegpu}/run_job.sh
RENAMED
|
File without changes
|
root_gnn_dgl/jobs/{singlegpu → training/singlegpu}/run_job_image.sh
RENAMED
|
File without changes
|
root_gnn_dgl/jobs/{singlegpu/submit_job.sh → training/singlegpu/submit.sh}
RENAMED
|
File without changes
|