| Universe = vanilla |
|
|
| RequestCpus = 4 |
| RequestMemory = 30G |
| +RequestWallTime = 30000 |
| request_GPUs = 1 |
|
|
| Requirements = (Has_avx)\ |
| &&((GPUs_GlobalMemoryMb >= 15000)||(CUDAGlobalMemoryMb >= 15000)) |
|
|
| NiceUser = true |
| initialdir = /users/spraak/jponcele/whisper/finetune_event/whisper_runs |
|
|
| Executable = /esat/spchtemp/scratch/jponcele/anaconda3/envs/whisper/bin/python |
| Arguments = "run_eval_whisper_streaming_local.py --model_size=$(size) --dataset=$(dataset) --device=gpu --batch_size=$(bs)" |
|
|
| Log = /esat/audioslave/jponcele/whisper/finetuning_event/CGN/condor/condor-eval-$(dataset)-$(size).log |
| Output = /esat/audioslave/jponcele/whisper/finetuning_event/CGN/condor/condor-eval-$(dataset)-$(size).out |
| Error = /esat/audioslave/jponcele/whisper/finetuning_event/CGN/condor/condor-eval-$(dataset)-$(size).err |
|
|
| Queue |
|
|