BoxOfColors commited on
Commit
1d1ce31
·
1 Parent(s): 8635f79

Fix TARO GPU abort: increase load_overhead 15→120s to cover model loading time

Browse files

TARO loads CAVP extractor, onset model, MMDiT, AudioLDM2 VAE, and BigVGAN
vocoder before inference. This took >60s total, exhausting the GPU budget
before _load_taro_models could complete. 120s overhead + inference time
gives enough headroom for all model loading + feature extraction.

Files changed (1) hide show
  1. app.py +1 -1
app.py CHANGED
@@ -512,7 +512,7 @@ MODEL_CONFIGS = {
512
  "window_s": TARO_MODEL_DUR, # 8.192 s
513
  "sr": TARO_SR, # 16000 (output resampled to TARGET_SR)
514
  "secs_per_step": 0.025, # measured 0.023 s/step on H200
515
- "load_overhead": 15, # model load + CAVP feature extraction
516
  "tab_prefix": "taro",
517
  "label": "TARO",
518
  "regen_fn": None, # set after function definitions (avoids forward-ref)
 
512
  "window_s": TARO_MODEL_DUR, # 8.192 s
513
  "sr": TARO_SR, # 16000 (output resampled to TARGET_SR)
514
  "secs_per_step": 0.025, # measured 0.023 s/step on H200
515
+ "load_overhead": 120, # CAVP+onset+MMDiT+VAE+vocoder load + feature extraction
516
  "tab_prefix": "taro",
517
  "label": "TARO",
518
  "regen_fn": None, # set after function definitions (avoids forward-ref)