Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -475,13 +475,13 @@ def auto_ensemble_process(audio, model_keys, seg_size=64, overlap=0.1, out_forma
|
|
| 475 |
scipy.io.wavfile.write(temp_audio_path, sample_rate, data)
|
| 476 |
audio = temp_audio_path
|
| 477 |
|
| 478 |
-
#
|
| 479 |
-
if ensemble_state["current_audio"] != audio:
|
| 480 |
ensemble_state["current_audio"] = audio
|
| 481 |
ensemble_state["current_model_idx"] = 0
|
| 482 |
ensemble_state["processed_stems"] = []
|
| 483 |
ensemble_state["model_outputs"] = {model_key: {"vocals": [], "other": []} for model_key in model_keys}
|
| 484 |
-
logger.info("New audio
|
| 485 |
|
| 486 |
use_tta = use_tta == "True"
|
| 487 |
base_name = os.path.splitext(os.path.basename(audio))[0]
|
|
@@ -510,7 +510,6 @@ def auto_ensemble_process(audio, model_keys, seg_size=64, overlap=0.1, out_forma
|
|
| 510 |
for model_key, stems_dict in ensemble_state["model_outputs"].items():
|
| 511 |
for stem_type in ["vocals", "other"]:
|
| 512 |
if stems_dict[stem_type]:
|
| 513 |
-
# Stem tipini kontrol et, excluded listesinde varsa atla
|
| 514 |
if stem_type.lower() in excluded_stems_list:
|
| 515 |
logger.info(f"Excluding {stem_type} for {model_key} from ensemble")
|
| 516 |
continue
|
|
@@ -670,7 +669,7 @@ def auto_ensemble_process(audio, model_keys, seg_size=64, overlap=0.1, out_forma
|
|
| 670 |
if torch.cuda.is_available():
|
| 671 |
torch.cuda.empty_cache()
|
| 672 |
logger.info("GPU memory cleared")
|
| 673 |
-
|
| 674 |
def update_roformer_models(category):
|
| 675 |
"""Update Roformer model dropdown based on selected category."""
|
| 676 |
choices = list(ROFORMER_MODELS.get(category, {}).keys()) or []
|
|
|
|
| 475 |
scipy.io.wavfile.write(temp_audio_path, sample_rate, data)
|
| 476 |
audio = temp_audio_path
|
| 477 |
|
| 478 |
+
# State yönetimini kontrol et ve sıfırlama
|
| 479 |
+
if ensemble_state["current_audio"] != audio or ensemble_state["current_model_idx"] >= len(model_keys):
|
| 480 |
ensemble_state["current_audio"] = audio
|
| 481 |
ensemble_state["current_model_idx"] = 0
|
| 482 |
ensemble_state["processed_stems"] = []
|
| 483 |
ensemble_state["model_outputs"] = {model_key: {"vocals": [], "other": []} for model_key in model_keys}
|
| 484 |
+
logger.info("New audio or completed cycle detected, resetting ensemble state.")
|
| 485 |
|
| 486 |
use_tta = use_tta == "True"
|
| 487 |
base_name = os.path.splitext(os.path.basename(audio))[0]
|
|
|
|
| 510 |
for model_key, stems_dict in ensemble_state["model_outputs"].items():
|
| 511 |
for stem_type in ["vocals", "other"]:
|
| 512 |
if stems_dict[stem_type]:
|
|
|
|
| 513 |
if stem_type.lower() in excluded_stems_list:
|
| 514 |
logger.info(f"Excluding {stem_type} for {model_key} from ensemble")
|
| 515 |
continue
|
|
|
|
| 669 |
if torch.cuda.is_available():
|
| 670 |
torch.cuda.empty_cache()
|
| 671 |
logger.info("GPU memory cleared")
|
| 672 |
+
|
| 673 |
def update_roformer_models(category):
|
| 674 |
"""Update Roformer model dropdown based on selected category."""
|
| 675 |
choices = list(ROFORMER_MODELS.get(category, {}).keys()) or []
|