File size: 38,415 Bytes
cc64221
 
 
 
 
5e51096
f87864c
cc64221
 
 
 
 
 
489680b
fd43400
 
 
a05e815
76a0025
489680b
57b3cc7
 
 
 
fd43400
 
 
 
1bed755
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fd43400
 
 
 
489680b
fd43400
489680b
fd43400
 
489680b
fd43400
 
 
 
 
 
 
1bed755
489680b
fd43400
5e51096
cc64221
 
 
a05e815
5e51096
a05e815
cc64221
 
 
fdd4054
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cc64221
 
 
fdd4054
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cc64221
 
 
fdd4054
cc64221
 
 
fdd4054
 
 
 
 
 
 
 
cc64221
 
 
fdd4054
 
 
cc64221
 
 
fdd4054
 
cc64221
 
 
fdd4054
 
 
 
 
 
 
 
 
 
 
cc64221
 
 
 
 
a05e815
3b02063
cc64221
bfd82aa
 
 
cc64221
bfd82aa
 
 
 
cc64221
 
 
bfd82aa
 
 
 
 
 
 
cc64221
bfd82aa
cc64221
bfd82aa
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cc64221
 
bfd82aa
 
 
 
cc64221
 
bfd82aa
cc64221
 
 
bfd82aa
 
cc64221
 
bfd82aa
cc64221
 
bfd82aa
 
 
 
 
cc64221
 
 
 
 
bfd82aa
 
 
cc64221
 
bc8e842
 
db94774
cc64221
 
bfd82aa
bc8e842
 
 
 
 
 
 
 
 
 
 
 
 
 
bfd82aa
 
bc8e842
 
a05e815
bc8e842
bfd82aa
bc8e842
a05e815
bc8e842
 
 
 
db94774
 
 
 
bc8e842
 
 
 
 
 
 
bfd82aa
cc64221
bfd82aa
 
 
 
 
 
cc64221
bfd82aa
 
 
 
 
 
cc64221
bfd82aa
 
 
 
cc64221
bfd82aa
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cc64221
 
 
9e5c4b5
 
4f96bfa
9e5c4b5
6db3d10
9e5c4b5
 
cc64221
9e5c4b5
e121019
 
 
 
 
 
67568d2
 
 
 
 
e121019
bc8e842
 
1bed755
bc8e842
cc64221
e121019
9e5c4b5
 
e121019
4f96bfa
 
6db3d10
390f3f3
 
 
6db3d10
 
 
 
6bc0914
6db3d10
 
 
 
 
 
 
9e5c4b5
6bc0914
6db3d10
 
 
 
9e5c4b5
e121019
9e5c4b5
390f3f3
9e5c4b5
 
4f96bfa
9e5c4b5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6db3d10
9e5c4b5
a05e815
cc64221
 
 
 
 
7b79193
cc64221
7b79193
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cc64221
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
aecae1e
7b79193
 
 
 
 
 
 
cc64221
a05e815
8205184
e14bd0f
8205184
7b79193
8205184
 
e14bd0f
8205184
7b79193
 
 
 
 
cc64221
8205184
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7b79193
cc64221
8205184
7b79193
 
 
 
 
 
 
 
8205184
7b79193
8205184
 
7b79193
 
 
 
 
8205184
7b79193
 
8205184
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7b79193
 
 
 
8205184
7b79193
 
 
 
 
 
 
 
 
 
 
8205184
7b79193
 
e14bd0f
7b79193
 
 
 
 
8205184
 
7b79193
8205184
 
 
7b79193
8205184
7b79193
cc64221
 
fd43400
 
bc8e842
cc64221
 
 
fd43400
 
bc8e842
cc64221
b956722
 
 
 
cc64221
 
 
1bed755
cc64221
 
 
 
 
 
 
67568d2
cc64221
489680b
 
cc64221
 
 
 
 
 
 
 
 
67568d2
cc64221
b956722
cc64221
 
 
a4f6734
cc64221
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
67568d2
cc64221
b956722
cc64221
 
 
a4f6734
cc64221
 
 
 
 
 
 
 
 
 
 
 
 
 
6bc0914
b956722
6bc0914
b956722
6bc0914
cc64221
6bc0914
cc64221
 
 
 
 
 
 
 
6bc0914
b956722
6bc0914
b956722
6bc0914
cc64221
489680b
cc64221
 
bb358dc
cc64221
 
 
 
 
 
a3908c3
 
 
 
 
 
 
 
 
cb4dec1
a3908c3
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
import os
import sys
import torch
import logging
from yt_dlp import YoutubeDL
import gradio as gr
import argparse
from audio_separator.separator import Separator
import numpy as np
import librosa
import soundfile as sf
from ensemble import ensemble_files
import shutil
import gradio_client.utils as client_utils
import validators
import matchering as mg
from typing import Any, Optional
import spaces  # Added for @spaces.GPU
import gdown 

# Add this import if not already present
import yt_dlp
import scipy.io.wavfile  # For audio processing

# Logging setup
logging.basicConfig(level=logging.INFO)
logger = logging.getLogger(__name__)

# Configuration directory
CONFIG_DIR = "/tmp/SESA-Config"

def ensure_config_dir():
    """Ensure the configuration directory exists and is writable."""
    try:
        os.makedirs(CONFIG_DIR, exist_ok=True)
        logger.info(f"Configuration directory ensured: {CONFIG_DIR}")
    except PermissionError as e:
        logger.error(f"Failed to create config directory {CONFIG_DIR}: {e}")
        raise RuntimeError(f"Cannot create config directory: {e}")
    except Exception as e:
        logger.error(f"Unexpected error creating config directory {CONFIG_DIR}: {e}")
        raise

# Call early in the script
ensure_config_dir()

# Patch gradio_client.utils._json_schema_to_python_type
original_json_schema_to_python_type = client_utils._json_schema_to_python_type

def patched_json_schema_to_python_type(schema: Any, defs: Optional[dict] = None) -> str:
    logger.debug(f"Parsing schema: {schema}")
    if isinstance(schema, bool):
        logger.info("Found boolean schema, returning 'boolean'")
        return "boolean"
    if not isinstance(schema, dict):
        logger.warning(f"Unexpected schema type: {type(schema)}, returning 'Any'")
        return "Any"
    if "enum" in schema and schema.get("type") == "string":
        logger.info(f"Handling enum schema: {schema['enum']}")
        return f"Literal[{', '.join(repr(e) for e in schema['enum'])}]"
    try:
        return original_json_schema_to_python_type(schema, defs)
    except client_utils.APIInfoParseError as e:
        logger.error(f"Failed to parse schema {schema}: {e}")
        return "str"

client_utils._json_schema_to_python_type = patched_json_schema_to_python_type

# Device and autocast setup
device = "cuda" if torch.cuda.is_available() else "cpu"
use_autocast = device == "cuda"
logger.info(f"Using device: {device}")

# ROFORMER_MODELS and OUTPUT_FORMATS
ROFORMER_MODELS = {
    "Vocals": {
        'MelBand Roformer | Big Beta 6X by unwa': 'melband_roformer_big_beta6x.ckpt',
        'MelBand Roformer Kim | Big Beta 4 FT by unwa': 'melband_roformer_big_beta4.ckpt',
        'MelBand Roformer Kim | Big Beta 5e FT by unwa': 'melband_roformer_big_beta5e.ckpt',
        'MelBand Roformer | Big Beta 6 by unwa': 'melband_roformer_big_beta6.ckpt',
        'MelBand Roformer | Vocals by Kimberley Jensen': 'vocals_mel_band_roformer.ckpt',
        'MelBand Roformer Kim | FT 3 by unwa': 'mel_band_roformer_kim_ft3_unwa.ckpt',
        'MelBand Roformer Kim | FT by unwa': 'mel_band_roformer_kim_ft_unwa.ckpt',
        'MelBand Roformer Kim | FT 2 by unwa': 'mel_band_roformer_kim_ft2_unwa.ckpt',
        'MelBand Roformer Kim | FT 2 Bleedless by unwa': 'mel_band_roformer_kim_ft2_bleedless_unwa.ckpt',
        'MelBand Roformer | Vocals by becruily': 'mel_band_roformer_vocals_becruily.ckpt',
        'MelBand Roformer | Vocals Fullness by Aname': 'mel_band_roformer_vocal_fullness_aname.ckpt',
        'BS Roformer | Vocals by Gabox': 'bs_roformer_vocals_gabox.ckpt',
        'MelBand Roformer | Vocals by Gabox': 'mel_band_roformer_vocals_gabox.ckpt',
        'MelBand Roformer | Vocals FV1 by Gabox': 'mel_band_roformer_vocals_fv1_gabox.ckpt',
        'MelBand Roformer | Vocals FV2 by Gabox': 'mel_band_roformer_vocals_fv2_gabox.ckpt',
        'MelBand Roformer | Vocals FV3 by Gabox': 'mel_band_roformer_vocals_fv3_gabox.ckpt',
        'MelBand Roformer | Vocals FV4 by Gabox': 'mel_band_roformer_vocals_fv4_gabox.ckpt',
        'BS Roformer | Chorus Male-Female by Sucial': 'model_chorus_bs_roformer_ep_267_sdr_24.1275.ckpt',
        'BS Roformer | Male-Female by aufr33': 'bs_roformer_male_female_by_aufr33_sdr_7.2889.ckpt',
    },
    "Instrumentals": {
        'MelBand Roformer | FVX by Gabox': 'mel_band_roformer_instrumental_fvx_gabox.ckpt',
        'MelBand Roformer | INSTV8N by Gabox': 'mel_band_roformer_instrumental_instv8n_gabox.ckpt',
        'MelBand Roformer | INSTV8 by Gabox': 'mel_band_roformer_instrumental_instv8_gabox.ckpt',
        'MelBand Roformer | INSTV7N by Gabox': 'mel_band_roformer_instrumental_instv7n_gabox.ckpt',
        'MelBand Roformer | Instrumental Bleedless V3 by Gabox': 'mel_band_roformer_instrumental_bleedless_v3_gabox.ckpt',
        'MelBand Roformer Kim | Inst V1 (E) Plus by Unwa': 'melband_roformer_inst_v1e_plus.ckpt',
        'MelBand Roformer Kim | Inst V1 Plus by Unwa': 'melband_roformer_inst_v1_plus.ckpt',
        'MelBand Roformer Kim | Inst V1 by Unwa': 'melband_roformer_inst_v1.ckpt',
        'MelBand Roformer Kim | Inst V1 (E) by Unwa': 'melband_roformer_inst_v1e.ckpt',
        'MelBand Roformer Kim | Inst V2 by Unwa': 'melband_roformer_inst_v2.ckpt',
        'MelBand Roformer | Instrumental by becruily': 'mel_band_roformer_instrumental_becruily.ckpt',
        'MelBand Roformer | Instrumental by Gabox': 'mel_band_roformer_instrumental_gabox.ckpt',
        'MelBand Roformer | Instrumental 2 by Gabox': 'mel_band_roformer_instrumental_2_gabox.ckpt',
        'MelBand Roformer | Instrumental 3 by Gabox': 'mel_band_roformer_instrumental_3_gabox.ckpt',
        'MelBand Roformer | Instrumental Bleedless V1 by Gabox': 'mel_band_roformer_instrumental_bleedless_v1_gabox.ckpt',
        'MelBand Roformer | Instrumental Bleedless V2 by Gabox': 'mel_band_roformer_instrumental_bleedless_v2_gabox.ckpt',
        'MelBand Roformer | Instrumental Fullness V1 by Gabox': 'mel_band_roformer_instrumental_fullness_v1_gabox.ckpt',
        'MelBand Roformer | Instrumental Fullness V2 by Gabox': 'mel_band_roformer_instrumental_fullness_v2_gabox.ckpt',
        'MelBand Roformer | Instrumental Fullness V3 by Gabox': 'mel_band_roformer_instrumental_fullness_v3_gabox.ckpt',
        'MelBand Roformer | Instrumental Fullness Noisy V4 by Gabox': 'mel_band_roformer_instrumental_fullness_noise_v4_gabox.ckpt',
        'MelBand Roformer | INSTV5 by Gabox': 'mel_band_roformer_instrumental_instv5_gabox.ckpt',
        'MelBand Roformer | INSTV5N by Gabox': 'mel_band_roformer_instrumental_instv5n_gabox.ckpt',
        'MelBand Roformer | INSTV6 by Gabox': 'mel_band_roformer_instrumental_instv6_gabox.ckpt',
        'MelBand Roformer | INSTV6N by Gabox': 'mel_band_roformer_instrumental_instv6n_gabox.ckpt',
        'MelBand Roformer | INSTV7 by Gabox': 'mel_band_roformer_instrumental_instv7_gabox.ckpt',
    },
    "InstVoc Duality": {
        'MelBand Roformer Kim | InstVoc Duality V1 by Unwa': 'melband_roformer_instvoc_duality_v1.ckpt',
        'MelBand Roformer Kim | InstVoc Duality V2 by Unwa': 'melband_roformer_instvox_duality_v2.ckpt',
    },
    "De-Reverb": {
        'BS-Roformer-De-Reverb': 'deverb_bs_roformer_8_384dim_10depth.ckpt',
        'MelBand Roformer | De-Reverb by anvuew': 'dereverb_mel_band_roformer_anvuew_sdr_19.1729.ckpt',
        'MelBand Roformer | De-Reverb Less Aggressive by anvuew': 'dereverb_mel_band_roformer_less_aggressive_anvuew_sdr_18.8050.ckpt',
        'MelBand Roformer | De-Reverb Mono by anvuew': 'dereverb_mel_band_roformer_mono_anvuew.ckpt',
        'MelBand Roformer | De-Reverb Big by Sucial': 'dereverb_big_mbr_ep_362.ckpt',
        'MelBand Roformer | De-Reverb Super Big by Sucial': 'dereverb_super_big_mbr_ep_346.ckpt',
        'MelBand Roformer | De-Reverb-Echo by Sucial': 'dereverb-echo_mel_band_roformer_sdr_10.0169.ckpt',
        'MelBand Roformer | De-Reverb-Echo V2 by Sucial': 'dereverb-echo_mel_band_roformer_sdr_13.4843_v2.ckpt',
        'MelBand Roformer | De-Reverb-Echo Fused by Sucial': 'dereverb_echo_mbr_fused.ckpt',
    },
    "Denoise": {
        'Mel-Roformer-Denoise-Aufr33': 'denoise_mel_band_roformer_aufr33_sdr_27.9959.ckpt',
        'Mel-Roformer-Denoise-Aufr33-Aggr': 'denoise_mel_band_roformer_aufr33_aggr_sdr_27.9768.ckpt',
        'MelBand Roformer | Denoise-Debleed by Gabox': 'mel_band_roformer_denoise_debleed_gabox.ckpt',
        'MelBand Roformer | Bleed Suppressor V1 by unwa-97chris': 'mel_band_roformer_bleed_suppressor_v1.ckpt',
    },
    "Karaoke": {
        'Mel-Roformer-Karaoke-Aufr33-Viperx': 'mel_band_roformer_karaoke_aufr33_viperx_sdr_10.1956.ckpt',
        'MelBand Roformer | Karaoke by Gabox': 'mel_band_roformer_karaoke_gabox.ckpt',
        'MelBand Roformer | Karaoke by becruily': 'mel_band_roformer_karaoke_becruily.ckpt',
    },
    "General Purpose": {
        'BS-Roformer-Viperx-1297': 'model_bs_roformer_ep_317_sdr_12.9755.ckpt',
        'BS-Roformer-Viperx-1296': 'model_bs_roformer_ep_368_sdr_12.9628.ckpt',
        'BS-Roformer-Viperx-1053': 'model_bs_roformer_ep_937_sdr_10.5309.ckpt',
        'Mel-Roformer-Viperx-1143': 'model_mel_band_roformer_ep_3005_sdr_11.4360.ckpt',
        'Mel-Roformer-Crowd-Aufr33-Viperx': 'mel_band_roformer_crowd_aufr33_viperx_sdr_8.7144.ckpt',
        'MelBand Roformer Kim | SYHFT by SYH99999': 'MelBandRoformerSYHFT.ckpt',
        'MelBand Roformer Kim | SYHFT V2 by SYH99999': 'MelBandRoformerSYHFTV2.ckpt',
        'MelBand Roformer Kim | SYHFT V2.5 by SYH99999': 'MelBandRoformerSYHFTV2.5.ckpt',
        'MelBand Roformer Kim | SYHFT V3 by SYH99999': 'MelBandRoformerSYHFTV3Epsilon.ckpt',
        'MelBand Roformer Kim | Big SYHFT V1 by SYH99999': 'MelBandRoformerBigSYHFTV1.ckpt',
        'MelBand Roformer | Aspiration by Sucial': 'aspiration_mel_band_roformer_sdr_18.9845.ckpt',
        'MelBand Roformer | Aspiration Less Aggressive by Sucial': 'aspiration_mel_band_roformer_less_aggr_sdr_18.1201.ckpt',
    }
}

OUTPUT_FORMATS = ['wav', 'flac', 'mp3', 'ogg', 'opus', 'm4a', 'aiff', 'ac3']

# CSS (unchanged from provided version)
CSS = """
body {
    background: linear-gradient(to bottom, rgba(45, 11, 11, 0.9), rgba(0, 0, 0, 0.8)), url('/content/logo.jpg') no-repeat center center fixed;
    background-size: cover;
    min-height: 100vh;
    margin: 0;
    padding: 1rem;
    font-family: 'Poppins', sans-serif;
    color: #C0C0C0;
    overflow-x: hidden;
}
.header-text {
    text-align: center;
    padding: 100px 20px 20px;
    color: #ff4040;
    font-size: 3rem;
    font-weight: 900;
    text-shadow: 0 0 10px rgba(255, 64, 64, 0.5);
    z-index: 1500;
    animation: text-glow 2s infinite;
}
.header-subtitle {
    text-align: center;
    color: #C0C0C0;
    font-size: 1.2rem;
    font-weight: 300;
    margin-top: -10px;
    text-shadow: 0 0 5px rgba(255, 64, 64, 0.3);
}
.gr-tab {
    background: rgba(128, 0, 0, 0.5) !important;
    border-radius: 12px 12px 0 0 !important;
    margin: 0 5px !important;
    color: #C0C0C0 !important;
    border: 1px solid #ff4040 !important;
    z-index: 1500;
    transition: background 0.3s ease, color 0.3s ease;
    padding: 10px 20px !important;
    font-size: 1.1rem !important;
}
button {
    transition: all 0.3s cubic-bezier(0.4, 0, 0.2, 1) !important;
    background: #800000 !important;
    border: 1px solid #ff4040 !important;
    color: #C0C0C0 !important;
    border-radius: 8px !important;
    padding: 8px 16px !important;
    box-shadow: 0 2px 10px rgba(255, 64, 64, 0.3);
}
button:hover {
    transform: scale(1.05) !important;
    box-shadow: 0 10px 40px rgba(255, 64, 64, 0.7) !important;
    background: #ff4040 !important;
}
.compact-upload.horizontal {
    display: inline-flex !important;
    align-items: center !important;
    gap: 8px !important;
    max-width: 400px !important;
    height: 40px !important;
    padding: 0 12px !important;
    border: 1px solid #ff4040 !important;
    background: rgba(128, 0, 0, 0.5) !important;
    border-radius: 8px !important;
}
.compact-dropdown {
    padding: 8px 12px !important;
    border-radius: 8px !important;
    border: 2px solid #ff6b6b !important; /* Match previous style */
    background: rgba(46, 26, 71, 0.7) !important; /* Match previous style */
    color: #e0e0e0 !important; /* Match previous style */
    width: 100%;
    font-size: 1rem !important;
    transition: border-color 0.3s ease, box-shadow 0.3s ease !important;
    position: relative;
    z-index: 100; /* Base z-index for dropdown */
}
.compact-dropdown:hover {
    border-color: #ff8787 !important; /* Lighter shade for hover */
    box-shadow: 0 2px 8px rgba(255, 107, 107, 0.4) !important;
}
.compact-dropdown select, .compact-dropdown .gr-dropdown {
    background: transparent !important;
    color: #e0e0e0 !important;
    border: none !important;
    width: 100% !important;
    padding: 8px !important;
    font-size: 1rem !important;
    appearance: none !important;
    -webkit-appearance: none !important;
    -moz-appearance: none !important;
}
.compact-dropdown .gr-dropdown-menu {
    background: rgba(46, 26, 71, 0.95) !important; /* Match previous style */
    border: 2px solid #ff6b6b !important; /* Match previous style */
    border-radius: 8px !important;
    color: #e0e0e0 !important;
    max-height: 300px !important;
    overflow-y: auto !important;
    z-index: 300 !important; /* Ensure it stays above other elements */
    width: 100% !important;
    opacity: 1 !important;
    visibility: visible !important;
    position: absolute !important;
    top: 100% !important;
    left: 0 !important;
    pointer-events: auto !important;
}
.compact-dropdown:hover .gr-dropdown-menu {
    display: block !important; /* Ensure menu stays open on hover */
}
.compact-dropdown .gr-dropdown-menu option {
    padding: 8px !important;
    color: #e0e0e0 !important;
    background: transparent !important;
}
.compact-dropdown .gr-dropdown-menu option:hover {
    background: rgba(255, 107, 107, 0.3) !important; /* Match hover effect */
}
#custom-progress {
    margin-top: 10px;
    padding: 10px;
    background: rgba(128, 0, 0, 0.3);
    border-radius: 8px;
    border: 1px solid #ff4040;
}
#progress-bar {
    height: 20px;
    background: linear-gradient(to right, #6e8efb, #ff4040);
    border-radius: 5px;
    transition: width 0.5s ease-in-out;
    max-width: 100% !important;
}
.gr-accordion {
    background: rgba(128, 0, 0, 0.5) !important;
    border-radius: 10px !important;
    border: 1px solid #ff4040 !important;
}
.footer {
    text-align: center;
    padding: 20px;
    color: #ff4040;
    font-size: 14px;
    margin-top: 40px;
    background: rgba(128, 0, 0, 0.3);
    border-top: 1px solid #ff4040;
}
#log-accordion {
    max-height: 400px;
    overflow-y: auto;
    background: rgba(0, 0, 0, 0.7) !important;
    padding: 10px;
    border-radius: 8px;
}
@keyframes text-glow {
    0% { text-shadow: 0 0 5px rgba(192, 192, 192, 0); }
    50% { text-shadow: 0 0 15px rgba(192, 192, 192, 1); }
    100% { text-shadow: 0 0 5px rgba(192, 192, 192, 0); }
}
"""

import os
import yt_dlp
import gdown
from scipy.io import wavfile
from pydub import AudioSegment

def download_audio(url, cookie_file=None):
    ydl_opts = {
        'format': 'bestaudio[ext=webm]/bestaudio[ext=m4a]/bestaudio[ext=opus]/bestaudio[ext=aac]/bestaudio -video',
        'postprocessors': [{
            'key': 'FFmpegExtractAudio',
            'preferredcodec': 'wav',
            'preferredquality': '192',
        }],
        'outtmpl': 'ytdl/%(title)s.%(ext)s',
        'user_agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36',
        'geo_bypass': True,
        'force_ipv4': True,
        'referer': 'https://www.youtube.com/',
        'noplaylist': True,
        'cookiefile': cookie_file.name if cookie_file else None,
        'extractor_retries': 5,
        'ignoreerrors': False,
        'no_check_certificate': True,
        'verbose': True,
    }

    # Check if it's a Google Drive link
    if 'drive.google.com' in url or 'https://drive.google.com' in url:
        try:
            # Create the 'ytdl' directory if it doesn't exist
            os.makedirs('ytdl', exist_ok=True)

            # Extract file ID from the URL
            file_id = url.split('/d/')[1].split('/')[0]
            download_url = f'https://drive.google.com/uc?id={file_id}'
            temp_output_path = 'ytdl/gdrive_temp_audio'  # Temporary file for raw download
            gdown.download(download_url, temp_output_path, quiet=False)

            if not os.path.exists(temp_output_path):
                return None, "Downloaded file not found", None

            # Convert the downloaded file to WAV using pydub
            output_path = 'ytdl/gdrive_audio.wav'
            audio = AudioSegment.from_file(temp_output_path)
            audio.export(output_path, format="wav")

            # Read the converted WAV file
            sample_rate, data = wavfile.read(output_path)
            audio_data = (sample_rate, data)
            
            # Clean up the temporary file
            os.remove(temp_output_path)
            
            return output_path, "Download successful", audio_data
        except Exception as e:
            return None, f"Google Drive download failed: {str(e)}", None

    # Handle YouTube link
    else:
        os.makedirs('ytdl', exist_ok=True)
        with yt_dlp.YoutubeDL(ydl_opts) as ydl:
            try:
                info_dict = ydl.extract_info(url, download=True)
                base_file_path = ydl.prepare_filename(info_dict)
                file_path = base_file_path
                for ext in ['.webm', '.m4a', '.opus', '.aac']:
                    file_path = file_path.replace(ext, '.wav')
                if not os.path.exists(file_path):
                    return None, "Downloaded file not found", None
                sample_rate, data = wavfile.read(file_path)
                audio_data = (sample_rate, data)
                return file_path, "Download successful", audio_data
            except yt_dlp.utils.ExtractorError as e:
                if "Sign in to confirm you’re not a bot" in str(e) or "The provided YouTube account cookies are no longer valid" in str(e):
                    return None, "Authentication failed. Please upload updated cookies from a logged-in browser session in the respective tab. See https://github.com/yt-dlp/yt-dlp/wiki/Extractors#exporting-youtube-cookies for instructions.", None
                elif "HTTP Error 403: Forbidden" in str(e):
                    return None, "Download failed: HTTP Error 403. This format requires a GVS PO Token, or the cookies are invalid. Please upload fresh cookies. See https://github.com/yt-dlp/yt-dlp/wiki/PO-Token-Guide for advanced troubleshooting.", None
                return None, f"Download failed: {str(e)}", None
            except Exception as e:
                return None, f"Unexpected error: {str(e)}", None 
                
@spaces.GPU
def roformer_separator(audio, model_key, seg_size, override_seg_size, overlap, pitch_shift, model_dir, output_dir, out_format, norm_thresh, amp_thresh, batch_size, exclude_stems="", progress=gr.Progress(track_tqdm=True)):
    """Separate audio into stems using a Roformer model."""
    if not audio:
        raise ValueError("No audio file provided.")
    
    temp_audio_path = None  # Initialize to None to avoid undefined variable in finally
    try:
        # If audio is a tuple (sample_rate, data), save it as a temporary file
        if isinstance(audio, tuple):
            sample_rate, data = audio
            temp_audio_path = os.path.join("/tmp", "temp_audio.wav")
            scipy.io.wavfile.write(temp_audio_path, sample_rate, data)
            audio = temp_audio_path
        
        override_seg_size = override_seg_size == "True"
        
        if os.path.exists(output_dir):
            shutil.rmtree(output_dir)
        os.makedirs(output_dir, exist_ok=True)
        
        base_name = os.path.splitext(os.path.basename(audio))[0]
        for category, models in ROFORMER_MODELS.items():
            if model_key in models:
                model = models[model_key]
                break
        else:
            raise ValueError(f"Model '{model_key}' not found.")
        
        logger.info(f"Separating {base_name} with {model_key} on {device}")
        separator = Separator(
            log_level=logging.INFO,
            model_file_dir=model_dir,
            output_dir=output_dir,
            output_format=out_format,
            normalization_threshold=norm_thresh,
            amplification_threshold=amp_thresh,
            use_autocast=use_autocast,
            mdxc_params={"segment_size": seg_size, "override_model_segment_size": override_seg_size, "batch_size": batch_size, "overlap": overlap, "pitch_shift": pitch_shift}
        )
        progress(0.2, desc="Loading model...")
        separator.load_model(model_filename=model)
        progress(0.7, desc="Separating audio...")
        separation = separator.separate(audio)
        stems = [os.path.join(output_dir, file_name) for file_name in separation]
        
        if exclude_stems.strip():
            excluded = [s.strip().lower() for s in exclude_stems.split(',')]
            filtered_stems = [stem for stem in stems if not any(ex in os.path.basename(stem).lower() for ex in excluded)]
            return filtered_stems[0] if filtered_stems else None, filtered_stems[1] if len(filtered_stems) > 1 else None
        return stems[0], stems[1] if len(stems) > 1 else None
    except Exception as e:
        logger.error(f"Separation failed: {e}")
        raise RuntimeError(f"Separation failed: {e}")
    finally:
        # Clean up temporary file if it was created
        if temp_audio_path and os.path.exists(temp_audio_path):
            try:
                os.remove(temp_audio_path)
                logger.info(f"Cleaned up temporary file: {temp_audio_path}")
            except Exception as e:
                logger.warning(f"Failed to clean up temporary file {temp_audio_path}: {e}")

@spaces.GPU
def auto_ensemble_process(audio, model_keys, seg_size=128, overlap=0.1, out_format="wav", use_tta="False", model_dir="/tmp/audio-separator-models/", output_dir="output", norm_thresh=0.9, amp_thresh=0.9, batch_size=1, ensemble_method="avg_wave", exclude_stems="", weights_str=""):
    temp_audio_path = None
    chunk_paths = []
    try:
        if not audio or not model_keys:
            raise ValueError("Audio or models missing.")
        
        # Handle tuple input (sample_rate, data)
        if isinstance(audio, tuple):
            sample_rate, data = audio
            temp_audio_path = os.path.join("/tmp", "temp_audio.wav")
            scipy.io.wavfile.write(temp_audio_path, sample_rate, data)
            audio = temp_audio_path
        
        # Load audio to check duration
        audio_data, sr = librosa.load(audio, sr=None, mono=False)
        duration = librosa.get_duration(y=audio_data, sr=sr)
        logger.info(f"Audio duration: {duration:.2f} seconds")
        
        # Split audio if longer than 15 minutes (900 seconds)
        chunk_duration = 300  # 5 minutes in seconds
        chunks = []
        if duration > 900:
            logger.info(f"Audio exceeds 15 minutes, splitting into {chunk_duration}-second chunks")
            num_chunks = int(np.ceil(duration / chunk_duration))
            for i in range(num_chunks):
                start = i * chunk_duration * sr
                end = min((i + 1) * chunk_duration * sr, audio_data.shape[-1])
                chunk_data = audio_data[:, start:end] if audio_data.ndim == 2 else audio_data[start:end]
                chunk_path = os.path.join("/tmp", f"chunk_{i}.wav")
                sf.write(chunk_path, chunk_data.T if audio_data.ndim == 2 else chunk_data, sr)
                chunks.append(chunk_path)
                chunk_paths.append(chunk_path)
                logger.info(f"Created chunk {i}: {chunk_path}")
        else:
            chunks = [audio]
        
        use_tta = use_tta == "True"
        
        # Create output directory
        if os.path.exists(output_dir):
            shutil.rmtree(output_dir)
        os.makedirs(output_dir, exist_ok=True)
        
        base_name = os.path.splitext(os.path.basename(audio))[0]
        logger.info(f"Ensemble for {base_name} with {model_keys} on {device}")
        
        all_stems = []
        model_stems = {}  # Store stems per model for concatenation
        
        for model_key in model_keys:
            model_stems[model_key] = {"vocals": [], "other": []}
            for category, models in ROFORMER_MODELS.items():
                if model_key in models:
                    model = models[model_key]
                    break
            else:
                logger.warning(f"Model {model_key} not found, skipping")
                continue
            
            for chunk_idx, chunk_path in enumerate(chunks):
                separator = Separator(
                    log_level=logging.INFO,
                    model_file_dir=model_dir,
                    output_dir=output_dir,
                    output_format=out_format,
                    normalization_threshold=norm_thresh,
                    amplification_threshold=amp_thresh,
                    use_autocast=use_autocast,
                    mdxc_params={"segment_size": seg_size, "overlap": overlap, "use_tta": use_tta, "batch_size": batch_size}
                )
                logger.info(f"Loading {model_key} for chunk {chunk_idx}")
                separator.load_model(model_filename=model)
                logger.info(f"Separating chunk {chunk_idx} with {model_key}")
                separation = separator.separate(chunk_path)
                stems = [os.path.join(output_dir, file_name) for file_name in separation]
                
                # Store stems for this chunk
                for stem in stems:
                    if "vocals" in os.path.basename(stem).lower():
                        model_stems[model_key]["vocals"].append(stem)
                    elif "other" in os.path.basename(stem).lower():
                        model_stems[model_key]["other"].append(stem)
                
                # Clean up memory
                separator = None
                gc.collect()
                if torch.cuda.is_available():
                    torch.cuda.empty_cache()
                    logger.info(f"Cleared CUDA cache after {model_key} chunk {chunk_idx}")
        
        # Combine stems for each model
        for model_key, stems_dict in model_stems.items():
            for stem_type in ["vocals", "other"]:
                if stems_dict[stem_type]:
                    combined_path = os.path.join(output_dir, f"{base_name}_{stem_type}_{model_key.replace(' | ', '_').replace(' ', '_')}.wav")
                    combined_data = []
                    for stem_path in stems_dict[stem_type]:
                        data, _ = librosa.load(stem_path, sr=sr, mono=False)
                        combined_data.append(data)
                    combined_data = np.concatenate(combined_data, axis=-1) if combined_data[0].ndim == 2 else np.concatenate(combined_data)
                    sf.write(combined_path, combined_data.T if combined_data.ndim == 2 else combined_data, sr)
                    logger.info(f"Combined {stem_type} for {model_key}: {combined_path}")
                    if exclude_stems.strip() and stem_type.lower() in [s.strip().lower() for s in exclude_stems.split(',')]:
                        logger.info(f"Excluding {stem_type} for {model_key}")
                        continue
                    all_stems.append(combined_path)
        
        if not all_stems:
            raise ValueError("No valid stems for ensemble after exclusion.")
        
        # Ensemble the combined stems
        weights = [float(w.strip()) for w in weights_str.split(',')] if weights_str.strip() else [1.0] * len(all_stems)
        if len(weights) != len(all_stems):
            weights = [1.0] * len(all_stems)
        
        output_file = os.path.join(output_dir, f"{base_name}_ensemble_{ensemble_method}.{out_format}")
        ensemble_args = [
            "--files", *all_stems,
            "--type", ensemble_method,
            "--weights", *[str(w) for w in weights],
            "--output", output_file
        ]
        logger.info(f"Running ensemble with args: {ensemble_args}")
        ensemble_files(ensemble_args)
        
        logger.info("Ensemble complete")
        return output_file, f"Ensemble completed with {ensemble_method}, excluded: {exclude_stems if exclude_stems else 'None'}"
    except Exception as e:
        logger.error(f"Ensemble failed: {e}")
        raise RuntimeError(f"Ensemble failed: {e}")
    finally:
        # Clean up temporary files
        for path in chunk_paths + ([temp_audio_path] if temp_audio_path and os.path.exists(temp_audio_path) else []):
            try:
                if os.path.exists(path):
                    os.remove(path)
                    logger.info(f"Successfully cleaned up {path}")
            except Exception as e:
                logger.error(f"Failed to clean up {path}: {e}")
                
def update_roformer_models(category):
    """Update Roformer model dropdown based on selected category."""
    choices = list(ROFORMER_MODELS.get(category, {}).keys()) or []
    logger.debug(f"Updating roformer models for category {category}: {choices}")
    return gr.update(choices=choices, value=choices[0] if choices else None)

def update_ensemble_models(category):
    """Update ensemble model dropdown based on selected category."""
    choices = list(ROFORMER_MODELS.get(category, {}).keys()) or []
    logger.debug(f"Updating ensemble models for category {category}: {choices}")
    return gr.update(choices=choices, value=[])

def download_audio_wrapper(url, cookie_file):
    file_path, status, audio_data = download_audio(url, cookie_file)
    return audio_data, status  # Return audio_data for gr.Audio, status for gr.Textbox    

def create_interface():
    with gr.Blocks(title="🎡 SESA Fast Separation 🎡", css=CSS, elem_id="app-container") as app:
        gr.Markdown("<h1 class='header-text'>🎡 SESA Fast Separation 🎡</h1>")
        gr.Markdown("**Note**: If YouTube downloads fail, try uploading an audio file directly or use a valid cookies file.")
        
        with gr.Tabs():
            # Settings Tab
            with gr.Tab("βš™οΈ Settings"):
                with gr.Group(elem_classes="dubbing-theme"):
                    gr.Markdown("### General Settings")
                    model_file_dir = gr.Textbox(value="/tmp/audio-separator-models/", label="πŸ“‚ Model Cache", placeholder="Path to model directory", interactive=True)
                    output_dir = gr.Textbox(value="output", label="πŸ“€ Output Directory", placeholder="Where to save results", interactive=True)
                    output_format = gr.Dropdown(value="wav", choices=OUTPUT_FORMATS, label="🎢 Output Format", interactive=True)
                    norm_threshold = gr.Slider(0.1, 1.0, value=0.9, step=0.1, label="πŸ”Š Normalization Threshold", interactive=True)
                    amp_threshold = gr.Slider(0.1, 1.0, value=0.3, step=0.1, label="πŸ“ˆ Amplification Threshold", interactive=True)
                    batch_size = gr.Slider(1, 16, value=4, step=1, label="⚑ Batch Size", interactive=True)

            # Roformer Tab
            with gr.Tab("🎀 Roformer"):
                with gr.Group(elem_classes="dubbing-theme"):
                    gr.Markdown("### Audio Separation")
                    with gr.Row():
                        roformer_audio = gr.Audio(label="🎧 Upload Audio", type="filepath", interactive=True)
                        url_ro = gr.Textbox(label="πŸ”— Or Paste URL", placeholder="YouTube or audio URL", interactive=True)
                        cookies_ro = gr.File(label="πŸͺ Cookies File", file_types=[".txt"], interactive=True)
                        download_roformer = gr.Button("⬇️ Download", variant="secondary")
                    roformer_download_status = gr.Textbox(label="πŸ“’ Download Status", interactive=False)  # Added status output
                    roformer_exclude_stems = gr.Textbox(label="🚫 Exclude Stems", placeholder="e.g., vocals, drums (comma-separated)", interactive=True)
                    with gr.Row():
                        roformer_category = gr.Dropdown(label="πŸ“š Category", choices=list(ROFORMER_MODELS.keys()), value="General Purpose", interactive=True)
                        roformer_model = gr.Dropdown(label="πŸ› οΈ Model", choices=list(ROFORMER_MODELS["General Purpose"].keys()), interactive=True, allow_custom_value=True)
                    with gr.Row():
                        roformer_seg_size = gr.Slider(32, 4000, value=256, step=32, label="πŸ“ Segment Size", interactive=True)
                        roformer_overlap = gr.Slider(2, 10, value=8, step=1, label="πŸ”„ Overlap", interactive=True)
                    with gr.Row():
                        roformer_pitch_shift = gr.Slider(-12, 12, value=0, step=1, label="🎡 Pitch Shift", interactive=True)
                        roformer_override_seg_size = gr.Dropdown(choices=["True", "False"], value="False", label="πŸ”§ Override Segment Size", interactive=True)
                    roformer_button = gr.Button("βœ‚οΈ Separate Now!", variant="primary")
                    with gr.Row():
                        roformer_stem1 = gr.Audio(label="🎸 Stem 1", type="filepath", interactive=False)
                        roformer_stem2 = gr.Audio(label="πŸ₯ Stem 2", type="filepath", interactive=False)

            # Auto Ensemble Tab
            with gr.Tab("🎚️ Auto Ensemble"):
                with gr.Group(elem_classes="dubbing-theme"):
                    gr.Markdown("### Ensemble Processing")
                    with gr.Row():
                        ensemble_audio = gr.Audio(label="🎧 Upload Audio", type="filepath", interactive=True)
                        url_ensemble = gr.Textbox(label="πŸ”— Or Paste URL", placeholder="YouTube or audio URL", interactive=True)
                        cookies_ensemble = gr.File(label="πŸͺ Cookies File", file_types=[".txt"], interactive=True)
                        download_ensemble = gr.Button("⬇️ Download", variant="secondary")
                    ensemble_download_status = gr.Textbox(label="πŸ“’ Download Status", interactive=False)  # Added status output
                    ensemble_exclude_stems = gr.Textbox(label="🚫 Exclude Stems", placeholder="e.g., vocals, drums (comma-separated)", interactive=True)
                    with gr.Row():
                        ensemble_category = gr.Dropdown(label="πŸ“š Category", choices=list(ROFORMER_MODELS.keys()), value="Instrumentals", interactive=True)
                        ensemble_models = gr.Dropdown(label="πŸ› οΈ Models", choices=list(ROFORMER_MODELS["Instrumentals"].keys()), multiselect=True, interactive=True, allow_custom_value=True)
                    with gr.Row():
                        ensemble_seg_size = gr.Slider(32, 4000, value=256, step=32, label="πŸ“ Segment Size", interactive=True)
                        ensemble_overlap = gr.Slider(2, 10, value=8, step=1, label="πŸ”„ Overlap", interactive=True)
                        ensemble_use_tta = gr.Dropdown(choices=["True", "False"], value="False", label="πŸ” Use TTA", interactive=True)
                    ensemble_method = gr.Dropdown(label="βš™οΈ Ensemble Method", choices=['avg_wave', 'median_wave', 'max_wave', 'min_wave', 'avg_fft', 'median_fft', 'max_fft', 'min_fft'], value='avg_wave', interactive=True)
                    ensemble_weights = gr.Textbox(label="βš–οΈ Weights", placeholder="e.g., 1.0, 1.0 (comma-separated)", interactive=True)
                    ensemble_button = gr.Button("πŸŽ›οΈ Run Ensemble!", variant="primary")
                    ensemble_output = gr.Audio(label="🎢 Ensemble Result", type="filepath", interactive=False)
                    ensemble_status = gr.Textbox(label="πŸ“’ Status", interactive=False)

        gr.HTML("<div class='footer'>Powered by Audio-Separator 🌟🎢 | Made with ❀️</div>")

        # Event Handlers
        roformer_category.change(update_roformer_models, inputs=[roformer_category], outputs=[roformer_model])
        download_roformer.click(
            fn=download_audio_wrapper,
            inputs=[url_ro, cookies_ro],
            outputs=[roformer_audio, roformer_download_status]  # Updated to use wrapper and output status
        )
        roformer_button.click(
            fn=roformer_separator,
            inputs=[
                roformer_audio, roformer_model, roformer_seg_size, roformer_override_seg_size,
                roformer_overlap, roformer_pitch_shift, model_file_dir, output_dir,
                output_format, norm_threshold, amp_threshold, batch_size, roformer_exclude_stems
            ],
            outputs=[roformer_stem1, roformer_stem2]
        )
        ensemble_category.change(update_ensemble_models, inputs=[ensemble_category], outputs=[ensemble_models])
        download_ensemble.click(
            fn=download_audio_wrapper,
            inputs=[url_ensemble, cookies_ensemble],
            outputs=[ensemble_audio, ensemble_download_status]  # Updated to use wrapper and output status
        )
        ensemble_button.click(
            fn=auto_ensemble_process,
            inputs=[
                ensemble_audio, ensemble_models, ensemble_seg_size, ensemble_overlap,
                output_format, ensemble_use_tta, model_file_dir, output_dir,
                norm_threshold, amp_threshold, batch_size, ensemble_method,
                ensemble_exclude_stems, ensemble_weights
            ],
            outputs=[ensemble_output, ensemble_status]
        )
    
    return app

if __name__ == "__main__":
    parser = argparse.ArgumentParser(description="Music Source Separation Web UI")
    parser.add_argument("--port", type=int, default=7860, help="Port to run the UI on")
    args = parser.parse_args()

    app = create_interface()
    try:
        app.launch(server_name="0.0.0.0", server_port=args.port, share=True)
    except Exception as e:
        logger.error(f"Failed to launch app: {e}")
        raise
    finally:
        app.close()