File size: 25,353 Bytes
ea6ca86
80ec8c9
9abb457
 
 
67828d3
9abb457
c926b76
 
 
 
9abb457
ea6ca86
67828d3
 
 
 
 
 
 
f92e9c8
84eb043
9e73ac9
6cbb4e6
1ea1920
6cbb4e6
 
c7db8c6
1ea1920
6cbb4e6
1ea1920
9abb457
 
 
 
 
67828d3
 
 
 
 
80ec8c9
67828d3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
80ec8c9
67828d3
 
 
 
1d95413
 
 
67828d3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
80ec8c9
9e73ac9
 
 
 
 
 
ea6ca86
80ec8c9
2489a92
9e73ac9
 
f787e37
 
 
 
d7e5ee7
 
 
 
f787e37
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22e39eb
f787e37
 
 
 
 
 
 
 
22e39eb
 
f787e37
 
 
 
 
 
 
 
 
 
 
ed66f22
 
 
 
 
22e39eb
ed66f22
 
22e39eb
 
ed66f22
22e39eb
ed66f22
cea444d
ed66f22
f787e37
 
 
 
 
8e617aa
2489a92
 
80ec8c9
 
 
2489a92
80ec8c9
 
 
 
 
2489a92
 
 
 
 
 
80ec8c9
2489a92
 
 
 
 
80ec8c9
 
 
 
2489a92
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
80ec8c9
 
 
364042e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f787e37
 
364042e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f787e37
364042e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9abb457
 
 
 
f1b62d1
9abb457
 
 
e3e1895
9abb457
 
 
e3e1895
 
 
9abb457
e263c76
c404e15
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e3e1895
 
 
 
 
 
 
 
 
 
 
286d24d
 
e3e1895
 
 
 
 
 
 
 
286d24d
c404e15
 
 
e3e1895
 
 
 
c404e15
 
 
 
 
 
e3e1895
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c404e15
 
 
f54c21c
c926b76
f54c21c
 
c926b76
 
f54c21c
c926b76
f54c21c
 
 
 
 
c926b76
 
f54c21c
 
 
 
 
 
c926b76
f54c21c
 
 
 
 
c926b76
 
 
 
 
 
 
 
 
 
 
 
 
 
362178d
c926b76
 
 
 
 
f787e37
 
 
 
 
 
 
 
c926b76
f787e37
 
 
c926b76
f787e37
 
c926b76
 
 
 
 
362178d
 
 
 
 
 
 
 
 
c926b76
 
f54c21c
c926b76
95542db
 
c926b76
 
 
 
 
 
 
 
 
 
f787e37
 
 
c926b76
 
 
 
 
 
 
 
 
 
 
 
 
 
f787e37
 
c926b76
362178d
 
 
 
c926b76
362178d
c926b76
19dd021
 
 
e3e1895
 
 
 
7b1b82b
 
e6ba84f
 
 
921e6ff
 
 
e6ba84f
 
348b09b
921e6ff
 
 
e6ba84f
 
 
 
19dd021
 
 
9abb457
 
f1b62d1
e6ba84f
 
f1b62d1
e6ba84f
e3e1895
 
e6ba84f
 
 
 
 
f1b62d1
e6ba84f
80ec8c9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f1b62d1
9abb457
80ec8c9
9abb457
921e6ff
80ec8c9
 
e6ba84f
80ec8c9
e6ba84f
80ec8c9
 
 
f1b62d1
06fe876
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
084ae9e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c926b76
 
 
 
 
 
 
 
f787e37
 
 
 
 
 
 
 
 
 
c926b76
 
 
 
 
362178d
c926b76
 
 
 
 
 
362178d
c926b76
 
9e73ac9
46a312e
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
import gradio as gr
from sentence_transformers import SentenceTransformer, CrossEncoder
from supabase import create_client
import os
from dotenv import load_dotenv
from google import genai

import pandas as pd
import time
import math

load_dotenv()

GOOGLE_API_KEY = os.getenv("GEMINI_API")
if not GOOGLE_API_KEY:
    print("⚠️ Peringatan: GOOGLE_API_KEY tidak ditemukan, Gemini akan dinonaktifkan.")
    gemini_client = None
else:
    gemini_client = genai.Client(api_key=GOOGLE_API_KEY)
    
embedder = SentenceTransformer("sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2")
reranker = CrossEncoder("cross-encoder/ms-marco-MiniLM-L-6-v2") 

print("--- Daftar Model yang Tersedia ---")

if gemini_client:
    for m in gemini_client.models.list():
        print(f"Model: {m.name} | Name: {m.display_name}")

print("----------------------------------")

# === Supabase ===
supabase_url = os.getenv("SUPABASE_URL")
supabase_key = os.getenv("SUPABASE_KEY")
supabase = create_client(supabase_url, supabase_key)

def expand_query(query: str, num_variations: int = 3) -> str:
    """
    Memperluas query menggunakan Gemini 2.5 API resmi.
    - Otomatis fallback jika API error.
    - Cache hasil untuk menghemat pemanggilan API.
    """
    if not query.strip():
        return query

    if gemini_client is None:
        return f"Kegiatan usaha yang berkaitan dengan {query}"

    prompt = f"""
    Anda adalah ahli dalam sistem pencarian KBLI (Klasifikasi Baku Lapangan Usaha Indonesia) 2020. 
    Tugas Anda adalah membuat {num_variations} variasi dari kueri berikut
    untuk meningkatkan hasil pencarian.

    Kueri pengguna: "{query}"

    Buatkan {num_variations} variasi kueri yang:
    1. Menggunakan bahasa formal atau teknis (mis. istilah industri).
    2. Menggunakan bahasa sehari-hari.
    3. Mengandung kata kunci relevan lain.

    Format keluaran HARUS seperti ini:
    Variasi 1: ...
    Variasi 2: ...
    Variasi 3: ...
    """

    try:
        # Panggil Gemini
        response = gemini_client.models.generate_content(
            # model="gemini-2.5-flash",
            # # model="gemini-robotics-er-1.5-preview",
            model="gemini-3.1-flash-lite",
            contents=prompt,
        )

        text_output = response.text.strip()
        variations = []

        for line in text_output.splitlines():
            if line.lower().startswith("variasi"):
                parts = line.split(":", 1)
                if len(parts) > 1:
                    variations.append(parts[1].strip())

        if not variations:
            print("[Gemini Warning] Tidak ada variasi ditemukan. Gunakan fallback.")
            return f"Kegiatan usaha yang berkaitan dengan {query}"

        # Gabungkan hasil
        expanded = query + ". " + " ".join(variations)
        print(f"[Gemini Expand] {query} -> {expanded}")
        return expanded

    except Exception as e:
        print(f"[Gemini Error] {e}. Menggunakan fallback lokal.")
        return f"Kegiatan usaha yang berkaitan dengan {query}"

def get_embedding(text: str):
    """
    Menghasilkan embedding vector dari teks menggunakan model SentenceTransformer.
    """
    if not text:
        return []
    
    expanded_text = expand_query(text)
    embedding = embedder.encode(expanded_text, normalize_embeddings=True).tolist()
    return embedding

# ==========================================
# ABLATION STUDY
# ==========================================

# Helper Function
def apply_sigmoid(logit):
    return 1 / (1 + math.exp(-logit))

def bm25_only(query: str, match_count: int = 50):
    """Hanya Lexical / Full-Text Search (Tanpa Vector, Tanpa Gemini, Tanpa Reranker)"""
    # Catatan: Pastikan Anda sudah membuat RPC 'search_kbli_lexical' di Supabase
    response = supabase.rpc(
        "search_kbli_lexical",
        {"query_text": query, "match_count": match_count}
    ).execute()
    
    candidates = response.data or []
    return {"results": candidates[:10]}

def dense_only(query: str, match_count: int = 50):
    """Hanya Semantic Vector (Tanpa BM25, Tanpa Gemini, Tanpa Reranker)"""
    # Gunakan query asli (tanpa expand)
    embedding_q = embedder.encode(query, normalize_embeddings=True).tolist()

    response = supabase.rpc(
        "search_kbli",
        {"query_embedding": embedding_q, "match_count": match_count}
    ).execute()
    
    candidates = response.data or []
    return {"results": candidates[:10]}

def semantic_no_gemini(query: str, match_count: int = 50):
    """Semantic Vector + Reranker (TANPA Gemini Query Expansion)"""
    expanded = query # Bypass Gemini
    embedding_q = embedder.encode(expanded, normalize_embeddings=True).tolist()

    response = supabase.rpc(
        "search_kbli",
        {"query_embedding": embedding_q, "match_count": match_count}
    ).execute()
    candidates = response.data or []

    if not candidates:
        return {"results": []}

    pairs = [(expanded, c["judul"] + " " + c["deskripsi"]) for c in candidates]

    try:
        scores = reranker.predict(pairs)
    except Exception as e:
        print("Reranker error:", e)
        return {"results": sorted(candidates, key=lambda x: x.get("similarity", 0), reverse=True)[:10]}

    rerank_vals = [float(s) for s in scores]
    rmin, rmax = min(rerank_vals), max(rerank_vals)
    
    for c, s in zip(candidates, rerank_vals):
        c["rerank_score"] = s
        if rmax - rmin > 1e-9:
            c["rerank_norm"] = (s - rmin) / (rmax - rmin)
        else:
            c["rerank_norm"] = 0.0

        sim = c.get("similarity", 0.0)
        c["hybrid_score"] = 0.6 * sim + 0.4 * c["rerank_norm"]

    candidates = sorted(candidates, key=lambda x: x["hybrid_score"], reverse=True)
    return {"results": candidates[:10]}

def hybrid_search_no_gemini(query: str, match_count: int = 50):
    """Hybrid (BM25 + Dense) + Reranker (TANPA Gemini Query Expansion)"""
    expanded = query 
    embedding_q = embedder.encode(expanded, normalize_embeddings=True).tolist()

    response = supabase.rpc(
        "search_kbli_hybrid",
        {
            "query_text": query,          
            "query_embedding": embedding_q, 
            "match_count": match_count,
            "lexical_weight": 0.4, # Diabaikan di SQL, tapi wajib dikirim
            "dense_weight": 0.6    # Diabaikan di SQL, tapi wajib dikirim
        }
    ).execute()
    
    candidates = response.data or []

    if not candidates:
        return {"results": []}

    pairs = [(query, c["judul"] + " " + c["deskripsi"]) for c in candidates]

    try:
        scores = reranker.predict(pairs)
    except Exception as e:
        print("Reranker error:", e)
        return {"results": sorted(candidates, key=lambda x: x.get("similarity", 0), reverse=True)[:10]}

    # Reranker sebagai Hakim Tunggal
    for c, s in zip(candidates, scores):
        c["rerank_score"] = float(s)
        # Gunakan Sigmoid agar nilainya menjadi probabilitas pasti
        c["final_score"] = apply_sigmoid(float(s))

    # Urutkan berdasarkan keputusan mutlak dari Reranker
    candidates = sorted(candidates, key=lambda x: x["final_score"], reverse=True)
    
    return {"results": candidates[:10]}

# ==========================================
# CORE APPS
# ==========================================

def fn_semantic(query: str, match_count: int = 50):
    expanded = expand_query(query)
    embedding_q = embedder.encode(expanded, normalize_embeddings=True).tolist()

    response = supabase.rpc(
        "search_kbli",
        {"query_embedding": embedding_q, "match_count": match_count}
    ).execute()
    candidates = response.data or []

    if not candidates:
        return {"results": []}
    
    print("=== Candidates BEFORE rerank (top 10) ===")
    for c in candidates[:10]:
        print(c.get("kode"), c.get("judul")[:80], "sim=", c.get("similarity"))

    pairs = [(expanded, c["judul"] + " " + c["deskripsi"]) for c in candidates]

    try:
        scores = reranker.predict(pairs)
    except Exception as e:
        print("Reranker error:", e)
        return {"results": sorted(candidates, key=lambda x: x.get("similarity", 0), reverse=True)[:10]}

    for c, s in zip(candidates, scores):
        c["rerank_score"] = float(s)

    print("=== Candidates AFTER rerank (top 10) ===")
    for c in candidates[:10]:
        print(c.get("kode"), c.get("judul")[:80], "sim=", c.get("similarity"), "rerank=", c.get("rerank_score"))

    rerank_vals = [c["rerank_score"] for c in candidates]
    rmin, rmax = min(rerank_vals), max(rerank_vals)
    for c in candidates:
        if rmax - rmin > 1e-9:
            c["rerank_norm"] = (c["rerank_score"] - rmin) / (rmax - rmin)
        else:
            c["rerank_norm"] = 0.0

    for c in candidates:
        sim = c.get("similarity", 0.0)
        c["hybrid_score"] = 0.6 * sim + 0.4 * c["rerank_norm"]

    candidates = sorted(candidates, key=lambda x: x["hybrid_score"], reverse=True)

    return {"results": candidates[:10]}

def hybrid_search(query: str, match_count: int = 50):
    # 1. Query Expansion
    expanded = expand_query(query)
    
    # 2. Embedding
    # Kita encode query yang sudah di-expand untuk pencarian dense
    embedding_q = embedder.encode(expanded, normalize_embeddings=True).tolist()

    # 3. Panggil Hybrid Search di Supabase
    # Kita kirimkan query ASLI untuk Lexical (agar tidak terlalu banyak noise kata), 
    # dan query EXPANDED untuk Dense (embedding).
    response = supabase.rpc(
        "search_kbli_hybrid",
        {
            "query_text": query,          # Untuk Lexical Match (tsvector)
            "query_embedding": embedding_q, # Untuk Dense Match (pgvector)
            "match_count": match_count,
            "lexical_weight": 0.2,        # Bobot Lexical (bisa disesuaikan untuk Ablation Study)
            "dense_weight": 0.8           # Bobot Dense
        }
    ).execute()
    
    candidates = response.data or []

    if not candidates:
        return {"results": []}
    
    print("=== Candidates dari Hybrid DB BEFORE rerank (top 10) ===")
    for c in candidates[:10]:
        # Tampilkan similarity yang sekarang merupakan gabungan Lexical & Dense
        print(c.get("kode"), c.get("judul")[:80], "hybrid_db_sim=", c.get("similarity"))

    # 4. Reranking dengan Cross-Encoder
    # Evaluasi kecocokan antara query asli dengan dokumen kandidat
    pairs = [(expanded, c["judul"] + " " + c["deskripsi"]) for c in candidates]

    try:
        scores = reranker.predict(pairs)
    except Exception as e:
        print("Reranker error:", e)
        return {"results": sorted(candidates, key=lambda x: x.get("similarity", 0), reverse=True)[:10]}

    # 5. Normalisasi skor Reranker & Kalkulasi Final Score
    rerank_vals = [float(s) for s in scores]
    rmin, rmax = min(rerank_vals), max(rerank_vals)
    
    for c, s in zip(candidates, rerank_vals):
        c["rerank_score"] = s
        
        # Normalisasi Min-Max
        if rmax - rmin > 1e-9:
            c["rerank_norm"] = (s - rmin) / (rmax - rmin)
        else:
            c["rerank_norm"] = 0.0

        # Skor Final Sistem Neural IR Anda (gabungan Stage 1: Hybrid Retrieval + Stage 2: Reranking)
        # Anda bisa menyesuaikan bobot ini nanti
        db_hybrid_sim = c.get("similarity", 0.0)
        c["final_score"] = (0.5 * db_hybrid_sim) + (0.5 * c["rerank_norm"])

    print("=== Candidates AFTER Cross-Encoder rerank (top 10) ===")
    # Urutkan berdasarkan final_score
    candidates = sorted(candidates, key=lambda x: x["final_score"], reverse=True)
    
    for c in candidates[:10]:
        print(c.get("kode"), c.get("judul")[:80], "final_score=", c.get("final_score"), "rerank=", c.get("rerank_score"))

    # Kembalikan 10 teratas (sesuai logika asli Anda)
    return {"results": candidates[:10]}

def search_kbli(text: str):
    if not text:
        return {"embedding": [], "results": []}

    embedding = get_embedding(text)

    response = supabase.rpc(
        "search_kbli",
        {"query_embedding": embedding, "match_count": 25}
    ).execute()

    results = response.data if response.data else []
    
    if not results:
        return "<p>Tidak ditemukan hasil.</p>"

    html = """
    <style>
        .kbli-item {
            border: 1px solid #ddd;
            border-radius: 8px;
            padding: 10px;
            margin-bottom: 8px;
            transition: background 0.2s ease;
        }
        .kbli-item:hover {
            background: #f9fafb;
        }
        .kbli-title {
            font-weight: 600;
            margin: 0;
        }
        .kbli-desc {
            font-size: 13px;
            color: #4b5563;
            margin-top: 4px;
        }
        details {
            margin-top: 16px;
            border: 1px solid #ddd;
            border-radius: 6px;
            padding: 8px;
        }
        details summary {
            cursor: pointer;
            font-weight: 600;
            color: #2563eb;
        }
        
        @media (prefers-color-scheme: dark) {
            .kbli-item { border: 1px solid #374151; }
            .kbli-item:hover { background: #1f2937; }
            .kbli-title { color: #f3f4f6; }
            .kbli-desc { color: #d1d5db; }
            .kbli-item:hover .kbli-title { color: #93c5fd; }
            .kbli-item:hover .kbli-desc { color: #e5e7eb; }
            details { border: 1px solid #374151; }
            details summary { color: #60a5fa; }
        }
    </style>
    <div>
    """

    # Top 10 == // for r in results
    top_10 = results[:10]
    for r in top_10:
        html += f"""
        <div class="kbli-item">
            <p class="kbli-title">{r['kode']}{r['judul']}</p>
            <p class="kbli-desc">{r['deskripsi']}</p>
        </div>
        """

    # Expandable for
    others = results[10:]
    if others:
        html += "<details><summary>Lihat hasil lainnya</summary><div style='margin-top:10px;'>"
        for r in others:
            html += f"""
            <div class="kbli-item">
                <p class="kbli-title">{r['kode']}{r['judul']}</p>
                <p class="kbli-desc">{r['deskripsi']}</p>
            </div>
            """
        html += "</div></details>"
    # End

    html += "</div>"
    return html

def calculate_mrr(retrieved_kodes, relevant_kodes_set):
    for i, kode in enumerate(retrieved_kodes):
        if kode in relevant_kodes_set:
            return 1.0 / (i + 1)
    return 0.0

def calculate_recall(retrieved_kodes, relevant_kodes_set, k=10):
    retrieved_k_set = set(retrieved_kodes[:k])

    if not relevant_kodes_set:
        return 0.0

    return len(retrieved_k_set & relevant_kodes_set) / len(relevant_kodes_set)

def calculate_ndcg(retrieved_kodes, relevance_dict, k=10):

    dcg = 0
    for i, kode in enumerate(retrieved_kodes[:k]):
        rel = relevance_dict.get(kode, 0)
        dcg += rel / math.log2(i + 2)

    ideal_rels = sorted(relevance_dict.values(), reverse=True)[:k]

    idcg = 0
    for i, rel in enumerate(ideal_rels):
        idcg += rel / math.log2(i + 2)

    return dcg / idcg if idcg > 0 else 0.0

def run_evaluation(file_obj, scenario):
    if file_obj is None:
        return "Peringatan: Silakan unggah file ground_truth.csv terlebih dahulu.", None, None
    
    df = pd.read_csv(file_obj.name)
    queries = df.groupby('query_id').first()['query'].to_dict()
    
    ground_truth = {}
    for q_id, group in df.groupby('query_id'):
        ground_truth[q_id] = dict(zip(group['kode_kbli'].astype(str), group['relevance']))

    results_list = []
    retrieval_rows = []
    
    for q_id, query_text in queries.items():
        start_time = time.perf_counter() # Mulai hitung latensi
        
        # Eksekusi fungsi berdasarkan skenario yang dipilih

        if scenario == "BM25 Only (Lexical)":
            response = bm25_only(query_text, match_count=50)
        elif scenario == "Dense Only (Semantic)":
            response = dense_only(query_text, match_count=50)
        elif scenario == "Semantic + Reranker (No Gemini)":
            response = semantic_no_gemini(query_text, match_count=50)
        elif scenario == "Semantic + Reranker (With Gemini)":
            response = fn_semantic(query_text, match_count=50)
        elif scenario == "Hybrid + Reranker (No Gemini)":
            response = hybrid_search_no_gemini(query_text, match_count=50)
        elif scenario == "Hybrid + Reranker (With Gemini)":
            response = hybrid_search(query_text, match_count=50)
        else:
            response = {"results": []}
            
        latency = time.perf_counter() - start_time # Hitung selisih waktu
        
        candidates = response.get("results", [])
        retrieved_kodes = [str(r.get('kode')) for r in candidates]

        for rank, kode in enumerate(retrieved_kodes, start=1):
            retrieval_rows.append({
                "query_id": q_id,
                "query": query_text,
                "scenario": scenario,
                "rank": rank,
                "kode_kbli": kode
            })
        
        rel_dict = ground_truth.get(q_id, {})
        relevant_kodes_set = {k for k, r in rel_dict.items() if r > 0}
        
        mrr = calculate_mrr(retrieved_kodes, relevant_kodes_set)
        recall = calculate_recall(retrieved_kodes, relevant_kodes_set, k=10)
        ndcg = calculate_ndcg(retrieved_kodes, rel_dict, k=10)
        
        results_list.append({
            "Query ID": q_id,
            "Query Text": query_text,
            "MRR@10": round(mrr, 4),
            "Recall@10": round(recall, 4),
            "nDCG@10": round(ndcg, 4),
            "Latency (sec)": round(latency, 4) # Menyimpan data latensi per kueri
        })

        if "With Gemini" in scenario:
            time.sleep(1) # Hindari rate limit Gemini API

    results_df = pd.DataFrame(results_list)
    
    # Hitung rata-rata
    summary = {
        "Skenario": scenario,
        "Total Query": len(queries),
        "Avg MRR@10": round(results_df["MRR@10"].mean(), 4),
        "Avg Recall@10": round(results_df["Recall@10"].mean(), 4),
        "Avg nDCG@10": round(results_df["nDCG@10"].mean(), 4),
        "Avg Latency (sec)": round(results_df["Latency (sec)"].mean(), 4)
    }
    
    # Export ke Excel
    safe_scenario_name = scenario.replace(" ", "_").replace("(", "").replace(")", "").replace("+", "plus")
    output_filename = f"Evaluasi_{safe_scenario_name}.xlsx"
    results_df.to_excel(output_filename, index=False)

    retrieval_df = pd.DataFrame(retrieval_rows)
    retrieval_filename = f"retrieval_results_{safe_scenario_name}.csv"
    retrieval_df.to_csv(retrieval_filename, index=False)
    
    return summary, results_df, output_filename, retrieval_filename

with gr.Blocks(css="""
    .title {font-size: 22px; font-weight: 700; color: #111827; margin-bottom: 4px;}
    .desc {font-size: 14px; color: #6b7280; margin-bottom: 16px;}
    button.gr-button {
        border-radius: 6px;
    }
    button.gr-button-primary, button.gr-button-secondary {
        border-radius: 6px;
    }
    
    .btn-row {display: flex; gap: 8px;}
    .btn-row > * {flex: 1;}
    
    .btn-row-search {display: flex; gap: 8px;}
    .btn-row-search > * {flex: none;}

    @media (max-width: 640px) {
        
        .btn-row, .btn-row-search {flex-direction: column-reverse;}
        .btn-row > button,
        .btn-row-search > button {
            width: 100% !important;
            flex: none;
        }
    }
""") as demo:
    gr.Markdown("<div class='title'>Semantic KBLI Search</div>")
    gr.Markdown("<div class='desc'>Cari kode KBLI dengan semantic search (Embedding + Matching)</div>")

    with gr.Tab("Embedding Only"):
        with gr.Row():
            with gr.Column(scale=1):
                inp1 = gr.Textbox(label="Masukkan teks")

                with gr.Row(elem_classes="btn-row"):
                    btn_clear1 = gr.Button("Clear", variant="secondary")
                    btn_submit1 = gr.Button("Submit", variant="primary")

            with gr.Column(scale=1):
                out1 = gr.JSON(label="Embedding Vector")
        
        inp1.submit(get_embedding, inp1, out1, api_name="get_embedding")
        btn_clear1.click(lambda: ("", None), None, [inp1, out1])
        btn_submit1.click(get_embedding, inp1, out1, api_name="get_embedding")
    
    with gr.Tab("Embedding Fine-tuned"):
        with gr.Row():
            with gr.Column(scale=1):
                inp2 = gr.Textbox(label="Masukkan teks")

                with gr.Row(elem_classes="btn-row"):
                    btn_clear2 = gr.Button("Clear", variant="secondary")
                    btn_submit2 = gr.Button("Submit", variant="primary")

            with gr.Column(scale=1):
                out2 = gr.JSON(label="Embedding Vector")
        
        inp2.submit(fn_semantic, inp2, out2, api_name="fn_semantic")
        btn_clear2.click(lambda: ("", None), None, [inp2, out2])
        btn_submit2.click(fn_semantic, inp2, out2, api_name="fn_semantic")

    with gr.Tab("Search KBLI"):
        inp3 = gr.Textbox(label="Masukkan teks")

        with gr.Row(elem_classes="btn-row-search"):
            btn_clear3 = gr.Button("Clear", variant="secondary")
            btn_submit3 = gr.Button("Submit", variant="primary")
        
        out3 = gr.HTML(label="Hasil Pencarian Semantic")
        
        inp3.submit(search_kbli, inp3, out3, api_name="search_kbli")
        btn_clear3.click(lambda: ("", None), None, [inp3, out3])
        btn_submit3.click(search_kbli, inp3, out3, api_name="search_kbli")

    with gr.Tab("Hybrid Search (Final)"):
        with gr.Row():
            with gr.Column(scale=1):
                inp4 = gr.Textbox(label="Masukkan teks")

                with gr.Row(elem_classes="btn-row"):
                    btn_clear4 = gr.Button("Clear", variant="secondary")
                    btn_submit4 = gr.Button("Submit", variant="primary")

            with gr.Column(scale=1):
                out4 = gr.JSON(label="Hasil Hybrid Search")
        
        inp4.submit(hybrid_search, inp4, out4, api_name="hybrid_search")
        btn_clear4.click(lambda: ("", None), None, [inp4, out4])
        btn_submit4.click(hybrid_search, inp4, out4, api_name="hybrid_search")

    with gr.Tab("Ablation Endpoints (API)"):
        gr.Markdown("### Individual Model")
        
        with gr.Row():
            with gr.Column(scale=1):
                inp5 = gr.Textbox(label="Masukkan kueri teks")

                with gr.Row(elem_classes="btn-row"):
                    btn_bm25 = gr.Button("BM25 Only", variant="primary")
                    btn_dense = gr.Button("Dense Only", variant="primary")
                
                with gr.Row(elem_classes="btn-row"):
                    btn_sem_no_gem = gr.Button("Semantic (No Gemini)", variant="primary")
                    btn_hyb_no_gem = gr.Button("Hybrid (No Gemini)", variant="primary")

                with gr.Row():
                    btn_clear5 = gr.Button("Clear", variant="secondary")

            with gr.Column(scale=1):
                out5 = gr.JSON(label="Hasil Pencarian Ablation")
        
        # Clear Button
        btn_clear5.click(lambda: ("", None), None, [inp5, out5])
        
        # Registrasi Event dan API Name
        btn_bm25.click(bm25_only, inputs=[inp5], outputs=[out5], api_name="bm25_only")
        btn_dense.click(dense_only, inputs=[inp5], outputs=[out5], api_name="dense_only")
        btn_sem_no_gem.click(semantic_no_gemini, inputs=[inp5], outputs=[out5], api_name="semantic_no_gemini")
        btn_hyb_no_gem.click(hybrid_search_no_gemini, inputs=[inp5], outputs=[out5], api_name="hybrid_search_no_gemini")
        
        inp5.submit(dense_only, inputs=[inp5], outputs=[out5])

    with gr.Tab("Ablation Study"):
        gr.Markdown("### Metrics & Latency")
        gr.Markdown("Unggah file `ground_truth.csv` Anda untuk menjalankan *batch testing* dan membandingkan skenario.")
        
        with gr.Row():
            with gr.Column(scale=1):
                eval_file = gr.File(label="Upload ground_truth.csv", file_types=[".csv"])
                eval_scenario = gr.Dropdown(
                    choices=[
                            "BM25 Only (Lexical)",
                            "Dense Only (Semantic)",
                            "Semantic + Reranker (No Gemini)",
                            "Semantic + Reranker (With Gemini)",
                            "Hybrid + Reranker (No Gemini)",
                            "Hybrid + Reranker (With Gemini)"
                        ],
                        label="Pilih Skenario Evaluasi"
                    )
                btn_run_eval = gr.Button("Jalankan Evaluasi Otomatis", variant="primary")
            
            with gr.Column(scale=1):
                eval_summary = gr.JSON(label="Ringkasan Skor Rata-rata & Latensi")
                eval_download = gr.File(label="Download Laporan (Excel)")
                eval_retrieval_download = gr.File(label="Download Retrieval Results (CSV)")
                
        eval_table = gr.Dataframe(label="Detail Per-Kueri")
        
        btn_run_eval.click(
            run_evaluation, 
            inputs=[eval_file, eval_scenario], 
            outputs=[eval_summary, eval_table, eval_download, eval_retrieval_download]
        )

if __name__ == "__main__":
    demo.queue().launch(show_error=True)