galihboy commited on
Commit
c5ff0a8
Β·
verified Β·
1 Parent(s): a1aa315

Update src/streamlit_app.py

Browse files

import os

os.environ["TOKENIZERS_PARALLELISM"] = "false"

# Sekarang baru import streamlit
import streamlit as st
import faiss
import pickle
from sentence_transformers import SentenceTransformer
import pandas as pd

# Konfigurasi
MODEL_NAME = "Qwen/Qwen3-Embedding-0.6B"

# Get absolute path for data directory (independent from maintenance_web)
SCRIPT_DIR = os.path.dirname(os.path.abspath(__file__))
INDEX_DIR = os.path.join(SCRIPT_DIR, "data")



@st
.cache_resource(show_spinner=True)
def load_model():
"""Load embedding model"""
# Model akan di-cache otomatis
model = SentenceTransformer(MODEL_NAME)
return model



@st
.cache_resource
def load_index():
"""Load FAISS index and metadata"""
index_path = os.path.join(INDEX_DIR, "skripsi.faiss")
metadata_path = os.path.join(INDEX_DIR, "metadata.pkl")

if not os.path.exists(index_path):
st.error(f"Index not found: {index_path}")
return None, None

if not os.path.exists(metadata_path):
st.error(f"Metadata not found: {metadata_path}")
return None, None

index = faiss.read_index(index_path)

with open(metadata_path, 'rb') as f:
metadata = pickle.load(f)

return index, metadata

def search(query, model, index, metadata, top_k=10):
"""Perform semantic search"""
# Generate query embedding
query_embedding = model.encode([query])

# Search
distances, indices = index.search(query_embedding, top_k)

# Get data list from metadata
data_list = metadata.get('data', [])

# Format results
results = []
for i, (dist, idx) in enumerate(zip(distances[0], indices[0])):
if idx < len(data_list):
meta = data_list[idx]
# Combine pembimbing info
pembimbing = meta.get('nama_pembimbing', 'N/A')
gelar_depan = meta.get('gelar_depan_pembimbing', '')
gelar_belakang = meta.get('gelar_belakang_pembimbing', '')
if gelar_depan or gelar_belakang:
pembimbing = f"{gelar_depan} {pembimbing}, {gelar_belakang}".strip(', ')

results.append({
'Rank': i + 1,
'Score': f"{dist:.4f}",
'Judul': meta.get('judul', 'N/A'),
'NIM': meta.get('nim', 'N/A'),
'Nama': meta.get('nama', 'N/A'),
'Pembimbing': pembimbing,
'Tahun': meta.get('tahun', 'N/A'),
'Semester': meta.get('semester', 'N/A')
})

return results

# Streamlit UI
st.set_page_config(page_title="Semantic Search - Skripsi UNIKOM", layout="wide")

st.title("πŸ” Semantic Search - Database Skripsi Prodi Teknik Informatika UNIKOM")
st.markdown("*Pencarian semantik berdasarkan kemiripan makna judul skripsi*")
st.markdown("---")

# Sidebar
with st.sidebar:
st.header("βš™οΈ Settings")
top_k = st.slider("Number of results", min_value=5, max_value=50, value=10, step=5)

st.markdown("---")
st.markdown("### πŸ“Š Model Info")
st.info(f"""
**Model**: {MODEL_NAME}
**Index**: {INDEX_DIR}
""")

# Load resources
try:
model = load_model()
index, metadata = load_index()

if index is None or metadata is None:
st.error("Failed to load index or metadata")
st.stop()

st.success(f"βœ… Model loaded | Index: {index.ntotal} vectors | Dimension: {index.d}")

except Exception as e:
st.error(f"Error loading resources: {e}")
st.stop()

# Search interface
st.markdown("### πŸ’¬ Enter your search query")
query = st.text_input("Search Query", placeholder="e.g., machine learning, web application, sistem informasi...", label_visibility="collapsed")

if st.button("πŸ” Search", type="primary") or query:
if query.strip():
with st.spinner("Searching..."):
results = search(query, model, index, metadata, top_k)

st.markdown(f"### πŸ“‹ Found {len(results)} results")

# Display as dataframe
if results:
df = pd.DataFrame(results)
st.dataframe(df, use_container_width=True, hide_index=True)

# Detailed view
st.markdown("---")
st.markdown("### πŸ“– Detailed Results")
for result in results:
with st.expander(f"#{result['Rank']} - {result['Judul'][:100]}... (Score: {result['Score']})"):
col1, col2 = st.columns(2)
with col1:
st.markdown(f"**NIM**: {result['NIM']}")
st.markdown(f"**Nama**: {result['Nama']}")
st.markdown(f"**Pembimbing**: {result['Pembimbing']}")
with col2:
st.markdown(f"**Tahun**: {result['Tahun']}")
st.markdown(f"**Semester**: {result['Semester']}")
st.markdown(f"**Judul Lengkap**: {result['Judul']}")
else:
st.warning("No results found")
else:
st.warning("Please enter a search query")

# Footer
st.markdown("---")
st.markdown("""
<div style='text-align: center;'>
<p><a href='https://galih.eu'>Galih Hermawan</a> | Akabot Research Group</p>
<p>Prodi Teknik Informatika | Universitas Komputer Indonesia</p>
<p>Powered by Qwen3 Embedding Model</p>
</div>
""", unsafe_allow_html=True)

Files changed (1) hide show
  1. src/streamlit_app.py +168 -38
src/streamlit_app.py CHANGED
@@ -1,40 +1,170 @@
1
- import altair as alt
2
- import numpy as np
3
- import pandas as pd
 
 
 
 
 
 
 
 
 
 
 
 
4
  import streamlit as st
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
 
6
- """
7
- # Welcome to Streamlit!
8
-
9
- Edit `/streamlit_app.py` to customize this app to your heart's desire :heart:.
10
- If you have any questions, checkout our [documentation](https://docs.streamlit.io) and [community
11
- forums](https://discuss.streamlit.io).
12
-
13
- In the meantime, below is an example of what you can do with just a few lines of code:
14
- """
15
-
16
- num_points = st.slider("Number of points in spiral", 1, 10000, 1100)
17
- num_turns = st.slider("Number of turns in spiral", 1, 300, 31)
18
-
19
- indices = np.linspace(0, 1, num_points)
20
- theta = 2 * np.pi * num_turns * indices
21
- radius = indices
22
-
23
- x = radius * np.cos(theta)
24
- y = radius * np.sin(theta)
25
-
26
- df = pd.DataFrame({
27
- "x": x,
28
- "y": y,
29
- "idx": indices,
30
- "rand": np.random.randn(num_points),
31
- })
32
-
33
- st.altair_chart(alt.Chart(df, height=700, width=700)
34
- .mark_point(filled=True)
35
- .encode(
36
- x=alt.X("x", axis=None),
37
- y=alt.Y("y", axis=None),
38
- color=alt.Color("idx", legend=None, scale=alt.Scale()),
39
- size=alt.Size("rand", legend=None, scale=alt.Scale(range=[1, 150])),
40
- ))
 
1
+ import os
2
+ from pathlib import Path
3
+
4
+ # FIX: Set cache ke /tmp (writable di HF Spaces)
5
+ os.environ["HOME"] = "/tmp"
6
+ os.environ["HF_HOME"] = "/tmp/.cache/huggingface"
7
+ os.environ["TRANSFORMERS_CACHE"] = "/tmp/.cache/huggingface/transformers"
8
+ os.environ["SENTENCE_TRANSFORMERS_HOME"] = "/tmp/.cache/sentence-transformers"
9
+ os.environ["TOKENIZERS_PARALLELISM"] = "false"
10
+
11
+ # Buat folder cache
12
+ Path("/tmp/.cache/huggingface").mkdir(parents=True, exist_ok=True)
13
+ Path("/tmp/.cache/sentence-transformers").mkdir(parents=True, exist_ok=True)
14
+
15
+ # Sekarang baru import streamlit
16
  import streamlit as st
17
+ import faiss
18
+ import pickle
19
+ from sentence_transformers import SentenceTransformer
20
+ import pandas as pd
21
+
22
+ # Konfigurasi
23
+ MODEL_NAME = "Qwen/Qwen3-Embedding-0.6B"
24
+
25
+ # Get absolute path for data directory (independent from maintenance_web)
26
+ SCRIPT_DIR = os.path.dirname(os.path.abspath(__file__))
27
+ INDEX_DIR = os.path.join(SCRIPT_DIR, "data")
28
+
29
+ @st.cache_resource(show_spinner=True)
30
+ def load_model():
31
+ """Load embedding model"""
32
+ # Model akan di-cache otomatis
33
+ model = SentenceTransformer(MODEL_NAME)
34
+ return model
35
+
36
+ @st.cache_resource
37
+ def load_index():
38
+ """Load FAISS index and metadata"""
39
+ index_path = os.path.join(INDEX_DIR, "skripsi.faiss")
40
+ metadata_path = os.path.join(INDEX_DIR, "metadata.pkl")
41
+
42
+ if not os.path.exists(index_path):
43
+ st.error(f"Index not found: {index_path}")
44
+ return None, None
45
+
46
+ if not os.path.exists(metadata_path):
47
+ st.error(f"Metadata not found: {metadata_path}")
48
+ return None, None
49
+
50
+ index = faiss.read_index(index_path)
51
+
52
+ with open(metadata_path, 'rb') as f:
53
+ metadata = pickle.load(f)
54
+
55
+ return index, metadata
56
+
57
+ def search(query, model, index, metadata, top_k=10):
58
+ """Perform semantic search"""
59
+ # Generate query embedding
60
+ query_embedding = model.encode([query])
61
+
62
+ # Search
63
+ distances, indices = index.search(query_embedding, top_k)
64
+
65
+ # Get data list from metadata
66
+ data_list = metadata.get('data', [])
67
+
68
+ # Format results
69
+ results = []
70
+ for i, (dist, idx) in enumerate(zip(distances[0], indices[0])):
71
+ if idx < len(data_list):
72
+ meta = data_list[idx]
73
+ # Combine pembimbing info
74
+ pembimbing = meta.get('nama_pembimbing', 'N/A')
75
+ gelar_depan = meta.get('gelar_depan_pembimbing', '')
76
+ gelar_belakang = meta.get('gelar_belakang_pembimbing', '')
77
+ if gelar_depan or gelar_belakang:
78
+ pembimbing = f"{gelar_depan} {pembimbing}, {gelar_belakang}".strip(', ')
79
+
80
+ results.append({
81
+ 'Rank': i + 1,
82
+ 'Score': f"{dist:.4f}",
83
+ 'Judul': meta.get('judul', 'N/A'),
84
+ 'NIM': meta.get('nim', 'N/A'),
85
+ 'Nama': meta.get('nama', 'N/A'),
86
+ 'Pembimbing': pembimbing,
87
+ 'Tahun': meta.get('tahun', 'N/A'),
88
+ 'Semester': meta.get('semester', 'N/A')
89
+ })
90
+
91
+ return results
92
+
93
+ # Streamlit UI
94
+ st.set_page_config(page_title="Semantic Search - Skripsi UNIKOM", layout="wide")
95
+
96
+ st.title("πŸ” Semantic Search - Database Skripsi Prodi Teknik Informatika UNIKOM")
97
+ st.markdown("*Pencarian semantik berdasarkan kemiripan makna judul skripsi*")
98
+ st.markdown("---")
99
+
100
+ # Sidebar
101
+ with st.sidebar:
102
+ st.header("βš™οΈ Settings")
103
+ top_k = st.slider("Number of results", min_value=5, max_value=50, value=10, step=5)
104
+
105
+ st.markdown("---")
106
+ st.markdown("### πŸ“Š Model Info")
107
+ st.info(f"""
108
+ **Model**: {MODEL_NAME}
109
+ **Index**: {INDEX_DIR}
110
+ """)
111
+
112
+ # Load resources
113
+ try:
114
+ model = load_model()
115
+ index, metadata = load_index()
116
+
117
+ if index is None or metadata is None:
118
+ st.error("Failed to load index or metadata")
119
+ st.stop()
120
+
121
+ st.success(f"βœ… Model loaded | Index: {index.ntotal} vectors | Dimension: {index.d}")
122
+
123
+ except Exception as e:
124
+ st.error(f"Error loading resources: {e}")
125
+ st.stop()
126
+
127
+ # Search interface
128
+ st.markdown("### πŸ’¬ Enter your search query")
129
+ query = st.text_input("Search Query", placeholder="e.g., machine learning, web application, sistem informasi...", label_visibility="collapsed")
130
+
131
+ if st.button("πŸ” Search", type="primary") or query:
132
+ if query.strip():
133
+ with st.spinner("Searching..."):
134
+ results = search(query, model, index, metadata, top_k)
135
+
136
+ st.markdown(f"### πŸ“‹ Found {len(results)} results")
137
+
138
+ # Display as dataframe
139
+ if results:
140
+ df = pd.DataFrame(results)
141
+ st.dataframe(df, use_container_width=True, hide_index=True)
142
+
143
+ # Detailed view
144
+ st.markdown("---")
145
+ st.markdown("### πŸ“– Detailed Results")
146
+ for result in results:
147
+ with st.expander(f"#{result['Rank']} - {result['Judul'][:100]}... (Score: {result['Score']})"):
148
+ col1, col2 = st.columns(2)
149
+ with col1:
150
+ st.markdown(f"**NIM**: {result['NIM']}")
151
+ st.markdown(f"**Nama**: {result['Nama']}")
152
+ st.markdown(f"**Pembimbing**: {result['Pembimbing']}")
153
+ with col2:
154
+ st.markdown(f"**Tahun**: {result['Tahun']}")
155
+ st.markdown(f"**Semester**: {result['Semester']}")
156
+ st.markdown(f"**Judul Lengkap**: {result['Judul']}")
157
+ else:
158
+ st.warning("No results found")
159
+ else:
160
+ st.warning("Please enter a search query")
161
 
162
+ # Footer
163
+ st.markdown("---")
164
+ st.markdown("""
165
+ <div style='text-align: center;'>
166
+ <p><a href='https://galih.eu'>Galih Hermawan</a> | Akabot Research Group</p>
167
+ <p>Prodi Teknik Informatika | Universitas Komputer Indonesia</p>
168
+ <p>Powered by Qwen3 Embedding Model</p>
169
+ </div>
170
+ """, unsafe_allow_html=True)