Spaces:
Running
Running
n min words default to 3 and change GPU to allow for cuda
Browse files
app.py
CHANGED
|
@@ -668,11 +668,31 @@ def generate_and_save_embeddings(
|
|
| 668 |
|
| 669 |
model = load_embedding_model(selected_embedding_model)
|
| 670 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 671 |
encode_device = None
|
| 672 |
batch_size = 32
|
|
|
|
|
|
|
| 673 |
if device == "CPU":
|
| 674 |
encode_device = "cpu"
|
| 675 |
batch_size = 64
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 676 |
|
| 677 |
embeddings = model.encode(
|
| 678 |
docs,
|
|
@@ -843,7 +863,7 @@ min_words = st.sidebar.slider(
|
|
| 843 |
f"Remove {granularity_label} shorter than N words",
|
| 844 |
min_value=1,
|
| 845 |
max_value=20,
|
| 846 |
-
value=
|
| 847 |
step=1,
|
| 848 |
help="Units (sentences or reports) with fewer words than this will be discarded "
|
| 849 |
"during preprocessing. After changing, click 'Prepare Data for This Configuration'.",
|
|
@@ -871,7 +891,7 @@ selected_embedding_model = st.sidebar.selectbox(
|
|
| 871 |
|
| 872 |
selected_device = st.sidebar.radio(
|
| 873 |
"Processing device",
|
| 874 |
-
["GPU
|
| 875 |
index=0,
|
| 876 |
)
|
| 877 |
|
|
|
|
| 668 |
|
| 669 |
model = load_embedding_model(selected_embedding_model)
|
| 670 |
|
| 671 |
+
# encode_device = None
|
| 672 |
+
# batch_size = 32
|
| 673 |
+
# if device == "CPU":
|
| 674 |
+
# encode_device = "cpu"
|
| 675 |
+
# batch_size = 64
|
| 676 |
+
|
| 677 |
encode_device = None
|
| 678 |
batch_size = 32
|
| 679 |
+
|
| 680 |
+
# If user selected CPU explicitly, skip all checks
|
| 681 |
if device == "CPU":
|
| 682 |
encode_device = "cpu"
|
| 683 |
batch_size = 64
|
| 684 |
+
else:
|
| 685 |
+
# User selected GPU. We try CUDA -> MPS -> CPU
|
| 686 |
+
import torch
|
| 687 |
+
if torch.cuda.is_available():
|
| 688 |
+
encode_device = "cuda"
|
| 689 |
+
st.toast("Using NVIDIA GPU (CUDA)")
|
| 690 |
+
elif torch.backends.mps.is_available():
|
| 691 |
+
encode_device = "mps"
|
| 692 |
+
st.toast("Using Apple GPU (MPS)")
|
| 693 |
+
else:
|
| 694 |
+
encode_device = "cpu"
|
| 695 |
+
st.warning("No GPU found (neither CUDA nor MPS). Falling back to CPU.")
|
| 696 |
|
| 697 |
embeddings = model.encode(
|
| 698 |
docs,
|
|
|
|
| 863 |
f"Remove {granularity_label} shorter than N words",
|
| 864 |
min_value=1,
|
| 865 |
max_value=20,
|
| 866 |
+
value=3, # default = 3 words
|
| 867 |
step=1,
|
| 868 |
help="Units (sentences or reports) with fewer words than this will be discarded "
|
| 869 |
"during preprocessing. After changing, click 'Prepare Data for This Configuration'.",
|
|
|
|
| 891 |
|
| 892 |
selected_device = st.sidebar.radio(
|
| 893 |
"Processing device",
|
| 894 |
+
["GPU", "CPU"],
|
| 895 |
index=0,
|
| 896 |
)
|
| 897 |
|