Spaces:
Sleeping
Sleeping
๐w๐
Browse files
app.py
CHANGED
|
@@ -9,6 +9,7 @@ import spaces
|
|
| 9 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
|
| 10 |
import torch
|
| 11 |
from threading import Thread
|
|
|
|
| 12 |
|
| 13 |
token = os.environ["HF_TOKEN"]
|
| 14 |
model = AutoModelForCausalLM.from_pretrained("google/gemma-7b-it",
|
|
@@ -25,9 +26,12 @@ title_text_dataset = load_dataset(
|
|
| 25 |
).select_columns(["title", "text"])
|
| 26 |
|
| 27 |
# Load the int8 and binary indices. Int8 is loaded as a view to save memory, as we never actually perform search with it.
|
| 28 |
-
|
|
|
|
|
|
|
|
|
|
| 29 |
binary_index: faiss.IndexBinaryFlat = faiss.read_index_binary(
|
| 30 |
-
|
| 31 |
)
|
| 32 |
|
| 33 |
# Load the SentenceTransformer model for embedding the queries
|
|
@@ -154,5 +158,6 @@ demo = gr.ChatInterface(fn=talk,
|
|
| 154 |
chatbot=gr.Chatbot(show_label=True, show_share_button=True, show_copy_button=True, likeable=True, layout="bubble", bubble_full_width=False),
|
| 155 |
theme="Soft",
|
| 156 |
examples=[["what is machine learning"]],
|
| 157 |
-
title=
|
|
|
|
| 158 |
demo.launch()
|
|
|
|
| 9 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
|
| 10 |
import torch
|
| 11 |
from threading import Thread
|
| 12 |
+
from huggingface_hub import hf_hub_download
|
| 13 |
|
| 14 |
token = os.environ["HF_TOKEN"]
|
| 15 |
model = AutoModelForCausalLM.from_pretrained("google/gemma-7b-it",
|
|
|
|
| 26 |
).select_columns(["title", "text"])
|
| 27 |
|
| 28 |
# Load the int8 and binary indices. Int8 is loaded as a view to save memory, as we never actually perform search with it.
|
| 29 |
+
path_int8_view = hf_hub_download(repo_id="sentence-transformers/quantized-retrieval",repo_type="space", filename="wikipedia_ubinary_faiss_1m.index")
|
| 30 |
+
int8_view = Index.restore(path_int8_view, view=True)
|
| 31 |
+
|
| 32 |
+
path_binary_index = hf_hub_download(repo_id="sentence-transformers/quantized-retrieval",repo_type="space", filename="wikipedia_ubinary_faiss_1m.index")
|
| 33 |
binary_index: faiss.IndexBinaryFlat = faiss.read_index_binary(
|
| 34 |
+
path_binary_index
|
| 35 |
)
|
| 36 |
|
| 37 |
# Load the SentenceTransformer model for embedding the queries
|
|
|
|
| 158 |
chatbot=gr.Chatbot(show_label=True, show_share_button=True, show_copy_button=True, likeable=True, layout="bubble", bubble_full_width=False),
|
| 159 |
theme="Soft",
|
| 160 |
examples=[["what is machine learning"]],
|
| 161 |
+
title=TITLE,
|
| 162 |
+
description=DESCRIPTION)
|
| 163 |
demo.launch()
|