Spaces:
Sleeping
Sleeping
Commit ·
7a5d2ae
1
Parent(s): b88f075
modify the pipeline
Browse files
app.py
CHANGED
|
@@ -121,12 +121,6 @@ def get_retrieval_qa_chain(text_file, hf_model):
|
|
| 121 |
if text_file != default_text_file or default_text_file is None:
|
| 122 |
if text_file is not None and os.path.exists(text_file):
|
| 123 |
retriever, vectorstore = prepare_vector_store_retriever(text_file)
|
| 124 |
-
else:
|
| 125 |
-
# Create a dummy retriever if no file is available
|
| 126 |
-
dummy_doc = Document(page_content="No document loaded. Please upload a file to get started.")
|
| 127 |
-
dummy_vectorstore = FAISS.from_documents([dummy_doc], embeddings)
|
| 128 |
-
retriever = VectorStoreRetriever(vectorstore=dummy_vectorstore, search_kwargs={"k": 1})
|
| 129 |
-
vectorstore = dummy_vectorstore
|
| 130 |
|
| 131 |
chain = RetrievalQA.from_chain_type(
|
| 132 |
llm=hf_model,
|
|
@@ -146,15 +140,12 @@ def generate(question, answer, text_file, max_new_tokens):
|
|
| 146 |
# Create pipeline without streamer first to test
|
| 147 |
phi2_pipeline = pipeline(
|
| 148 |
"text-generation",
|
| 149 |
-
tokenizer=tokenizer,
|
| 150 |
model=model,
|
|
|
|
| 151 |
max_new_tokens=max_new_tokens,
|
|
|
|
| 152 |
pad_token_id=tokenizer.eos_token_id,
|
| 153 |
eos_token_id=tokenizer.eos_token_id,
|
| 154 |
-
do_sample=True,
|
| 155 |
-
temperature=0.7,
|
| 156 |
-
top_p=0.9,
|
| 157 |
-
repetition_penalty=1.1,
|
| 158 |
)
|
| 159 |
|
| 160 |
hf_model = HuggingFacePipeline(pipeline=phi2_pipeline)
|
|
|
|
| 121 |
if text_file != default_text_file or default_text_file is None:
|
| 122 |
if text_file is not None and os.path.exists(text_file):
|
| 123 |
retriever, vectorstore = prepare_vector_store_retriever(text_file)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 124 |
|
| 125 |
chain = RetrievalQA.from_chain_type(
|
| 126 |
llm=hf_model,
|
|
|
|
| 140 |
# Create pipeline without streamer first to test
|
| 141 |
phi2_pipeline = pipeline(
|
| 142 |
"text-generation",
|
|
|
|
| 143 |
model=model,
|
| 144 |
+
tokenizer=tokenizer,
|
| 145 |
max_new_tokens=max_new_tokens,
|
| 146 |
+
do_sample=False, # ← greedy
|
| 147 |
pad_token_id=tokenizer.eos_token_id,
|
| 148 |
eos_token_id=tokenizer.eos_token_id,
|
|
|
|
|
|
|
|
|
|
|
|
|
| 149 |
)
|
| 150 |
|
| 151 |
hf_model = HuggingFacePipeline(pipeline=phi2_pipeline)
|