Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -12,8 +12,8 @@ def load_model():
|
|
| 12 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 13 |
model = AutoModelForCausalLM.from_pretrained(
|
| 14 |
model_name,
|
| 15 |
-
torch_dtype=torch.
|
| 16 |
-
device_map="
|
| 17 |
)
|
| 18 |
return model, tokenizer
|
| 19 |
|
|
@@ -79,6 +79,7 @@ def main():
|
|
| 79 |
with st.spinner("Réflexion en cours..."):
|
| 80 |
prompt = format_prompt(question)
|
| 81 |
response = generate_response(prompt, model, tokenizer)
|
|
|
|
| 82 |
st.markdown(response)
|
| 83 |
st.session_state.messages.append({"role": "assistant", "content": response})
|
| 84 |
|
|
|
|
| 12 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 13 |
model = AutoModelForCausalLM.from_pretrained(
|
| 14 |
model_name,
|
| 15 |
+
torch_dtype=torch.float32,
|
| 16 |
+
device_map="cpu"
|
| 17 |
)
|
| 18 |
return model, tokenizer
|
| 19 |
|
|
|
|
| 79 |
with st.spinner("Réflexion en cours..."):
|
| 80 |
prompt = format_prompt(question)
|
| 81 |
response = generate_response(prompt, model, tokenizer)
|
| 82 |
+
response = response.replace('<think>', '')
|
| 83 |
st.markdown(response)
|
| 84 |
st.session_state.messages.append({"role": "assistant", "content": response})
|
| 85 |
|