Spaces:
Sleeping
Sleeping
Update src/streamlit_app.py
Browse files- src/streamlit_app.py +5 -4
src/streamlit_app.py
CHANGED
|
@@ -180,17 +180,18 @@ Equipment: {equipment_list}
|
|
| 180 |
with st.spinner("Generating your AI workout plan..."):
|
| 181 |
|
| 182 |
inputs = tokenizer(prompt, return_tensors="pt", truncation=True)
|
| 183 |
-
|
| 184 |
outputs = model.generate(
|
| 185 |
-
inputs,
|
| 186 |
max_new_tokens=600,
|
| 187 |
temperature=0.3,
|
| 188 |
do_sample=False,
|
| 189 |
-
top_p=0.9,
|
| 190 |
repetition_penalty=1.2
|
| 191 |
)
|
|
|
|
|
|
|
| 192 |
|
| 193 |
-
|
| 194 |
|
| 195 |
st.subheader("🏋️ Your Personalized Workout Plan")
|
| 196 |
st.write(result)
|
|
|
|
| 180 |
with st.spinner("Generating your AI workout plan..."):
|
| 181 |
|
| 182 |
inputs = tokenizer(prompt, return_tensors="pt", truncation=True)
|
| 183 |
+
|
| 184 |
outputs = model.generate(
|
| 185 |
+
**inputs,
|
| 186 |
max_new_tokens=600,
|
| 187 |
temperature=0.3,
|
| 188 |
do_sample=False,
|
|
|
|
| 189 |
repetition_penalty=1.2
|
| 190 |
)
|
| 191 |
+
|
| 192 |
+
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
| 193 |
|
| 194 |
+
st.write(response)
|
| 195 |
|
| 196 |
st.subheader("🏋️ Your Personalized Workout Plan")
|
| 197 |
st.write(result)
|