Spaces:
Runtime error
Runtime error
Trying with Falcon
Browse files
app.py
CHANGED
|
@@ -15,10 +15,15 @@ from transformers import pipeline
|
|
| 15 |
#def resp(text):
|
| 16 |
#summarize
|
| 17 |
# return response(text)[0]["summary_text"]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 18 |
|
| 19 |
|
| 20 |
-
hub_model_id = "
|
| 21 |
-
response = pipeline("
|
| 22 |
|
| 23 |
def resp(text):
|
| 24 |
return response(text)
|
|
|
|
| 15 |
#def resp(text):
|
| 16 |
#summarize
|
| 17 |
# return response(text)[0]["summary_text"]
|
| 18 |
+
#hub_model_id = "WizardLM/WizardLM-7B-V1.0"
|
| 19 |
+
#response = pipeline("text2text-generation", model=hub_model_id)
|
| 20 |
+
#Obtuve este error, no encontro el modelo
|
| 21 |
+
#OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like WizardLM/WizardLM-7B-V1.0 is not the path to a directory containing a file named config.json.
|
| 22 |
+
#Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'.
|
| 23 |
|
| 24 |
|
| 25 |
+
hub_model_id = "tiiuae/falcon-7b-instruct"
|
| 26 |
+
response = pipeline("text-generation", model=hub_model_id)
|
| 27 |
|
| 28 |
def resp(text):
|
| 29 |
return response(text)
|