Spaces:
Sleeping
Sleeping
Upload app.py
Browse files
app.py
CHANGED
|
@@ -11,9 +11,10 @@ from huggingface_hub import InferenceClient
|
|
| 11 |
# CONFIG
|
| 12 |
|
| 13 |
# You can hard-code the model here, or override with MODEL_ID as a Space variable.
|
| 14 |
-
|
| 15 |
-
|
| 16 |
-
HF_TOKEN = os.environ.get("HF_TOKEN") #
|
|
|
|
| 17 |
|
| 18 |
REPO_PATH = "/data/questions.json" # where we store generated questions
|
| 19 |
|
|
|
|
| 11 |
# CONFIG
|
| 12 |
|
| 13 |
# You can hard-code the model here, or override with MODEL_ID as a Space variable.
|
| 14 |
+
# Use a free, public chat model that doesn't need a novita api_key
|
| 15 |
+
MODEL_ID = os.environ.get("MODEL_ID", "TinyLlama/TinyLlama-1.1B-Chat-v1.0")
|
| 16 |
+
HF_TOKEN = os.environ.get("HF_TOKEN") # can stay empty
|
| 17 |
+
|
| 18 |
|
| 19 |
REPO_PATH = "/data/questions.json" # where we store generated questions
|
| 20 |
|