Update app.py
Browse files
app.py
CHANGED
|
@@ -9,14 +9,17 @@ hf_token = os.getenv("HF_TOKEN")
|
|
| 9 |
MODEL_ID = "Qwen/Qwen2.5-Coder-1.0B-Instruct"
|
| 10 |
|
| 11 |
print("Loading pipeline (model + tokenizer)...")
|
|
|
|
|
|
|
|
|
|
| 12 |
# The pipeline automatically downloads everything needed
|
| 13 |
-
generator = pipeline(
|
| 14 |
-
|
| 15 |
-
|
| 16 |
-
|
| 17 |
-
|
| 18 |
-
|
| 19 |
-
)
|
| 20 |
print("Pipeline loaded successfully")
|
| 21 |
|
| 22 |
app = Flask(__name__)
|
|
|
|
| 9 |
MODEL_ID = "Qwen/Qwen2.5-Coder-1.0B-Instruct"
|
| 10 |
|
| 11 |
print("Loading pipeline (model + tokenizer)...")
|
| 12 |
+
|
| 13 |
+
generator = pipeline("text-generation", model="Qwen/Qwen2.5-Coder-1.5B-Instruct")
|
| 14 |
+
|
| 15 |
# The pipeline automatically downloads everything needed
|
| 16 |
+
# generator = pipeline(
|
| 17 |
+
# "text-generation",
|
| 18 |
+
# model=MODEL_ID,
|
| 19 |
+
# device_map="cpu", # Force CPU for free-tier Spaces
|
| 20 |
+
# torch_dtype="auto",
|
| 21 |
+
# token=hf_token
|
| 22 |
+
# )
|
| 23 |
print("Pipeline loaded successfully")
|
| 24 |
|
| 25 |
app = Flask(__name__)
|