Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -3,7 +3,8 @@ from transformers import AutoTokenizer, AutoModelForCausalLM
|
|
| 3 |
import torch
|
| 4 |
|
| 5 |
# Load model and tokenizer from Hugging Face Hub
|
| 6 |
-
model_id = "deepseek-ai/deepseek-coder-1.3b-base"
|
|
|
|
| 7 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
| 8 |
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32)
|
| 9 |
|
|
|
|
| 3 |
import torch
|
| 4 |
|
| 5 |
# Load model and tokenizer from Hugging Face Hub
|
| 6 |
+
# model_id = "deepseek-ai/deepseek-coder-1.3b-base"
|
| 7 |
+
model_id = "gpt2"
|
| 8 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
| 9 |
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32)
|
| 10 |
|