Spaces:
Sleeping
Sleeping
| import os | |
| import torch | |
| from transformers import AutoModelForCausalLM, AutoTokenizer | |
| def load_model_and_tok(): | |
| AUTH_TOKEN = os.environ.get("HF_TOKEN", False) | |
| model = AutoModelForCausalLM.from_pretrained( | |
| "meta-llama/Llama-3.2-3B-Instruct", | |
| token=AUTH_TOKEN, | |
| ) | |
| tok = AutoTokenizer.from_pretrained( | |
| "meta-llama/Llama-3.2-3B-Instruct", | |
| token=AUTH_TOKEN, | |
| ) | |
| model = model.to("cuda" if torch.cuda.is_available() else "cpu").eval() | |
| return model, tok | |