Spaces:
Sleeping
Sleeping
| import os | |
| import torch | |
| from dotenv import load_dotenv | |
| # Load .env file (for local development) | |
| load_dotenv() | |
| # Model path (uses default if environment variable is missing) | |
| BASE_MODEL = os.getenv("BASE_MODEL", "Qwen/Qwen2.5-3B-Instruct") | |
| ADAPTERS = os.getenv("ADAPTER_MODEL", "m97j/npc_LoRA-fps") | |
| # Device configuration | |
| DEVICE = os.getenv("DEVICE", "cuda" if torch.cuda.is_available() else "cpu") | |
| # Tokenizer/Model common parameters | |
| MAX_LENGTH = int(os.getenv("MAX_LENGTH", 1024)) | |
| NUM_FLAGS = int(os.getenv("NUM_FLAGS", 7)) # match withflags.json | |
| # Generation parameters (can be overridden at inference time) | |
| GEN_MAX_NEW_TOKENS = int(os.getenv("GEN_MAX_NEW_TOKENS", 400)) | |
| GEN_TEMPERATURE = float(os.getenv("GEN_TEMPERATURE", 0.7)) | |
| GEN_TOP_P = float(os.getenv("GEN_TOP_P", 0.9)) | |
| # Hugging Face Token (For Private Model Access) | |
| HF_TOKEN = os.getenv("HF_TOKEN") | |