dixiebone13-a11y commited on
Commit
e4c1061
·
1 Parent(s): 3500155

Fix: Pass HF_TOKEN to PeftModel for gated model access

Browse files
Files changed (1) hide show
  1. app.py +12 -2
app.py CHANGED
@@ -122,12 +122,21 @@ def compute_consciousness(
122
  print("🔮 Loading Oracle Engine (Qwen2.5-32B-Instruct 4-bit + LoRA)...")
123
  from transformers import AutoModelForCausalLM, AutoTokenizer
124
  from peft import PeftModel
 
125
 
126
  BASE_MODEL_ID = "unsloth/Qwen2.5-32B-Instruct-bnb-4bit"
127
  LORA_MODEL_ID = "Vikingdude81/oracle-engine-32b-lora"
128
 
 
 
 
 
 
 
 
 
129
  # Load tokenizer from base model (LoRA only has weights, not tokenizer)
130
- tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL_ID)
131
 
132
  # Load base model
133
  base_model = AutoModelForCausalLM.from_pretrained(
@@ -135,11 +144,12 @@ base_model = AutoModelForCausalLM.from_pretrained(
135
  device_map="auto",
136
  torch_dtype=torch.bfloat16,
137
  trust_remote_code=True,
 
138
  )
139
 
140
  # Apply LoRA adapter
141
  print("🔗 Applying LoRA adapter...")
142
- model = PeftModel.from_pretrained(base_model, LORA_MODEL_ID)
143
  model.eval()
144
 
145
  HIDDEN_DIM = model.config.hidden_size
 
122
  print("🔮 Loading Oracle Engine (Qwen2.5-32B-Instruct 4-bit + LoRA)...")
123
  from transformers import AutoModelForCausalLM, AutoTokenizer
124
  from peft import PeftModel
125
+ from huggingface_hub import login
126
 
127
  BASE_MODEL_ID = "unsloth/Qwen2.5-32B-Instruct-bnb-4bit"
128
  LORA_MODEL_ID = "Vikingdude81/oracle-engine-32b-lora"
129
 
130
+ # Authenticate with HF token from environment (set in Space secrets)
131
+ HF_TOKEN = os.environ.get("HF_TOKEN")
132
+ if HF_TOKEN:
133
+ print("🔑 Found HF_TOKEN, logging in...")
134
+ login(token=HF_TOKEN)
135
+ else:
136
+ print("⚠️ No HF_TOKEN found, attempting public access...")
137
+
138
  # Load tokenizer from base model (LoRA only has weights, not tokenizer)
139
+ tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL_ID, token=HF_TOKEN)
140
 
141
  # Load base model
142
  base_model = AutoModelForCausalLM.from_pretrained(
 
144
  device_map="auto",
145
  torch_dtype=torch.bfloat16,
146
  trust_remote_code=True,
147
+ token=HF_TOKEN,
148
  )
149
 
150
  # Apply LoRA adapter
151
  print("🔗 Applying LoRA adapter...")
152
+ model = PeftModel.from_pretrained(base_model, LORA_MODEL_ID, token=HF_TOKEN)
153
  model.eval()
154
 
155
  HIDDEN_DIM = model.config.hidden_size