eddddyy commited on
Commit
7883432
Β·
verified Β·
1 Parent(s): e15e364

Update model_loader.py

Browse files
Files changed (1) hide show
  1. model_loader.py +34 -43
model_loader.py CHANGED
@@ -1,47 +1,38 @@
1
- import torch
2
- from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
3
- from config import HF_TOKEN, MODEL_ID
4
 
5
- def load_model():
6
- try:
7
- print(f"πŸ”„ Loading tokenizer and model: {MODEL_ID}")
 
8
 
9
- # Load tokenizer
10
- tokenizer = AutoTokenizer.from_pretrained(
11
- MODEL_ID,
12
- token=HF_TOKEN,
13
- trust_remote_code=True,
14
- padding_side="left" # For chat-style models
15
- )
16
 
17
- # Set max length (MedGemma supports up to 8192 tokens)
18
- tokenizer.model_max_length = 8192
 
19
 
20
- # Load model
21
- model = AutoModelForCausalLM.from_pretrained(
22
- MODEL_ID,
23
- token=HF_TOKEN,
24
- trust_remote_code=True,
25
- device_map="auto" if torch.cuda.is_available() else "cpu",
26
- torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
27
- low_cpu_mem_usage=True
28
- )
29
-
30
- print("βœ… Model loaded successfully.")
31
-
32
- # Return generation pipeline with large max token output and context window
33
- return pipeline(
34
- "text-generation",
35
- model=model,
36
- tokenizer=tokenizer,
37
- max_new_tokens=1024, # Max response length
38
- truncation=True, # Safely truncate input if needed
39
- pad_token_id=tokenizer.eos_token_id,
40
- do_sample=True,
41
- temperature=0.7,
42
- top_p=0.9
43
- )
44
-
45
- except Exception as e:
46
- print(f"❌ Failed to load model: {e}")
47
- raise RuntimeError(f"Model loading failed: {e}")
 
1
+ import os
 
 
2
 
3
+ # Hugging Face Token
4
+ HF_TOKEN = os.getenv("HF_TOKEN")
5
+ if not HF_TOKEN:
6
+ raise ValueError("❌ HF_TOKEN not set in environment variables.")
7
 
8
+ # Model to Load
9
+ MODEL_ID = os.getenv("MODEL_ID", "meta-llama/Meta-Llama-3-8B-Instruct") # πŸ‘ˆ updated here
 
 
 
 
 
10
 
11
+ # Paths and Settings
12
+ TEMP_DIR = os.getenv("TEMP_DIR", "/tmp") # For screenshots, temporary files, etc.
13
+ ENABLE_EXECUTION = os.getenv("ENABLE_EXECUTION", "true").lower() == "true" # toggle command execution
14
 
15
+ # OS Commands (unchanged)
16
+ OS_COMMANDS = {
17
+ "open": {
18
+ "chrome": "google-chrome",
19
+ "firefox": "firefox",
20
+ "notepad": "notepad",
21
+ "calculator": "gnome-calculator",
22
+ "terminal": "gnome-terminal",
23
+ "vscode": "code"
24
+ },
25
+ "close": {
26
+ "chrome": "pkill chrome",
27
+ "firefox": "pkill firefox",
28
+ "notepad": "pkill notepad",
29
+ "calculator": "pkill gnome-calculator",
30
+ "vscode": "pkill code"
31
+ },
32
+ "screenshot": {
33
+ "default": "import -window root screenshot.png"
34
+ },
35
+ "shutdown": "shutdown now",
36
+ "restart": "reboot",
37
+ "lock": "gnome-screensaver-command -l"
38
+ }