AIencoder commited on
Commit
3ed108f
·
verified ·
1 Parent(s): f97dde7

Update entrypoint.sh

Browse files
Files changed (1) hide show
  1. entrypoint.sh +11 -5
entrypoint.sh CHANGED
@@ -3,18 +3,25 @@
3
  echo "🔥 Axon v6 - llama.cpp Edition"
4
  echo "📥 Downloading models..."
5
 
6
- # Create models directory
7
  mkdir -p /models
8
 
9
- # Download GGUFs from HuggingFace
10
  python3 << 'EOF'
11
  from huggingface_hub import hf_hub_download
12
- import os
13
 
14
  models = [
 
 
 
 
 
15
  ("Qwen/Qwen2.5-Coder-7B-Instruct-GGUF", "qwen2.5-coder-7b-instruct-q4_k_m.gguf"),
16
  ("Qwen/Qwen2.5-Coder-3B-Instruct-GGUF", "qwen2.5-coder-3b-instruct-q4_k_m.gguf"),
17
  ("Qwen/Qwen2.5-Coder-1.5B-Instruct-GGUF", "qwen2.5-coder-1.5b-instruct-q4_k_m.gguf"),
 
 
 
 
 
18
  ]
19
 
20
  for repo, filename in models:
@@ -24,11 +31,10 @@ for repo, filename in models:
24
  repo_id=repo,
25
  filename=filename,
26
  local_dir="/models",
27
- local_dir_use_symlinks=False
28
  )
29
  print(f"✅ {filename} ready!")
30
  except Exception as e:
31
- print(f" Failed to download {filename}: {e}")
32
 
33
  print("✅ All models downloaded!")
34
  EOF
 
3
  echo "🔥 Axon v6 - llama.cpp Edition"
4
  echo "📥 Downloading models..."
5
 
 
6
  mkdir -p /models
7
 
 
8
  python3 << 'EOF'
9
  from huggingface_hub import hf_hub_download
 
10
 
11
  models = [
12
+ # Qwen3 30B MoE (Best!)
13
+ ("unsloth/Qwen3-Coder-30B-A3B-Instruct-GGUF", "Qwen3-Coder-30B-A3B-Instruct-Q4_K_M.gguf"),
14
+
15
+ # Qwen2.5 family
16
+ ("Qwen/Qwen2.5-Coder-14B-Instruct-GGUF", "qwen2.5-coder-14b-instruct-q4_k_m.gguf"),
17
  ("Qwen/Qwen2.5-Coder-7B-Instruct-GGUF", "qwen2.5-coder-7b-instruct-q4_k_m.gguf"),
18
  ("Qwen/Qwen2.5-Coder-3B-Instruct-GGUF", "qwen2.5-coder-3b-instruct-q4_k_m.gguf"),
19
  ("Qwen/Qwen2.5-Coder-1.5B-Instruct-GGUF", "qwen2.5-coder-1.5b-instruct-q4_k_m.gguf"),
20
+ ("Qwen/Qwen2.5-Coder-0.5B-Instruct-GGUF", "qwen2.5-coder-0.5b-instruct-q4_k_m.gguf"),
21
+
22
+ # DeepSeek
23
+ ("bartowski/DeepSeek-Coder-V2-Lite-Instruct-GGUF", "DeepSeek-Coder-V2-Lite-Instruct-Q4_K_M.gguf"),
24
+ ("TheBloke/deepseek-coder-6.7B-instruct-GGUF", "deepseek-coder-6.7b-instruct.Q4_K_M.gguf"),
25
  ]
26
 
27
  for repo, filename in models:
 
31
  repo_id=repo,
32
  filename=filename,
33
  local_dir="/models",
 
34
  )
35
  print(f"✅ {filename} ready!")
36
  except Exception as e:
37
+ print(f"⚠️ {filename}: {e}")
38
 
39
  print("✅ All models downloaded!")
40
  EOF