Update models/loader.py
Browse files- models/loader.py +2 -2
models/loader.py
CHANGED
|
@@ -2,15 +2,14 @@
|
|
| 2 |
import torch
|
| 3 |
import os
|
| 4 |
from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig
|
|
|
|
| 5 |
|
| 6 |
-
# Configuration for loading models
|
| 7 |
QUANTIZATION_CONFIG = BitsAndBytesConfig(
|
| 8 |
load_in_4bit=True,
|
| 9 |
bnb_4bit_quant_type="nf4",
|
| 10 |
bnb_4bit_compute_dtype=torch.bfloat16,
|
| 11 |
)
|
| 12 |
|
| 13 |
-
# Use a registry to map agent roles to specific models
|
| 14 |
MODEL_REGISTRY = {
|
| 15 |
"ceo": "Qwen/Qwen3-0.6B",
|
| 16 |
"planner": "Qwen/Qwen3-0.6B",
|
|
@@ -21,6 +20,7 @@ MODEL_REGISTRY = {
|
|
| 21 |
"worker_backend_coder": "Qwen/Qwen3-0.6B",
|
| 22 |
"worker_front_end_coder": "Qwen/Qwen3-0.6B",
|
| 23 |
"worker_tester": "Qwen/Qwen3-0.6B",
|
|
|
|
| 24 |
}
|
| 25 |
_MODEL_CACHE = {}
|
| 26 |
|
|
|
|
| 2 |
import torch
|
| 3 |
import os
|
| 4 |
from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig
|
| 5 |
+
from backend.agents import ROLE_PROMPTS
|
| 6 |
|
|
|
|
| 7 |
QUANTIZATION_CONFIG = BitsAndBytesConfig(
|
| 8 |
load_in_4bit=True,
|
| 9 |
bnb_4bit_quant_type="nf4",
|
| 10 |
bnb_4bit_compute_dtype=torch.bfloat16,
|
| 11 |
)
|
| 12 |
|
|
|
|
| 13 |
MODEL_REGISTRY = {
|
| 14 |
"ceo": "Qwen/Qwen3-0.6B",
|
| 15 |
"planner": "Qwen/Qwen3-0.6B",
|
|
|
|
| 20 |
"worker_backend_coder": "Qwen/Qwen3-0.6B",
|
| 21 |
"worker_front_end_coder": "Qwen/Qwen3-0.6B",
|
| 22 |
"worker_tester": "Qwen/Qwen3-0.6B",
|
| 23 |
+
"code_analyst": "Qwen/Qwen3-0.6B"
|
| 24 |
}
|
| 25 |
_MODEL_CACHE = {}
|
| 26 |
|