Upload Cybersecurity-SLM v2
Browse files
config.py
CHANGED
|
@@ -29,8 +29,8 @@ class SLMConfig:
|
|
| 29 |
dropout: float = 0.1
|
| 30 |
bias: bool = False
|
| 31 |
ffn_multiplier: float = 2.667
|
| 32 |
-
max_position_embeddings: int =
|
| 33 |
-
rope_theta: float =
|
| 34 |
sliding_window: Optional[int] = None
|
| 35 |
|
| 36 |
batch_size: int = 4
|
|
@@ -46,7 +46,7 @@ class SLMConfig:
|
|
| 46 |
log_interval: int = 10
|
| 47 |
device: str = "auto"
|
| 48 |
|
| 49 |
-
max_new_tokens: int =
|
| 50 |
temperature: float = 0.8
|
| 51 |
top_k: int = 50
|
| 52 |
top_p: float = 0.9
|
|
|
|
| 29 |
dropout: float = 0.1
|
| 30 |
bias: bool = False
|
| 31 |
ffn_multiplier: float = 2.667
|
| 32 |
+
max_position_embeddings: int = 100_000_000_000 # 100B tokens via RoPE
|
| 33 |
+
rope_theta: float = 50_000_000_000.0 # Scaled for 100B context
|
| 34 |
sliding_window: Optional[int] = None
|
| 35 |
|
| 36 |
batch_size: int = 4
|
|
|
|
| 46 |
log_interval: int = 10
|
| 47 |
device: str = "auto"
|
| 48 |
|
| 49 |
+
max_new_tokens: int = 1_000_000 # 1M output tokens
|
| 50 |
temperature: float = 0.8
|
| 51 |
top_k: int = 50
|
| 52 |
top_p: float = 0.9
|