Update configuration_binaryllm.py
Browse files
configuration_binaryllm.py
CHANGED
|
@@ -14,6 +14,7 @@ class BinaryLLMConfig(PretrainedConfig):
|
|
| 14 |
max_position_embeddings: int = 2048,
|
| 15 |
dropout: float = 0.1,
|
| 16 |
activation: str = "gelu",
|
|
|
|
| 17 |
bos_token_id: int = 65536,
|
| 18 |
eos_token_id: int = 65537,
|
| 19 |
pad_token_id: int = 65537,
|
|
@@ -27,9 +28,10 @@ class BinaryLLMConfig(PretrainedConfig):
|
|
| 27 |
self.max_position_embeddings = int(max_position_embeddings)
|
| 28 |
self.dropout = float(dropout)
|
| 29 |
self.activation = str(activation)
|
|
|
|
| 30 |
|
| 31 |
self.bos_token_id = int(bos_token_id)
|
| 32 |
self.eos_token_id = int(eos_token_id)
|
| 33 |
self.pad_token_id = int(pad_token_id)
|
| 34 |
|
| 35 |
-
super().__init__(**kwargs)
|
|
|
|
| 14 |
max_position_embeddings: int = 2048,
|
| 15 |
dropout: float = 0.1,
|
| 16 |
activation: str = "gelu",
|
| 17 |
+
attn_backend: str = "auto",
|
| 18 |
bos_token_id: int = 65536,
|
| 19 |
eos_token_id: int = 65537,
|
| 20 |
pad_token_id: int = 65537,
|
|
|
|
| 28 |
self.max_position_embeddings = int(max_position_embeddings)
|
| 29 |
self.dropout = float(dropout)
|
| 30 |
self.activation = str(activation)
|
| 31 |
+
self.attn_backend = str(attn_backend)
|
| 32 |
|
| 33 |
self.bos_token_id = int(bos_token_id)
|
| 34 |
self.eos_token_id = int(eos_token_id)
|
| 35 |
self.pad_token_id = int(pad_token_id)
|
| 36 |
|
| 37 |
+
super().__init__(**kwargs)
|