test emb
Browse files- config.json +2 -1
- modeling_aria.py +2 -2
config.json
CHANGED
|
@@ -1,10 +1,11 @@
|
|
| 1 |
{
|
| 2 |
"architectures": [
|
| 3 |
-
"
|
| 4 |
],
|
| 5 |
"bos_token_id": 0,
|
| 6 |
"eos_token_id": 1,
|
| 7 |
"hidden_size": 1536,
|
|
|
|
| 8 |
"intermediate_size": 6144,
|
| 9 |
"max_position_embeddings": 8192,
|
| 10 |
"model_type": "aria",
|
|
|
|
| 1 |
{
|
| 2 |
"architectures": [
|
| 3 |
+
"AriaForSequenceEmbeddings"
|
| 4 |
],
|
| 5 |
"bos_token_id": 0,
|
| 6 |
"eos_token_id": 1,
|
| 7 |
"hidden_size": 1536,
|
| 8 |
+
"embedding_size": 512,
|
| 9 |
"intermediate_size": 6144,
|
| 10 |
"max_position_embeddings": 8192,
|
| 11 |
"model_type": "aria",
|
modeling_aria.py
CHANGED
|
@@ -617,7 +617,7 @@ class AriaForCausalLM(AriaPreTrainedModel, GenerationMixin):
|
|
| 617 |
)
|
| 618 |
|
| 619 |
|
| 620 |
-
class
|
| 621 |
"""Transformer decoder embedding head for contrastive learning.
|
| 622 |
|
| 623 |
Args:
|
|
@@ -744,5 +744,5 @@ __all__ = [
|
|
| 744 |
"AriaModel",
|
| 745 |
"TransformerBlock",
|
| 746 |
"AriaForCausalLM",
|
| 747 |
-
"
|
| 748 |
]
|
|
|
|
| 617 |
)
|
| 618 |
|
| 619 |
|
| 620 |
+
class AriaForSequenceEmbedding(AriaPreTrainedModel):
|
| 621 |
"""Transformer decoder embedding head for contrastive learning.
|
| 622 |
|
| 623 |
Args:
|
|
|
|
| 744 |
"AriaModel",
|
| 745 |
"TransformerBlock",
|
| 746 |
"AriaForCausalLM",
|
| 747 |
+
"AriaForSequenceEmbedding",
|
| 748 |
]
|