loua19 commited on
Commit
840c6a3
·
1 Parent(s): cea6439
Files changed (2) hide show
  1. config.json +2 -1
  2. modeling_aria.py +2 -2
config.json CHANGED
@@ -1,10 +1,11 @@
1
  {
2
  "architectures": [
3
- "AriaForCausalLM"
4
  ],
5
  "bos_token_id": 0,
6
  "eos_token_id": 1,
7
  "hidden_size": 1536,
 
8
  "intermediate_size": 6144,
9
  "max_position_embeddings": 8192,
10
  "model_type": "aria",
 
1
  {
2
  "architectures": [
3
+ "AriaForSequenceEmbeddings"
4
  ],
5
  "bos_token_id": 0,
6
  "eos_token_id": 1,
7
  "hidden_size": 1536,
8
+ "embedding_size": 512,
9
  "intermediate_size": 6144,
10
  "max_position_embeddings": 8192,
11
  "model_type": "aria",
modeling_aria.py CHANGED
@@ -617,7 +617,7 @@ class AriaForCausalLM(AriaPreTrainedModel, GenerationMixin):
617
  )
618
 
619
 
620
- class AriaForSequenceEmbeddings(AriaPreTrainedModel):
621
  """Transformer decoder embedding head for contrastive learning.
622
 
623
  Args:
@@ -744,5 +744,5 @@ __all__ = [
744
  "AriaModel",
745
  "TransformerBlock",
746
  "AriaForCausalLM",
747
- "AriaForSequenceEmbeddings",
748
  ]
 
617
  )
618
 
619
 
620
+ class AriaForSequenceEmbedding(AriaPreTrainedModel):
621
  """Transformer decoder embedding head for contrastive learning.
622
 
623
  Args:
 
744
  "AriaModel",
745
  "TransformerBlock",
746
  "AriaForCausalLM",
747
+ "AriaForSequenceEmbedding",
748
  ]