tendatngo commited on
Commit
c554a00
·
verified ·
1 Parent(s): b19a5a9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -6
README.md CHANGED
@@ -10,9 +10,9 @@ language:
10
  - en
11
  ---
12
 
13
- # PRAG Encoder Model
14
 
15
- This is a PRAG (Prompting for RAG-styled recommendation) encoder model trained on Amazon review data. It uses a two-tower architecture with a shared ModernBERT backbone to embed both user queries and product items into a common vector space for efficient retrieval.
16
 
17
  ## Model Details
18
 
@@ -48,7 +48,7 @@ This model can be used directly for:
48
  <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
49
  - Non-English text (not explicitly trained/validated).
50
  - General-purpose text embedding outside the e-commerce/recommendation domain.
51
- - Tasks requiring generative capabilities (this is an encoder-only model).
52
 
53
  ## Bias, Risks, and Limitations
54
 
@@ -115,7 +115,7 @@ except Exception as e:
115
 
116
  ## Training Details
117
 
118
- The model was trained following standard large language model (LLM) training practices for encoder-based retrieval systems.
119
 
120
  ### Training Data
121
 
@@ -125,7 +125,7 @@ The model is trained on diverse product-related datasets, including product revi
125
 
126
  #### Preprocessing
127
 
128
- - Standard tokenization using a ModernBERT-based tokenizer.
129
  - Input sequences are formatted to represent user intent (queries) and product characteristics (items).
130
  - Dynamic padding and truncation are applied to optimize training efficiency.
131
 
@@ -152,7 +152,7 @@ The model is evaluated using standard retrieval metrics:
152
 
153
  ### Model Architecture and Objective
154
 
155
- - **Backbone:** Standard transformer-based encoder for text representation.
156
  - **Objective:** Contrastive learning objective
157
 
158
  ### Software
 
10
  - en
11
  ---
12
 
13
+ # PRAG
14
 
15
+ This is PRAG, a LLM model trained for multi recommendation tasks and domains. It uses a two-tower architecture with a shared model to embed both user queries and product items into a common vector space for efficient retrieval.
16
 
17
  ## Model Details
18
 
 
48
  <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
49
  - Non-English text (not explicitly trained/validated).
50
  - General-purpose text embedding outside the e-commerce/recommendation domain.
51
+ - Tasks requiring generative capabilities.
52
 
53
  ## Bias, Risks, and Limitations
54
 
 
115
 
116
  ## Training Details
117
 
118
+ The model was trained following standard large language model (LLM) training practices for retrieval systems.
119
 
120
  ### Training Data
121
 
 
125
 
126
  #### Preprocessing
127
 
128
+ - Standard tokenization using a LLM tokenizer.
129
  - Input sequences are formatted to represent user intent (queries) and product characteristics (items).
130
  - Dynamic padding and truncation are applied to optimize training efficiency.
131
 
 
152
 
153
  ### Model Architecture and Objective
154
 
155
+ - **Backbone:** Standard LLM for text representation.
156
  - **Objective:** Contrastive learning objective
157
 
158
  ### Software