Fill-Mask
Transformers
Safetensors
gemma3_text
feature-extraction
ecommerce
e-commerce
retail
marketplace
shopping
amazon
ebay
alibaba
google
rakuten
bestbuy
walmart
flipkart
wayfair
shein
target
etsy
shopify
taobao
asos
carrefour
costco
overstock
pretraining
encoder
language-modeling
foundation-model
custom_code
text-generation-inference
Upload folder using huggingface_hub
Browse files- README.md +1 -1
- config.json +1 -1
README.md
CHANGED
|
@@ -32,7 +32,7 @@ This model is a BERT-style bidirectional encoder based on Gemma 3 architecture:
|
|
| 32 |
| Base Model | [`thebajajra/Gemma3-270M-encoder`](https://huggingface.co/thebajajra/Gemma3-270M-encoder) |
|
| 33 |
| Vocab Size | 262,145 |
|
| 34 |
| Sliding Window | 512 |
|
| 35 |
-
| Max Sequence Length |
|
| 36 |
| Attention | Bidirectional |
|
| 37 |
|
| 38 |
## Usage
|
|
|
|
| 32 |
| Base Model | [`thebajajra/Gemma3-270M-encoder`](https://huggingface.co/thebajajra/Gemma3-270M-encoder) |
|
| 33 |
| Vocab Size | 262,145 |
|
| 34 |
| Sliding Window | 512 |
|
| 35 |
+
| Max Sequence Length | 2048 |
|
| 36 |
| Attention | Bidirectional |
|
| 37 |
|
| 38 |
## Usage
|
config.json
CHANGED
|
@@ -41,7 +41,7 @@
|
|
| 41 |
"sliding_attention",
|
| 42 |
"full_attention"
|
| 43 |
],
|
| 44 |
-
"max_position_embeddings":
|
| 45 |
"model_type": "gemma3_text",
|
| 46 |
"num_attention_heads": 4,
|
| 47 |
"num_hidden_layers": 18,
|
|
|
|
| 41 |
"sliding_attention",
|
| 42 |
"full_attention"
|
| 43 |
],
|
| 44 |
+
"max_position_embeddings": 2048,
|
| 45 |
"model_type": "gemma3_text",
|
| 46 |
"num_attention_heads": 4,
|
| 47 |
"num_hidden_layers": 18,
|