Update README.md
Browse files
README.md
CHANGED
|
@@ -15,15 +15,26 @@ base_model:
|
|
| 15 |
pipeline_tag: sentence-similarity
|
| 16 |
---
|
| 17 |
|
| 18 |
-
# Zarra Arabic Static Embedding
|
| 19 |
|
| 20 |
-
|
|
|
|
|
|
|
| 21 |
|
| 22 |
-
|
|
|
|
| 23 |
|
| 24 |
-
|
| 25 |
-
|
| 26 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 27 |
|
| 28 |
## Installation
|
| 29 |
|
|
|
|
| 15 |
pipeline_tag: sentence-similarity
|
| 16 |
---
|
| 17 |
|
| 18 |
+
# Zarra: Arabic Static Embedding Model
|
| 19 |
|
| 20 |
+
**Zarra** is a static embedding model built using the Model2Vec distillation framework.
|
| 21 |
+
It is a distilled version of a Sentence Transformer, specifically optimized for the Arabic language.
|
| 22 |
+
Unlike traditional transformer-based models, Zarra produces static embeddings, enabling ultra-fast inference on both CPU and GPU—making it ideal for resource-constrained environments or real-time applications.
|
| 23 |
|
| 24 |
+
## Why Zarra?
|
| 25 |
+
⚡ Exceptional Speed: Delivers embeddings up to 500x faster than sentence transformers.
|
| 26 |
|
| 27 |
+
🧠 Compact & Efficient: Up to 50x smaller in size, allowing easy deployment on edge devices.
|
| 28 |
+
|
| 29 |
+
🧰 Versatile: Well-suited for search, clustering, classification, deduplication, and more.
|
| 30 |
+
|
| 31 |
+
🌍 Arabic-First: Specifically trained on high-quality Arabic data, ensuring relevance and performance across a range of Arabic NLP tasks.
|
| 32 |
+
|
| 33 |
+
## About Model2Vec
|
| 34 |
+
|
| 35 |
+
|
| 36 |
+
The Model2Vec distillation technique transfers knowledge from large transformer models into lightweight static embedding spaces, preserving semantic quality while dramatically improving speed and efficiency.
|
| 37 |
+
Zarra represents the best of both worlds: the semantic power of transformers and the speed and simplicity of static vectors.
|
| 38 |
|
| 39 |
## Installation
|
| 40 |
|