---
license: openrail
language:
- en
- ja
- de
- zh
model-index:
- name: Socrates-embedding
results:
- task:
type: classification
name: Multilingual Classification
dataset:
name: AmazonCounterfactualClassification
type: mteb/amazon_counterfactual
metrics:
- name: Accuracy (Japanese)
type: accuracy
value: 54.83
- name: Accuracy (German)
type: accuracy
value: 52.57
- name: Accuracy (English)
type: accuracy
value: 49.70
- name: Accuracy (English-Ext)
type: accuracy
value: 49.15
- task:
type: clustering
name: Clustering
dataset:
name: StackExchangeClustering
type: mteb/stackexchange_clustering
metrics:
- name: V-measure x 100
type: v_measure
value: 8.92
co2_footprint:
emissions: 0.17 # in KgCO2eq
source: "Estimated based on 1.2 hours of training on a single NVIDIA RTX 6000 (TDP ~300W)."
training_type: "from_scratch"
geographical_location: "Zaozhuang, China"
hardware_used: "1 x NVIDIA RTX 6000"
training_duration: 1.2 # in hours
---
# Model Card for Socrates-embedding
> Note: This model was stopped during the training process after one epoch because we were unable to afford the cost of AutoDL. The current version shows the result after training with nil for 18,000 steps.
  
## Model Details
Socrates-embedding is a lightweight, high-density text embedding model. Unlike contemporary models that rely on massive parameter counts to brute-force semantic understanding, Socrates-embedding leverages Low-Rank Decay (LoRD) to achieve high-quality vector representations with minimal computational overhead.
This model is part of the Chunjiang Intelligence edge-computing initiative, aiming to bring retrieval-augmented generation (RAG) and semantic search capabilities to consumer-grade hardware.
- **Developed by:** Chunjiang Intelligence
- **Model Type:** Dual-Encoder Transformer.
- **Language:** English, Japanese, German, Chinese
The model was evaluated on the `AmazonCounterfactualClassification` dataset across multiple languages.
| Language | Accuracy |
| :--- | :---: |
| Japanese (ja) | 54.83 |
| German (de) | 52.57 |
| English (en) | 49.70 |
| English-Ext (en-ext)| 49.15 |
To put the model's efficiency into perspective, we compare its single-task score on Japanese classification against the *overall MTEB average scores* of much larger models. (Our budget is insufficient to cover the bill for the GPU used for the ongoing tests.)
Figure 1: Our 83M model's score on a single challenging task rivals the average performance of models up to 85x larger.
Figure 2: Leading clustering performance among lightweight embedding models.