Update README.md
Browse files
README.md
CHANGED
|
@@ -13,7 +13,7 @@ language:
|
|
| 13 |
|
| 14 |
# ReasoningCore-Llama-3B-R1-aligned
|
| 15 |
|
| 16 |
-
**ReasoningCore
|
| 17 |
|
| 18 |
### We used GRPO technique:
|
| 19 |
|
|
@@ -34,7 +34,7 @@ To provide a comprehensive overview of Group Relative Policy Optimization (GRPO)
|
|
| 34 |
|
| 35 |
| | Training Data | Params | Input Modalities | Output Modalities | Context Length | GQA | Shared Embeddings | Token Count | Knowledge Cutoff |
|
| 36 |
|--------------------------------|--------------------------------------------------|--------|-----------------------|------------------------------|----------------|-----|-------------------|----------------|-------------------|
|
| 37 |
-
| **ReasoningCore
|
| 38 |
|
| 39 |
- **Supported Languages:**
|
| 40 |
Officially supports English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai. While the pretraining included a broader range of languages, additional languages can be fine‑tuned in compliance with the community license and acceptable use policies.
|
|
|
|
| 13 |
|
| 14 |
# ReasoningCore-Llama-3B-R1-aligned
|
| 15 |
|
| 16 |
+
**ReasoningCore-Llama-3B-R1-aligned** is a multilingual, reasoning‑enhanced large language model developed by EpitemeAI. Pretrained on vast amounts of publicly available data and instruction‑tuned to excel at nuanced reasoning, dialogue management, retrieval, and summarization tasks, it often outperforms many current open source and proprietary conversational models on a range of industry benchmarks. Fine tuned with reasoning dataset.
|
| 17 |
|
| 18 |
### We used GRPO technique:
|
| 19 |
|
|
|
|
| 34 |
|
| 35 |
| | Training Data | Params | Input Modalities | Output Modalities | Context Length | GQA | Shared Embeddings | Token Count | Knowledge Cutoff |
|
| 36 |
|--------------------------------|--------------------------------------------------|--------|-----------------------|------------------------------|----------------|-----|-------------------|----------------|-------------------|
|
| 37 |
+
| **ReasoningCore-Llama-3B-R1-aligned (text only)** | A new mix of publicly available online data. | 3B | Multilingual Text | Multilingual Text and code | 128k | Yes | Yes | Up to 9T tokens | December 2023 |
|
| 38 |
|
| 39 |
- **Supported Languages:**
|
| 40 |
Officially supports English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai. While the pretraining included a broader range of languages, additional languages can be fine‑tuned in compliance with the community license and acceptable use policies.
|