Update README.md
Browse files
README.md
CHANGED
|
@@ -18,7 +18,7 @@ library_name: transformers
|
|
| 18 |
|
| 19 |
## Introduction
|
| 20 |
|
| 21 |
-
**STAR-0b6** is a highly capable 0.6B parameter language model specialized in function calling, achieving
|
| 22 |
|
| 23 |
This model is the result of fine-tuning the `Qwen/Qwen3-0.6B` base model using the novel **STAR (Similarity-guided Teacher-Assisted Refinement)** framework. STAR is a holistic training curriculum designed to effectively transfer the advanced capabilities of large language models (LLMs) into "super-tiny" models, making them powerful, accessible, and efficient for real-world agentic applications.
|
| 24 |
|
|
@@ -100,7 +100,7 @@ For local use, applications such as Ollama, LMStudio, MLX-LM, llama.cpp, and KTr
|
|
| 100 |
|
| 101 |
## Evaluation & Performance
|
| 102 |
|
| 103 |
-
STAR-0b6 has
|
| 104 |
|
| 105 |
- BFCLv3: Achieved 51.70% overall accuracy, outperforming all baseline and recent methods.
|
| 106 |
- ACEBench: Achieved 53.00% summary score, demonstrating superior generalization and robustness. This score is significantly higher than its base model (27.20%) and even surpasses much larger models like Llama3.1-8B (46.60%).
|
|
|
|
| 18 |
|
| 19 |
## Introduction
|
| 20 |
|
| 21 |
+
**STAR-0b6** is a highly capable 0.6B parameter language model specialized in function calling, achieving excellent performances on the [Berkeley Function Calling Leaderboard (BFCL)](https://huggingface.co/spaces/gorilla-llm/berkeley-function-calling-leaderboard) for models in its size class.
|
| 22 |
|
| 23 |
This model is the result of fine-tuning the `Qwen/Qwen3-0.6B` base model using the novel **STAR (Similarity-guided Teacher-Assisted Refinement)** framework. STAR is a holistic training curriculum designed to effectively transfer the advanced capabilities of large language models (LLMs) into "super-tiny" models, making them powerful, accessible, and efficient for real-world agentic applications.
|
| 24 |
|
|
|
|
| 100 |
|
| 101 |
## Evaluation & Performance
|
| 102 |
|
| 103 |
+
STAR-0b6 has achieved outstanding performance for models of its size on renowned function calling benchmarks.
|
| 104 |
|
| 105 |
- BFCLv3: Achieved 51.70% overall accuracy, outperforming all baseline and recent methods.
|
| 106 |
- ACEBench: Achieved 53.00% summary score, demonstrating superior generalization and robustness. This score is significantly higher than its base model (27.20%) and even surpasses much larger models like Llama3.1-8B (46.60%).
|