Update README.md
Browse files
README.md
CHANGED
|
@@ -1,23 +1,34 @@
|
|
| 1 |
---
|
| 2 |
license: mit
|
| 3 |
datasets:
|
| 4 |
-
- glaiveai/glaive-function-calling-v2
|
| 5 |
-
- nickrosh/Evol-Instruct-Code-80k-v1
|
| 6 |
language:
|
| 7 |
-
- en
|
| 8 |
base_model:
|
| 9 |
-
- LiquidAI/LFM2.5-1.2B-Instruct
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 10 |
---
|
| 11 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 12 |
## 📉 Training Results & Metrics
|
| 13 |
|
| 14 |
-
This model was fine-tuned on
|
| 15 |
|
| 16 |
| Metric | Value | Description |
|
| 17 |
| :--- | :--- | :--- |
|
| 18 |
| **Final Loss** | `0.7431` | The model's error rate at the final step. |
|
| 19 |
| **Average Train Loss** | `0.8274` | The average error rate across the entire session. |
|
| 20 |
-
| **Epochs** | `0.96` |
|
| 21 |
| **Global Steps** | `60` | Total number of optimizer updates. |
|
| 22 |
| **Runtime** | `594s` (~10 min) | Total wall-clock time for training. |
|
| 23 |
| **Samples/Second** | `0.808` | Throughput speed on T4 GPU. |
|
|
|
|
| 1 |
---
|
| 2 |
license: mit
|
| 3 |
datasets:
|
| 4 |
+
- glaiveai/glaive-function-calling-v2
|
| 5 |
+
- nickrosh/Evol-Instruct-Code-80k-v1
|
| 6 |
language:
|
| 7 |
+
- en
|
| 8 |
base_model:
|
| 9 |
+
- LiquidAI/LFM2.5-1.2B-Instruct
|
| 10 |
+
tags:
|
| 11 |
+
- tool-use
|
| 12 |
+
- code
|
| 13 |
+
- unsloth
|
| 14 |
+
- liquid
|
| 15 |
+
- fine-tune
|
| 16 |
+
library_name: unsloth
|
| 17 |
---
|
| 18 |
|
| 19 |
+
# 🧠 LFM-2.5-1.2B-Coding-Tools
|
| 20 |
+
|
| 21 |
+
This is a fine-tuned version of **Liquid LFM-2.5-1.2B-Instruct**, specialized for **Python coding** and **native tool calling**. It was trained using [Unsloth](https://github.com/unslothai/unsloth) on a hybrid dataset of coding instructions and Pythonic function calls.
|
| 22 |
+
|
| 23 |
## 📉 Training Results & Metrics
|
| 24 |
|
| 25 |
+
This model was fine-tuned on a Google Colab **Tesla T4** instance. The following metrics were recorded during the final training run.
|
| 26 |
|
| 27 |
| Metric | Value | Description |
|
| 28 |
| :--- | :--- | :--- |
|
| 29 |
| **Final Loss** | `0.7431` | The model's error rate at the final step. |
|
| 30 |
| **Average Train Loss** | `0.8274` | The average error rate across the entire session. |
|
| 31 |
+
| **Epochs** | `0.96` | Completed ~1 full pass over the dataset. |
|
| 32 |
| **Global Steps** | `60` | Total number of optimizer updates. |
|
| 33 |
| **Runtime** | `594s` (~10 min) | Total wall-clock time for training. |
|
| 34 |
| **Samples/Second** | `0.808` | Throughput speed on T4 GPU. |
|