Tigdora commited on
Commit
2078953
·
verified ·
1 Parent(s): 23004d2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -6
README.md CHANGED
@@ -1,23 +1,34 @@
1
  ---
2
  license: mit
3
  datasets:
4
- - glaiveai/glaive-function-calling-v2
5
- - nickrosh/Evol-Instruct-Code-80k-v1
6
  language:
7
- - en
8
  base_model:
9
- - LiquidAI/LFM2.5-1.2B-Instruct
 
 
 
 
 
 
 
10
  ---
11
 
 
 
 
 
12
  ## 📉 Training Results & Metrics
13
 
14
- This model was fine-tuned on **Liquid LFM-2.5-1.2B-Instruct** using **Unsloth** and T4 GPUs. The following metrics were recorded during the final training run.
15
 
16
  | Metric | Value | Description |
17
  | :--- | :--- | :--- |
18
  | **Final Loss** | `0.7431` | The model's error rate at the final step. |
19
  | **Average Train Loss** | `0.8274` | The average error rate across the entire session. |
20
- | **Epochs** | `0.96` | completed ~1 full pass over the dataset. |
21
  | **Global Steps** | `60` | Total number of optimizer updates. |
22
  | **Runtime** | `594s` (~10 min) | Total wall-clock time for training. |
23
  | **Samples/Second** | `0.808` | Throughput speed on T4 GPU. |
 
1
  ---
2
  license: mit
3
  datasets:
4
+ - glaiveai/glaive-function-calling-v2
5
+ - nickrosh/Evol-Instruct-Code-80k-v1
6
  language:
7
+ - en
8
  base_model:
9
+ - LiquidAI/LFM2.5-1.2B-Instruct
10
+ tags:
11
+ - tool-use
12
+ - code
13
+ - unsloth
14
+ - liquid
15
+ - fine-tune
16
+ library_name: unsloth
17
  ---
18
 
19
+ # 🧠 LFM-2.5-1.2B-Coding-Tools
20
+
21
+ This is a fine-tuned version of **Liquid LFM-2.5-1.2B-Instruct**, specialized for **Python coding** and **native tool calling**. It was trained using [Unsloth](https://github.com/unslothai/unsloth) on a hybrid dataset of coding instructions and Pythonic function calls.
22
+
23
  ## 📉 Training Results & Metrics
24
 
25
+ This model was fine-tuned on a Google Colab **Tesla T4** instance. The following metrics were recorded during the final training run.
26
 
27
  | Metric | Value | Description |
28
  | :--- | :--- | :--- |
29
  | **Final Loss** | `0.7431` | The model's error rate at the final step. |
30
  | **Average Train Loss** | `0.8274` | The average error rate across the entire session. |
31
+ | **Epochs** | `0.96` | Completed ~1 full pass over the dataset. |
32
  | **Global Steps** | `60` | Total number of optimizer updates. |
33
  | **Runtime** | `594s` (~10 min) | Total wall-clock time for training. |
34
  | **Samples/Second** | `0.808` | Throughput speed on T4 GPU. |