Novachrono93 commited on
Commit
5e4fa6d
Β·
verified Β·
1 Parent(s): 6e09239

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +101 -11
README.md CHANGED
@@ -1,21 +1,111 @@
1
  ---
2
- base_model: unsloth/LFM2.5-1.2B-Instruct
 
3
  tags:
4
- - text-generation-inference
5
- - transformers
6
  - unsloth
7
- - lfm2
8
- license: apache-2.0
 
 
 
 
 
9
  language:
10
  - en
11
  ---
12
 
13
- # Uploaded finetuned model
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
- - **Developed by:** NovachronoAI
16
- - **License:** apache-2.0
17
- - **Finetuned from model :** unsloth/LFM2.5-1.2B-Instruct
 
18
 
19
- This lfm2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
 
 
 
20
 
21
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: apache-2.0
3
+ base_model: LiquidAI/LFM2.5-1.2B-Instruct
4
  tags:
5
+ - function-calling
6
+ - liquid-neural-network
7
  - unsloth
8
+ - tool-use
9
+ - gguf
10
+ - conversational
11
+ datasets:
12
+ - NovachronoAI/Nova-Synapse-Function-Calling
13
+ library_name: transformers
14
+ pipeline_tag: text-generation
15
  language:
16
  - en
17
  ---
18
 
19
+ # 🌊 LFM 2.5 1.2B - Nova Synapse (Function Calling)
20
+
21
+ <div align="center">
22
+
23
+ ![Unsloth Fine-tuning](https://img.shields.io/badge/Fine--Tuned%20with-Unsloth-blue?style=for-the-badge)
24
+ ![Liquid AI](https://img.shields.io/badge/Architecture-Liquid%20Neural%20Network-cyan?style=for-the-badge)
25
+ ![Function Calling](https://img.shields.io/badge/Task-SOTA%20Function%20Calling-orange?style=for-the-badge)
26
+ ![Size](https://img.shields.io/badge/Params-1.2B-green?style=for-the-badge)
27
+
28
+ </div>
29
+
30
+ ## πŸš€ Model Overview
31
+ **LFM2.5-1.2B-Nova-Function-Calling** is a specialized fine-tune of Liquid AI's revolutionary **Liquid Neural Network (LFM 2.5)**. Despite its small size (1.2B parameters), this model rivals 7B+ class models in specific tasks due to its hybrid architecture.
32
+
33
+ This model has been **specifically engineered for robust Function Calling**, allowing it to seamlessly convert natural language user queries into structured JSON inputs for tools, APIs, and software agents.
34
+
35
+ ### 🌟 Key Features
36
+ * **Hyper-Efficient:** Runs on extremely low-resource hardware (phones, Raspberry Pi, older laptops) thanks to the 1.2B Liquid architecture.
37
+ * **Precision Tuned:** Achieved a training loss of **2.63**, mastering structured JSON syntax without overfitting.
38
+ * **ChatML Native:** Uses the standard `<|im_start|>` format for easy integration.
39
+ * **GGUF Ready:** Available in all quantization levels (from 16-bit down to 2-bit).
40
+
41
+ ---
42
+
43
+ ## πŸ“š Dataset
44
+ This model was trained on **[NovachronoAI/Nova-Synapse-Function-Calling](https://huggingface.co/datasets/NovachronoAI/Nova-Synapse-Function-Calling)**.
45
+
46
+ * **Source:** A massive collection of 130k+ examples of complex user-agent interactions involving tool usage.
47
+ * **Selection:** A curated subset of 15,000 high-complexity examples was selected to maximize syntax learning while preventing catastrophic forgetting.
48
+ * **Focus:** The dataset emphasizes correct JSON schema adherence, argument extraction, and tool selection logic.
49
+
50
+ ---
51
+
52
+ ## πŸ’» Quick Start (Inference)
53
+
54
+ ### 1. Using Transformers
55
+ You need the latest `transformers` and `unsloth` libraries to run Liquid architectures.
56
+
57
+ ```python
58
+ from unsloth import FastLanguageModel
59
+ import torch
60
+
61
+ # Load the model
62
+ model, tokenizer = FastLanguageModel.from_pretrained(
63
+ model_name = "NovachronoAI/LFM2.5-1.2B-Nova-Function-Calling-Full", # or use the GGUF repo
64
+ max_seq_length = 4096,
65
+ dtype = None,
66
+ load_in_4bit = True,
67
+ )
68
+ FastLanguageModel.for_inference(model)
69
+
70
+ # Define the Prompt (ChatML Format)
71
+ prompt = """<|im_start|>user
72
+ I need to calculate the area of a circle with a radius of 5.
73
+ <|im_end|>
74
+ <|im_start|>assistant
75
+ """
76
 
77
+ # Generate
78
+ inputs = tokenizer([prompt], return_tensors = "pt").to("cuda")
79
+ outputs = model.generate(**inputs, max_new_tokens = 128, use_cache = True)
80
+ print(tokenizer.batch_decode(outputs)[0].split("<|im_start|>assistant")[-1])
81
 
82
+ Expected Output:
83
+ <tool_call>
84
+ {"name": "calculate_circle_area", "arguments": {"radius": 5}}
85
+ </tool_call>
86
 
87
+ 2. Using GGUF (llama.cpp / Ollama)
88
+ This model is available in GGUF format in the companion repository: NovachronoAI/LFM2.5-1.2B-Nova-Function-Calling-GGUF
89
+ * Recommended: q4_k_m.gguf (Balanced Speed/Quality - ~800MB)
90
+ * Max Quality: f16.gguf (Lossless - ~2.5GB)
91
+ * Max Speed: q2_k.gguf (Extreme Speed - ~400MB)
92
+ βš™οΈ Training Details
93
+ | Parameter | Value |
94
+ |---|---|
95
+ | Base Model | LiquidAI/LFM2.5-1.2B-Instruct |
96
+ | Framework | Unsloth + Hugging Face TRL |
97
+ | Hardware | NVIDIA Tesla T4 (Kaggle) |
98
+ | Epochs | ~2 (600 Steps) |
99
+ | Learning Rate | 2e-4 |
100
+ | Scheduler | Linear |
101
+ | Quantization | 4-bit (QLoRA) |
102
+ Training Trajectory
103
+ The model showed rapid adaptation to the JSON syntax, dropping from a random-guess loss of 11.6 to a highly capable 2.63.
104
+ * Start: Loss 11.68 (Step 10)
105
+ * Convergence: Loss ~3.0 (Step 160)
106
+ * Final: Loss 2.63 (Step 600)
107
+ πŸ“œ License
108
+ This model is fine-tuned from LiquidAI/LFM2.5-1.2B-Instruct. Please refer to the original Liquid AI license terms for commercial use. The fine-tuning dataset and adapters are released under Apache 2.0.
109
+ <div align="center">
110
+ Built with ❀️ by <b>NovachronoAI</b> using <a href="https://github.com/unslothai/unsloth">Unsloth</a>
111
+ </div>