|
|
--- |
|
|
license: mpl-2.0 |
|
|
library_name: transformers |
|
|
tags: |
|
|
- gemma-3 |
|
|
- synthetic-data |
|
|
- textbooks |
|
|
- distillation |
|
|
- utility |
|
|
- summarization |
|
|
- lightning |
|
|
- conversational |
|
|
base_model: google/gemma-3-270m |
|
|
datasets: |
|
|
- TitleOS/Spark-Lightning-Synthetic-Textbooks |
|
|
language: |
|
|
- en |
|
|
pipeline_tag: text-generation |
|
|
--- |
|
|
|
|
|
# Spark-270M |
|
|
|
|
|
**Spark-270M** is a highly compact, utility-focused language model with **270 million parameters**. It is a fine-tune of Google's [Gemma 3 270M](https://huggingface.co/google/gemma-3-270m), designed to punch significantly above its weight class by leveraging high-quality synthetic data distillation. |
|
|
|
|
|
The model functions as a "dense information engine"—specializing in generating concise title summaries, search engine queries, and logical follow-up questioning—while retaining the creative conversational flair inherited from its teacher model's lineage. |
|
|
|
|
|
## ⚡ Model Details |
|
|
|
|
|
- **Model Name:** Spark-270M |
|
|
- **Base Architecture:** [Google Gemma 3 270M](https://huggingface.co/google/gemma-3-270m) |
|
|
- **Parameters:** 270M active parameters |
|
|
- **Context Window:** 32k tokens |
|
|
- **Teacher Model:** Lightning-1.7B (Custom model fine-tuned on Hermes 3) |
|
|
- **Training Type:** Synthetic "Textbook" Distillation (SFT) |
|
|
|
|
|
## 📚 Training Methodology: "Textbooks Are All You Need" |
|
|
|
|
|
Spark-270M was trained using a distinct data pipeline inspired by the *Textbooks Are All You Need* (Microsoft Phi) research paper. |
|
|
|
|
|
Instead of training on raw web scrapes, Spark-270M was fine-tuned exclusively on a series of **synthetic textbooks** generated by a larger parent model, **Lightning-1.7B**. |
|
|
|
|
|
### The Teacher: Lightning-1.7B |
|
|
The data generator, Lightning-1.7B, was itself fine-tuned on the [Hermes 3 dataset](https://huggingface.co/nousresearch/hermes-3-llama-3.1-8b). This lineage allows Spark-270M to inherit specific behavioral traits from Hermes 3—namely creativity, steerability, and a refusal to be "boring"—despite being distilled into a rigid textbook format. |
|
|
|
|
|
The synthetic data focused on: |
|
|
1. **High-density reasoning chains:** Explaining complex topics in compressed formats. |
|
|
2. **Utility Tasks:** Converting conversational fluff into actionable queries. |
|
|
3. **Socratic Dialogue:** Modeling inquisitive follow-up questioning. |
|
|
|
|
|
## 🛠️ Intended Use & Capabilities |
|
|
|
|
|
Spark-270M is designed to be a lightweight **Utility Model**. It is ideal for edge devices, rapid prototyping, or functioning as a specific "node" in a larger agentic system (e.g., the summarizer node or the query-generator node). |
|
|
|
|
|
### Primary Capabilities |
|
|
* **Dense Title Summarization:** Converting long conversation threads into information-dense, short titles or abstracts. |
|
|
* **Search Query Generation:** Formulating precise, keyword-rich search queries based on vague user input. |
|
|
* **Proactive Questioning:** Generating relevant follow-up questions to clarify user intent or deepen a topic. |
|
|
|
|
|
## 💻 Example Usage |
|
|
|
|
|
```python |
|
|
from transformers import AutoTokenizer, AutoModelForCausalLM |
|
|
|
|
|
model_id = "TitleOS/Spark-270M" |
|
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained(model_id) |
|
|
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto") |
|
|
|
|
|
# Example: Generating a search query from a user problem |
|
|
input_text = """ |
|
|
User: I need to fix my sink, it's leaking from the bottom pipe where the U-shape thing is. |
|
|
Task: Generate 3 search engine queries for this problem. |
|
|
Response: |
|
|
""" |
|
|
|
|
|
input_ids = tokenizer(input_text, return_tensors="pt").to("cuda") |
|
|
|
|
|
outputs = model.generate(**input_ids, max_new_tokens=128) |
|
|
print(tokenizer.d ecode(outputs[0])) |
|
|
``` |
|
|
|
|
|
Quants: |
|
|
|
|
|
Q4_K_M: https://huggingface.co/TitleOS/Spark-270M-FP16-Q4_K_M-GGUF |
|
|
|
|
|
Q8: https://huggingface.co/TitleOS/Spark-270M-FP16-Q8_0-GGUF |
|
|
|
|
|
FP16: https://huggingface.co/TitleOS/Spark-270M-FP16 |
|
|
|
|
|
Adaptor: https://huggingface.co/TitleOS/Spark-270M-LoRA |