Cygnis Alpha 1
Table of Contents
Model Summary
Cygnis Alpha v1 is a compact, high-performance language model based on the SmolLM2-1.7B architecture. It is designed to be lightweight enough to run on-device while maintaining high reasoning and instruction-following capabilities.
The 1.7B variant represents a significant leap over previous generations, trained on 11 trillion tokens using a high-quality mix of FineWeb-Edu, DCLM, and specialized coding/math datasets. This model has been refined using Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO) to ensure logical consistency and helpfulness.
How to use
Transformers
from transformers import AutoModelForCausalLM, AutoTokenizer
checkpoint = "cygnisai/Cygnis-Alpha-1.7B-v0.1"
device = "cuda" # for GPU usage or "cpu" for CPU usage
tokenizer = AutoTokenizer.from_pretrained(checkpoint)
model = AutoModelForCausalLM.from_pretrained(checkpoint).to(device)
messages = [{"role": "user", "content": "What is the core philosophy of sovereign AI?"}]
input_text = tokenizer.apply_chat_template(messages, tokenize=False)
inputs = tokenizer.encode(input_text, return_tensors="pt").to(device)
outputs = model.generate(inputs, max_new_tokens=100, temperature=0.2, top_p=0.9, do_sample=True)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Transformers.js
import { pipeline } from "@huggingface/transformers";
const generator = await pipeline(
"text-generation",
"cygnisai/Cygnis-Alpha-1.7B-v0.1",
);
const messages = [
{ role: "system", content: "You are Cygnis, a sovereign AI assistant." },
{ role: "user", content: "Hello!" },
];
const output = await generator(messages, { max_new_tokens: 128 });
console.log(output[0].generated_text.at(-1).content);
Evaluation
Cygnis Alpha v1 inherits the state-of-the-art benchmarks of the SmolLM2-1.7B-Instruct core. Evaluations are zero-shot using lighteval.
Instruction Model Benchmarks
| Metric | Cygnis Alpha v1 (1.7B) | Llama-1B-Instruct | Qwen2.5-1.5B-Instruct |
|---|---|---|---|
| IFEval (Avg prompt/inst) | 56.7 | 53.5 | 47.4 |
| MT-Bench | 6.13 | 5.48 | 6.52 |
| HellaSwag | 66.1 | 56.1 | 60.9 |
| ARC (Average) | 51.7 | 41.6 | 46.2 |
| GSM8K (5-shot) | 48.2 | 26.8 | 42.8 |
Examples
Text Rewriting
System Prompt: "You are an AI writing assistant. Your task is to rewrite the user's email to make it more professional and approachable while maintaining its main points and key message."
Function Calling
Cygnis Alpha v1 supports tool-use and function calling. It scores 27% on the BFCL Leaderboard, allowing it to interface with APIs to fetch real-time data or perform calculations.
Limitations
Cygnis Alpha v1 Instruct primarily understands and generates content in English. While powerful for its size, it may produce factually inaccurate or logically inconsistent content for highly specialized tasks. Users should verify important information and use it as an assistive tool.
Training
Model Specifications
- Architecture: Transformer decoder
- Pretraining tokens: 11T
- Precision: bfloat16
- Hardware: 256 H100 GPUs (original base training)
Software
- Training Framework: nanotron
- Alignment: alignment-handbook
License
This model is licensed under Apache 2.0.
Citation
@misc{cygnis_alpha_1.7b_v0.1,
title={SmolLM2: When Smol Goes Big -- Data-Centric Training of a Small Language Model},
author={Simonc-44},
year={2026},
eprint={2502.02737},
archivePrefix={arXiv},
}
- Downloads last month
- 1,207