Cygnis Alpha 1

Table of Contents

  1. Model Summary
  2. Evaluation
  3. Examples
  4. Limitations
  5. Training
  6. License
  7. Citation

Model Summary

Cygnis Alpha v1 is a compact, high-performance language model based on the SmolLM2-1.7B architecture. It is designed to be lightweight enough to run on-device while maintaining high reasoning and instruction-following capabilities.

The 1.7B variant represents a significant leap over previous generations, trained on 11 trillion tokens using a high-quality mix of FineWeb-Edu, DCLM, and specialized coding/math datasets. This model has been refined using Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO) to ensure logical consistency and helpfulness.

How to use

Transformers

from transformers import AutoModelForCausalLM, AutoTokenizer

checkpoint = "cygnisai/Cygnis-Alpha-1.7B-v0.1"
device = "cuda" # for GPU usage or "cpu" for CPU usage

tokenizer = AutoTokenizer.from_pretrained(checkpoint)
model = AutoModelForCausalLM.from_pretrained(checkpoint).to(device)

messages = [{"role": "user", "content": "What is the core philosophy of sovereign AI?"}]
input_text = tokenizer.apply_chat_template(messages, tokenize=False)
inputs = tokenizer.encode(input_text, return_tensors="pt").to(device)
outputs = model.generate(inputs, max_new_tokens=100, temperature=0.2, top_p=0.9, do_sample=True)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Transformers.js

import { pipeline } from "@huggingface/transformers";

const generator = await pipeline(
  "text-generation",
  "cygnisai/Cygnis-Alpha-1.7B-v0.1",
);

const messages = [
  { role: "system", content: "You are Cygnis, a sovereign AI assistant." },
  { role: "user", content: "Hello!" },
];

const output = await generator(messages, { max_new_tokens: 128 });
console.log(output[0].generated_text.at(-1).content);

Evaluation

Cygnis Alpha v1 inherits the state-of-the-art benchmarks of the SmolLM2-1.7B-Instruct core. Evaluations are zero-shot using lighteval.

Instruction Model Benchmarks

Metric Cygnis Alpha v1 (1.7B) Llama-1B-Instruct Qwen2.5-1.5B-Instruct
IFEval (Avg prompt/inst) 56.7 53.5 47.4
MT-Bench 6.13 5.48 6.52
HellaSwag 66.1 56.1 60.9
ARC (Average) 51.7 41.6 46.2
GSM8K (5-shot) 48.2 26.8 42.8

Examples

Text Rewriting

System Prompt: "You are an AI writing assistant. Your task is to rewrite the user's email to make it more professional and approachable while maintaining its main points and key message."

Function Calling

Cygnis Alpha v1 supports tool-use and function calling. It scores 27% on the BFCL Leaderboard, allowing it to interface with APIs to fetch real-time data or perform calculations.


Limitations

Cygnis Alpha v1 Instruct primarily understands and generates content in English. While powerful for its size, it may produce factually inaccurate or logically inconsistent content for highly specialized tasks. Users should verify important information and use it as an assistive tool.

Training

Model Specifications

  • Architecture: Transformer decoder
  • Pretraining tokens: 11T
  • Precision: bfloat16
  • Hardware: 256 H100 GPUs (original base training)

Software

License

This model is licensed under Apache 2.0.

Citation

@misc{cygnis_alpha_1.7b_v0.1,
      title={SmolLM2: When Smol Goes Big -- Data-Centric Training of a Small Language Model}, 
      author={Simonc-44},
      year={2026},
      eprint={2502.02737},
      archivePrefix={arXiv},
}
Downloads last month
1,207
Safetensors
Model size
2B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for cygnisai/Cygnis-Alpha-1.7B-v0.1

Quantized
(35)
this model
Quantizations
3 models

Collection including cygnisai/Cygnis-Alpha-1.7B-v0.1

Paper for cygnisai/Cygnis-Alpha-1.7B-v0.1