AbdulElahGwaith's picture
Upload folder using huggingface_hub
a9bd396 verified

This model was released on 2025-09-02 and added to Hugging Face Transformers on 2025-08-28.

Apertus

PyTorch FlashAttention SDPA Tensor parallelism

Overview

Apertus is a family of large language models from the Swiss AI Initiative.

Coming soon

The example below demonstrates how to generate text with [Pipeline] or the [AutoModel], and from the command line.

import torch
from transformers import pipeline

pipeline = pipeline(
    task="text-generation",
    model="swiss-ai/Apertus-8B",
    dtype=torch.bfloat16,
    device=0
)
pipeline("Plants create energy through a process known as")
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer

tokenizer = AutoTokenizer.from_pretrained(
    "swiss-ai/Apertus-8B",
)
model = AutoModelForCausalLM.from_pretrained(
    "swiss-ai/Apertus-8B",
    dtype=torch.bfloat16,
    device_map="auto",
    attn_implementation="sdpa"
)
input_ids = tokenizer("Plants create energy through a process known as", return_tensors="pt").to("cuda")

output = model.generate(**input_ids)
print(tokenizer.decode(output[0], skip_special_tokens=True))
echo -e "Plants create energy through a process known as" | transformers run --task text-generation --model swiss-ai/Apertus-8B --device 0

ApertusConfig

[[autodoc]] ApertusConfig

ApertusModel

[[autodoc]] ApertusModel - forward

ApertusForCausalLM

[[autodoc]] ApertusForCausalLM - forward

ApertusForTokenClassification

[[autodoc]] ApertusForTokenClassification - forward