DistilBART CNN 12-6 โ€” ONNX

ONNX export of DistilBART CNN 12-6 (306M parameters) with encoder-decoder architecture and KV cache support.

A distilled version of BART Large CNN, fine-tuned for text summarization on the CNN/DailyMail dataset.

Converted for use with inference4j, an inference-only AI library for Java.

Original Source

Usage with inference4j

try (var summarizer = BartSummarizer.distilBartCnn().build()) {
    System.out.println(summarizer.summarize("Long article text..."));
}

Model Details

Property Value
Architecture BART encoder-decoder (306M parameters, 12 encoder + 6 decoder layers)
Task Text summarization
Training data CNN/DailyMail
Tokenizer BPE (50,265 tokens)
Original framework PyTorch (transformers)
Export method Hugging Face Optimum (encoder-decoder with KV cache)

License

This model is licensed under the Apache License 2.0. Original model by Sam Shleifer.

Downloads last month
13
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support