BART-Large Text Summarizer (Fine-Tuned)
"Concise summaries, human-level readability."
This model is a fine-tuned version of Facebook's BART-Large-CNN, optimized to generate professional, coherent, and factually accurate summaries.
Unlike standard base models, this version was trained on a strictly filtered dataset (20,000 high-quality samples), ensuring it avoids common pitfalls like repetitive sentences or hallucinations.
Key Features
- High Precision: Achieved a Validation Loss of ~1.42 (indicating ~95% confidence/perplexity).
- Smart Truncation: Trained to handle articles between 200โ1000 words perfectly.
- No "Robot Speak": Fine-tuned to produce natural, flowing English rather than broken sentence fragments.
๐ ๏ธ How to Use (Python)
First, install the required libraries:
pip install transformers sentencepiece
from transformers import pipeline
# Load the model directly from Hugging Face
summarizer = pipeline("summarization", model="ziaulkarim245/bart-large-cnn-Text-Summarizer")
# Sample
text = """
The James Webb Space Telescope has captured a lush, highly detailed landscape โ the iconic Pillars of Creation โ
where new stars are forming within dense clouds of gas and dust. The three-dimensional pillars look like majestic
rock formations, but are far more permeable. These columns are made up of cool interstellar gas and dust that appear
semi-transparent in near-infrared light.
"""
# Summary
summary = summarizer(text, max_length=60, min_length=20, do_sample=False)
print("Summary:", summary[0]['summary_text'])
- Downloads last month
- 20
Model tree for ziaulkarim245/bart-large-cnn-Text-Summarizer
Base model
facebook/bart-large-cnn