--- language: en library_name: transformers license: apache-2.0 base_model: sshleifer/distilbart-cnn-12-6 tags: - summarization - text-generation - fine-tuned-model - bart model-index: - name: General Text Summarizer results: - task: type: summarization name: Text Summarization dataset: name: CNN/DailyMail type: cnn_dailymail metrics: - name: Rouge1 type: rouge value: 36.61 - name: Rouge2 type: rouge value: 16.51 - name: RougeL type: rouge value: 26.24 - name: RougeLsum type: rouge value: 33.45 --- # 🧠 General Text Summarizer This model is a fine-tuned version of [sshleifer/distilbart-cnn-12-6](https://huggingface.co/sshleifer/distilbart-cnn-12-6), trained to generate **concise and fluent summaries** of general English text — including **news articles, essays, stories, and blog posts**. --- ## 🚀 Model Description - **Base model:** DistilBART (CNN/DailyMail) - **Framework:** 🤗 Transformers (PyTorch) - **Training goal:** Summarize text across multiple domains (not limited to one topic) - **Device optimized:** CPU & Apple M-series chips (MPS compatible) This model is suitable for lightweight summarization tasks on laptops or limited-resource machines. --- ## 🧾 Example Usage from transformers import pipeline summarizer = pipeline("summarization", model="Fathi7ma/general_text_summarizer_cpu") text = """ Climate change continues to affect weather patterns across the globe. Scientists warn that without immediate action, rising temperatures may lead to irreversible damage to ecosystems and human livelihoods. """ summary = summarizer(text, max_length=80, min_length=25, do_sample=False) print(summary[0]['summary_text']) ## Intended uses This model can summarize: • News articles • Research abstracts • Reports and blogs • Long paragraphs of general English text Example domains: general news, education, business summaries, and everyday content. ## Training • Dataset: A subset of CNN/DailyMail, filtered and balanced for general summarization. • Approx. 10,000 samples used for CPU-efficient fine-tuning. • Texts are trimmed and normalized for readability. ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 1 - eval_batch_size: 1 - seed: 42 - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:| | 2.2534 | 1.0 | 600 | 2.1023 | 36.61 | 16.51 | 26.24 | 33.45 | ### Framework versions - Transformers 4.57.1 - Pytorch 2.9.0 - Datasets 4.3.0 - Tokenizers 0.22.1