🧠 General Text Summarizer

This model is a fine-tuned version of sshleifer/distilbart-cnn-12-6, trained to generate concise and fluent summaries of general English text β€” including news articles, essays, stories, and blog posts.


πŸš€ Model Description

  • Base model: DistilBART (CNN/DailyMail)
  • Framework: πŸ€— Transformers (PyTorch)
  • Training goal: Summarize text across multiple domains (not limited to one topic)
  • Device optimized: CPU & Apple M-series chips (MPS compatible)

This model is suitable for lightweight summarization tasks on laptops or limited-resource machines.


🧾 Example Usage

from transformers import pipeline

summarizer = pipeline("summarization", model="Fathi7ma/general_text_summarizer_cpu")

text = """ Climate change continues to affect weather patterns across the globe. Scientists warn that without immediate action, rising temperatures may lead to irreversible damage to ecosystems and human livelihoods. """

summary = summarizer(text, max_length=80, min_length=25, do_sample=False) print(summary[0]['summary_text'])

Intended uses

This model can summarize: β€’ News articles β€’ Research abstracts β€’ Reports and blogs β€’ Long paragraphs of general English text

Example domains: general news, education, business summaries, and everyday content.

Training

β€’	Dataset: A subset of CNN/DailyMail, filtered and balanced for general summarization.
β€’	Approx. 10,000 samples used for CPU-efficient fine-tuning.
β€’	Texts are trimmed and normalized for readability.

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 1
  • eval_batch_size: 1
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • num_epochs: 1

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum
2.2534 1.0 600 2.1023 36.61 16.51 26.24 33.45

Framework versions

  • Transformers 4.57.1
  • Pytorch 2.9.0
  • Datasets 4.3.0
  • Tokenizers 0.22.1
Downloads last month
8
Safetensors
Model size
0.3B params
Tensor type
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for Fathi7ma/news_text_summarizer

Finetuned
(41)
this model

Evaluation results