a-text-summarizer / README.md
nppiech's picture
Update README.md
162b6cc verified
metadata
library_name: transformers
base_model: google/pegasus-xsum
tags:
  - summarization
  - transformers
  - fine-tuned
  - google-pegasus-xsum
  - ccdv/govreport-summarization
model-index:
  - name: a-text-summarizer
    results: []
language:
  - en

a-text-summarizer

This model is a fine-tuned version of the google/pegasus-xsum model (https://huggingface.co/google/pegasus-xsum). It has been trained to generate summaries for governmental reports based on the GovReport summarization dataset (https://huggingface.co/datasets/ccdv/govreport-summarization). It achieves the following results on the evaluation set:

  • Loss: 2.3989

Model description

This is a summarization model fine-tuned on the ccdv/govreport-summarization dataset.

Intended uses & limitations

This model is intended for generating concise summaries of governmental reports or similar long-form documents in an official or formal American English register.

The model's performance is limited by the data it was trained on (GovReport summarization dataset). It may not generalize well to other domains or types of text. Summarization models can sometimes hallucinate information or produce summaries that are not entirely accurate. Potential biases present in the training data may be reflected in the generated summaries. Further analysis is needed to identify and mitigate potential biases.

Training and evaluation data

The model was fine-tuned on a subset of the ccdv/govreport-summarization dataset. Specifically, a subset of 5000 training examples and 500 validation examples were used for fine-tuning.

The GovReport dataset contains governmental reports and their corresponding summaries.

Training procedure

The model was fine-tuned using the Hugging Face transformers library and Trainer API.

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 4
  • eval_batch_size: 4
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • num_epochs: 3

Training results

Training Loss Epoch Step Validation Loss
2.7786 1.0 1250 2.4630
2.6139 2.0 2500 2.4117
2.5811 3.0 3750 2.3989

Framework versions

  • Transformers 4.55.0
  • Pytorch 2.6.0+cu124
  • Datasets 4.0.0
  • Tokenizers 0.21.4