|
--- |
|
library_name: transformers |
|
base_model: google/pegasus-xsum |
|
tags: |
|
- summarization |
|
- transformers |
|
- fine-tuned |
|
- google-pegasus-xsum |
|
- ccdv/govreport-summarization |
|
model-index: |
|
- name: a-text-summarizer |
|
results: [] |
|
language: |
|
- en |
|
--- |
|
|
|
# a-text-summarizer |
|
|
|
This model is a fine-tuned version of the google/pegasus-xsum model (https://huggingface.co/google/pegasus-xsum). |
|
It has been trained to generate summaries for governmental reports based on the GovReport summarization dataset (https://huggingface.co/datasets/ccdv/govreport-summarization). |
|
It achieves the following results on the evaluation set: |
|
- Loss: 2.3989 |
|
|
|
## Model description |
|
|
|
This is a summarization model fine-tuned on the ccdv/govreport-summarization dataset. |
|
|
|
## Intended uses & limitations |
|
|
|
This model is intended for generating concise summaries of governmental reports or similar long-form documents in an official or formal American English register. |
|
|
|
The model's performance is limited by the data it was trained on (GovReport summarization dataset). It may not generalize well to other domains or types of text. |
|
Summarization models can sometimes hallucinate information or produce summaries that are not entirely accurate. |
|
Potential biases present in the training data may be reflected in the generated summaries. Further analysis is needed to identify and mitigate potential biases. |
|
|
|
## Training and evaluation data |
|
|
|
The model was fine-tuned on a subset of the ccdv/govreport-summarization dataset. |
|
Specifically, a subset of 5000 training examples and 500 validation examples were used for fine-tuning. |
|
|
|
The GovReport dataset contains governmental reports and their corresponding summaries. |
|
|
|
## Training procedure |
|
|
|
The model was fine-tuned using the Hugging Face transformers library and Trainer API. |
|
|
|
### Training hyperparameters |
|
|
|
The following hyperparameters were used during training: |
|
- learning_rate: 2e-05 |
|
- train_batch_size: 4 |
|
- eval_batch_size: 4 |
|
- seed: 42 |
|
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments |
|
- lr_scheduler_type: linear |
|
- num_epochs: 3 |
|
|
|
### Training results |
|
|
|
| Training Loss | Epoch | Step | Validation Loss | |
|
|:-------------:|:-----:|:----:|:---------------:| |
|
| 2.7786 | 1.0 | 1250 | 2.4630 | |
|
| 2.6139 | 2.0 | 2500 | 2.4117 | |
|
| 2.5811 | 3.0 | 3750 | 2.3989 | |
|
|
|
|
|
### Framework versions |
|
|
|
- Transformers 4.55.0 |
|
- Pytorch 2.6.0+cu124 |
|
- Datasets 4.0.0 |
|
- Tokenizers 0.21.4 |