samsum_t5-small / README.md
jialicheng's picture
Upload folder using huggingface_hub
1b56f99 verified
metadata
license: apache-2.0
base_model: google/t5-v1_1-small
tags:
  - generated_from_trainer
datasets:
  - samsum
metrics:
  - rouge
model-index:
  - name: samsum_42
    results:
      - task:
          name: Summarization
          type: summarization
        dataset:
          name: samsum
          type: samsum
        metrics:
          - name: Rouge1
            type: rouge
            value: 38.9743

samsum_42

This model is a fine-tuned version of google/t5-v1_1-small on the samsum dataset. It achieves the following results on the evaluation set:

  • Loss: 2.1130
  • Rouge1: 38.9743
  • Rouge2: 17.0483
  • Rougel: 32.8088
  • Rougelsum: 36.1047
  • Gen Len: 20.1785

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 16
  • eval_batch_size: 64
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 5

Training results

Training Loss Epoch Step Gen Len Validation Loss Rouge1 Rouge2 Rougel Rougelsum
No log 1.0 461 14.9658 2.3319 30.3301 13.4627 26.1544 28.1275
No log 2.0 922 17.4230 2.1737 35.9382 15.9281 30.5125 33.3475
3.9209 3.0 1383 19.5892 2.1288 37.9339 16.6056 31.9002 35.0597
3.9209 4.0 1844 20.1381 2.1201 38.625 16.8095 32.5753 35.825
2.9358 5.0 2305 20.1785 2.1130 38.9743 17.0483 32.8088 36.1047

Framework versions

  • Transformers 4.39.3
  • Pytorch 2.3.0+cu121
  • Datasets 2.19.0
  • Tokenizers 0.15.2