--- license: apache-2.0 base_model: google/flan-t5-small tags: - generated_from_trainer metrics: - rouge model-index: - name: flan-t5-small_twon-debug-generative-agent results: [] --- # flan-t5-small_twon-debug-generative-agent This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 3.8511 - Rouge1: 9.4574 - Rouge2: 3.2143 - Rougel: 8.6333 - Rougelsum: 8.8052 - Gen Len: 18.99 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 4e-05 - train_batch_size: 24 - eval_batch_size: 24 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:| | 4.2488 | 1.19 | 100 | 3.9027 | 9.2555 | 2.9689 | 8.4655 | 8.5977 | 18.988 | | 4.0882 | 2.38 | 200 | 3.8511 | 9.4574 | 3.2143 | 8.6333 | 8.8052 | 18.99 | | 4.021 | 3.57 | 300 | 3.8275 | 9.132 | 3.1204 | 8.3109 | 8.5109 | 18.994 | | 4.0126 | 4.76 | 400 | 3.8172 | 9.4354 | 3.1824 | 8.5889 | 8.7927 | 18.996 | ### Framework versions - Transformers 4.31.0 - Pytorch 2.0.1+cu117 - Datasets 2.14.3 - Tokenizers 0.13.3