datasets: | |
- tatsu-lab/alpaca | |
language: | |
- en | |
metrics: | |
- accuracy | |
base_model: openai-community/gpt2 | |
pipeline_tag: text-generation | |
library_name: transformers | |
This model is a fine-tuned version of gpt2 on an [tatsu-lab/alpaca](https://huggingface.co/datasets/tatsu-lab/alpaca) dataset. It achieves the following results on the evaluation set: | |
Loss: 1.826895 | |