MechDistilGPT2
This model is fine-tuned on 200k text scraped from Mechanical/Automotive pdf books.
Base model is DistilGPT2(https://huggingface.co/gpt2) (the smallest version of GPT2)
Fine-Tuning
- Default Training Args
- Epochs = 3
- Perplexity = 48
- Training set = 200k sentences
- Validation set = 40k sentences
Framework versions
- Transformers 4.7.0.dev0
- Pytorch 1.8.1+cu111
- Datasets 1.6.2
- Tokenizers 0.10.2