## MechDistilGPT2 | |
This model is fine-tuned on 200k text scraped from Mechanical/Automotive pdf books. | |
Base model is DistilGPT2(https://huggingface.co/gpt2) (the smallest version of GPT2) | |
## Fine-Tuning | |
Default Training Args | |
Epochs = 3 | |
Perplexity = 48 | |
Training set = 200k sentences | |
Validation set = 40k sentences | |
## Framework versions | |
Transformers 4.7.0.dev0 | |
Pytorch 1.8.1+cu111 | |
Datasets 1.6.2 | |
Tokenizers 0.10.2 |