## MechDistilGPT2 This model is fine-tuned on 200k text scraped from Mechanical/Automotive pdf books. Base model is DistilGPT2(https://huggingface.co/gpt2) (the smallest version of GPT2) ## Fine-Tuning Default Training Args Epochs = 3 Perplexity = 48 Training set = 200k sentences Validation set = 40k sentences ## Framework versions Transformers 4.7.0.dev0 Pytorch 1.8.1+cu111 Datasets 1.6.2 Tokenizers 0.10.2