gpt2-large-imdb3 This model is a fine-tuned version of output_models/gpt2-large-news2 on the customized dataset.
Model description More information needed
Intended uses & limitations More information needed
Training procedure Training hyperparameters The following hyperparameters were used during training:
learning_rate: 0.0003 train_batch_size: 1 eval_batch_size: 8 seed: 42 distributed_type: multi-GPU num_devices: 6 total_train_batch_size: 6 total_eval_batch_size: 48 optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 lr_scheduler_type: linear num_epochs: 3.0 Training results Framework versions Transformers 4.32.1 Pytorch 2.0.1+cu117 Datasets 2.10.1 Tokenizers 0.13.3
- Downloads last month
- 1,452
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.