gpt2-large-imdb3 This model is a fine-tuned version of output_models/gpt2-large-news2 on the customized dataset.
Model description More information needed
Intended uses & limitations More information needed
Training procedure Training hyperparameters The following hyperparameters were used during training:
learning_rate: 0.0003 train_batch_size: 1 eval_batch_size: 8 seed: 42 distributed_type: multi-GPU num_devices: 6 total_train_batch_size: 6 total_eval_batch_size: 48 optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 lr_scheduler_type: linear num_epochs: 3.0 Training results Framework versions Transformers 4.32.1 Pytorch 2.0.1+cu117 Datasets 2.10.1 Tokenizers 0.13.3
- Downloads last month
- 1,137
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.