ayjays132 commited on
Commit
0f2dbf8
ยท
1 Parent(s): cee3688

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -13
README.md CHANGED
@@ -1,6 +1,7 @@
1
  ---
2
  model_type: gpt2
3
  model_name_or_path: gpt2-medium
 
4
  hidden_size: 2048
5
  num_hidden_layers: 12
6
  num_attention_heads: 12
@@ -27,19 +28,31 @@ pipeline_tag: conversational
27
  ---
28
  ## Hyperameters used
29
 
30
- model_type: gpt2
31
- model_name_or_path: gpt2-medium # Replace with the actual model architecture name if different
32
- hidden_size: 2048
33
- num_hidden_layers: 12
34
- num_attention_heads: 12
35
- intermediate_size: 3072 # Adjusted for model efficiency
36
- hidden_dropout_prob: 0.1
37
- attention_probs_dropout_prob: 0.1
38
- max_position_embeddings: 1024
39
- type_vocab_size: 1
40
- initializer_range: 0.02
41
- layer_norm_eps: 1e-05
42
- vocab_size: 50257
 
 
 
 
 
 
 
 
 
 
 
 
43
  ---
44
 
45
  # QNetworkGPT2: Reinventing Text Generation with AI ๐Ÿ“๐Ÿค–
 
1
  ---
2
  model_type: gpt2
3
  model_name_or_path: gpt2-medium
4
+ model_filename: pytorch_model.bin
5
  hidden_size: 2048
6
  num_hidden_layers: 12
7
  num_attention_heads: 12
 
28
  ---
29
  ## Hyperameters used
30
 
31
+ Certainly! Here's a consolidated list of hyperparameters for your QNetworkGPT2 RL model:
32
+
33
+ - `input_dim`: Input dimension for the RL agent.
34
+ - `output_dim`: Output dimension for the RL agent.
35
+ - `hidden_dim`: Hidden dimension for the RL agent.
36
+ - `num_episodes`: Number of training episodes.
37
+ - `generate_interval`: Interval for text generation during training.
38
+ - `load_path`: Path to load a pre-trained model.
39
+ - `model_name`: GPT-2 model architecture name.
40
+ - `max_new_tokens`: Maximum new tokens allowed during text generation.
41
+ - `max_length`: Maximum sequence length for input data.
42
+ - `sequence_length`: Length of sequences in the dataset.
43
+ - `batch_size`: Batch size for training.
44
+ - `learning_rate`: Learning rate for optimization.
45
+ - `gamma`: Discount factor for rewards.
46
+ - `clip_epsilon`: Epsilon value for policy loss clipping.
47
+ - `entropy_beta`: Beta value for entropy regularization.
48
+ - `epsilon_start`: Initial epsilon for epsilon-greedy exploration.
49
+ - `epsilon_end`: Minimum epsilon value.
50
+ - `epsilon_decay`: Epsilon decay rate.
51
+ - `heuristic_fn`: Heuristic function for action selection.
52
+ - `max_new_tokens`: Maximum new tokens allowed during text generation.
53
+ - `save_path`: Path to save the trained model.
54
+
55
+ Researchers can use these hyperparameters to configure and train their QNetworkGPT2 RL models effectively for text generation tasks.
56
  ---
57
 
58
  # QNetworkGPT2: Reinventing Text Generation with AI ๐Ÿ“๐Ÿค–