Setpember commited on
Commit
a65e636
·
verified ·
1 Parent(s): 6a02bc2

Push model using huggingface_hub.

Browse files
Files changed (3) hide show
  1. README.md +3 -3
  2. config.json +3 -3
  3. pytorch_model.bin +1 -1
README.md CHANGED
@@ -25,7 +25,7 @@ You can then generate text as follows:
25
  ```python
26
  from transformers import pipeline
27
 
28
- generator = pipeline("text-generation", model="Setpember//tmp/tmp9xts5xnq/Setpember/Jon_GPT2L_PPO_epi_2")
29
  outputs = generator("Hello, my llama is cute")
30
  ```
31
 
@@ -35,8 +35,8 @@ If you want to use the model for training or to obtain the outputs from the valu
35
  from transformers import AutoTokenizer
36
  from trl import AutoModelForCausalLMWithValueHead
37
 
38
- tokenizer = AutoTokenizer.from_pretrained("Setpember//tmp/tmp9xts5xnq/Setpember/Jon_GPT2L_PPO_epi_2")
39
- model = AutoModelForCausalLMWithValueHead.from_pretrained("Setpember//tmp/tmp9xts5xnq/Setpember/Jon_GPT2L_PPO_epi_2")
40
 
41
  inputs = tokenizer("Hello, my llama is cute", return_tensors="pt")
42
  outputs = model(**inputs, labels=inputs["input_ids"])
 
25
  ```python
26
  from transformers import pipeline
27
 
28
+ generator = pipeline("text-generation", model="Setpember//tmp/tmp49gn1ihq/Setpember/Jon_GPT2L_PPO_epi_2")
29
  outputs = generator("Hello, my llama is cute")
30
  ```
31
 
 
35
  from transformers import AutoTokenizer
36
  from trl import AutoModelForCausalLMWithValueHead
37
 
38
+ tokenizer = AutoTokenizer.from_pretrained("Setpember//tmp/tmp49gn1ihq/Setpember/Jon_GPT2L_PPO_epi_2")
39
+ model = AutoModelForCausalLMWithValueHead.from_pretrained("Setpember//tmp/tmp49gn1ihq/Setpember/Jon_GPT2L_PPO_epi_2")
40
 
41
  inputs = tokenizer("Hello, my llama is cute", return_tensors="pt")
42
  outputs = model(**inputs, labels=inputs["input_ids"])
config.json CHANGED
@@ -2,7 +2,7 @@
2
  "accelerator_kwargs": {},
3
  "adap_kl_ctrl": true,
4
  "backward_batch_size": 1,
5
- "batch_size": 256,
6
  "cliprange": 0.2,
7
  "cliprange_value": 0.2,
8
  "compare_steps": 1,
@@ -11,7 +11,7 @@
11
  "forward_batch_size": null,
12
  "gamma": 1,
13
  "global_backward_batch_size": 1,
14
- "global_batch_size": 256,
15
  "gradient_accumulation_steps": 1,
16
  "horizon": 10000,
17
  "init_kl_coef": 0.2,
@@ -24,7 +24,7 @@
24
  "max_grad_norm": null,
25
  "mini_batch_size": 1,
26
  "model_name": "Setpember/sft_gpt2_large",
27
- "optimize_cuda_cache": null,
28
  "optimize_device_cache": false,
29
  "ppo_epochs": 4,
30
  "project_kwargs": {},
 
2
  "accelerator_kwargs": {},
3
  "adap_kl_ctrl": true,
4
  "backward_batch_size": 1,
5
+ "batch_size": 4,
6
  "cliprange": 0.2,
7
  "cliprange_value": 0.2,
8
  "compare_steps": 1,
 
11
  "forward_batch_size": null,
12
  "gamma": 1,
13
  "global_backward_batch_size": 1,
14
+ "global_batch_size": 4,
15
  "gradient_accumulation_steps": 1,
16
  "horizon": 10000,
17
  "init_kl_coef": 0.2,
 
24
  "max_grad_norm": null,
25
  "mini_batch_size": 1,
26
  "model_name": "Setpember/sft_gpt2_large",
27
+ "optimize_cuda_cache": true,
28
  "optimize_device_cache": false,
29
  "ppo_epochs": 4,
30
  "project_kwargs": {},
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1a93c231b14ff36f0991b1b928d2297aa30bfcac8833ab56483807a88309db53
3
  size 6652
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f46068fc07ed19236d6cfe10ec53c7d0803f356914542822a2ddbfabbbbf0f89
3
  size 6652