ruriko commited on
Commit
430b370
·
1 Parent(s): 19e0f32

First model version

Browse files
output-small/bacqua ADDED
@@ -0,0 +1 @@
 
 
1
+ Subproject commit 2f805f4607fa7e13451f52c156beae124c1ad8eb
output-small/config.json DELETED
@@ -1,36 +0,0 @@
1
- {
2
- "_name_or_path": "microsoft/DialoGPT-medium",
3
- "activation_function": "gelu_new",
4
- "architectures": [
5
- "GPT2LMHeadModel"
6
- ],
7
- "attn_pdrop": 0.1,
8
- "bos_token_id": 50256,
9
- "embd_pdrop": 0.1,
10
- "eos_token_id": 50256,
11
- "initializer_range": 0.02,
12
- "layer_norm_epsilon": 1e-05,
13
- "model_type": "gpt2",
14
- "n_ctx": 1024,
15
- "n_embd": 1024,
16
- "n_head": 16,
17
- "n_inner": null,
18
- "n_layer": 24,
19
- "n_positions": 1024,
20
- "resid_pdrop": 0.1,
21
- "scale_attn_weights": true,
22
- "summary_activation": null,
23
- "summary_first_dropout": 0.1,
24
- "summary_proj_to_labels": true,
25
- "summary_type": "cls_index",
26
- "summary_use_proj": true,
27
- "task_specific_params": {
28
- "conversational": {
29
- "max_length": 1000
30
- }
31
- },
32
- "torch_dtype": "float32",
33
- "transformers_version": "4.11.3",
34
- "use_cache": true,
35
- "vocab_size": 50257
36
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
output-small/eval_results.txt DELETED
@@ -1 +0,0 @@
1
- perplexity = tensor(4.3131)
 
 
output-small/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
output-small/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9cb3ab06b053eac1f74ed4dd4912cfb53d94f53772d6ef3aa690f643b01e480
3
- size 1444581337
 
 
 
 
output-small/special_tokens_map.json DELETED
@@ -1 +0,0 @@
1
- {"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
 
 
output-small/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
output-small/tokenizer_config.json DELETED
@@ -1 +0,0 @@
1
- {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-medium", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
 
 
output-small/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1eb9519afff5b9542e77b33502fb978c30d9bd258876abc43af16308daf436ad
3
- size 1327
 
 
 
 
output-small/vocab.json DELETED
The diff for this file is too large to render. See raw diff