phungkhaccuong commited on
Commit
5e1e84f
·
verified ·
1 Parent(s): 317cba6

Training in progress, step 13

Browse files
adapter_config.json CHANGED
@@ -22,11 +22,11 @@
22
  "target_modules": [
23
  "q_proj",
24
  "gate_proj",
25
- "k_proj",
26
- "o_proj",
27
- "down_proj",
28
  "v_proj",
29
- "up_proj"
 
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
22
  "target_modules": [
23
  "q_proj",
24
  "gate_proj",
 
 
 
25
  "v_proj",
26
+ "up_proj",
27
+ "down_proj",
28
+ "k_proj",
29
+ "o_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ea06e402bd867383d2a827e7f82ddee7b66f39575a8061d6ff4c712afee8664
3
  size 41581360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0e2c7903a224327e16927e3f48dbeffd0008dcf4f7e4d90890cb71f06652416
3
  size 41581360
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d2dfa294a25e9f8e0bec8c34d855b268449440b9ff739b8d419198af8d12db3e
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffdc8f3cae7f913a06a47d0411c9fa2323db15439ee34d80ff873c4fd63a9725
3
  size 6776