|
{ |
|
"alpha_pattern": {}, |
|
"auto_mapping": null, |
|
"base_model_name_or_path": "beomi/KoAlpaca-KoRWKV-6B", |
|
"bias": "none", |
|
"fan_in_fan_out": false, |
|
"inference_mode": false, |
|
"init_lora_weights": true, |
|
"layer_replication": null, |
|
"layers_pattern": null, |
|
"layers_to_transform": null, |
|
"loftq_config": {}, |
|
"lora_alpha": 32, |
|
"lora_dropout": 0.05, |
|
"megatron_config": null, |
|
"megatron_core": "megatron.core", |
|
"modules_to_save": null, |
|
"peft_type": "LORA", |
|
"r": 8, |
|
"rank_pattern": {}, |
|
"revision": null, |
|
"target_modules": [ |
|
"rwkv.blocks.3.attention.key", |
|
"rwkv.blocks.25.attention.key", |
|
"rwkv.blocks.16.attention.receptance", |
|
"rwkv.blocks.17.attention.receptance", |
|
"rwkv.blocks.25.attention.value", |
|
"rwkv.blocks.24.attention.receptance", |
|
"rwkv.blocks.23.attention.value", |
|
"rwkv.blocks.26.attention.key", |
|
"rwkv.blocks.18.attention.value", |
|
"rwkv.blocks.16.attention.key", |
|
"rwkv.blocks.5.attention.value", |
|
"rwkv.blocks.14.attention.key", |
|
"rwkv.blocks.0.attention.key", |
|
"rwkv.blocks.22.attention.key", |
|
"rwkv.blocks.8.attention.key", |
|
"rwkv.blocks.6.attention.receptance", |
|
"rwkv.blocks.9.attention.receptance", |
|
"rwkv.blocks.19.attention.receptance", |
|
"rwkv.blocks.21.attention.key", |
|
"rwkv.blocks.4.attention.receptance", |
|
"rwkv.blocks.9.attention.value", |
|
"rwkv.blocks.17.attention.key", |
|
"rwkv.blocks.18.attention.key", |
|
"rwkv.blocks.19.attention.key", |
|
"rwkv.blocks.23.attention.receptance", |
|
"rwkv.blocks.27.attention.receptance", |
|
"rwkv.blocks.23.attention.key", |
|
"rwkv.blocks.5.attention.key", |
|
"rwkv.blocks.16.attention.value", |
|
"rwkv.blocks.4.attention.value", |
|
"rwkv.blocks.26.attention.receptance", |
|
"rwkv.blocks.7.attention.key", |
|
"rwkv.blocks.11.attention.key", |
|
"rwkv.blocks.15.attention.key", |
|
"rwkv.blocks.0.attention.receptance", |
|
"rwkv.blocks.24.attention.key", |
|
"rwkv.blocks.13.attention.receptance", |
|
"rwkv.blocks.8.attention.receptance", |
|
"rwkv.blocks.10.attention.key", |
|
"rwkv.blocks.27.attention.key", |
|
"rwkv.blocks.5.attention.receptance", |
|
"rwkv.blocks.1.attention.receptance", |
|
"rwkv.blocks.17.attention.value", |
|
"rwkv.blocks.2.attention.receptance", |
|
"rwkv.blocks.3.attention.value", |
|
"rwkv.blocks.13.attention.value", |
|
"rwkv.blocks.2.attention.value", |
|
"rwkv.blocks.6.attention.value", |
|
"rwkv.blocks.10.attention.receptance", |
|
"rwkv.blocks.2.attention.key", |
|
"rwkv.blocks.12.attention.receptance", |
|
"rwkv.blocks.3.attention.receptance", |
|
"rwkv.blocks.11.attention.receptance", |
|
"rwkv.blocks.11.attention.value", |
|
"rwkv.blocks.26.attention.value", |
|
"rwkv.blocks.4.attention.key", |
|
"rwkv.blocks.0.attention.value", |
|
"rwkv.blocks.13.attention.key", |
|
"rwkv.blocks.24.attention.value", |
|
"rwkv.blocks.15.attention.receptance", |
|
"rwkv.blocks.10.attention.value", |
|
"rwkv.blocks.14.attention.value", |
|
"rwkv.blocks.21.attention.receptance", |
|
"rwkv.blocks.22.attention.value", |
|
"rwkv.blocks.20.attention.receptance", |
|
"rwkv.blocks.9.attention.key", |
|
"rwkv.blocks.20.attention.key", |
|
"rwkv.blocks.12.attention.value", |
|
"rwkv.blocks.7.attention.receptance", |
|
"rwkv.blocks.25.attention.receptance", |
|
"rwkv.blocks.21.attention.value", |
|
"rwkv.blocks.20.attention.value", |
|
"rwkv.blocks.7.attention.value", |
|
"rwkv.blocks.14.attention.receptance", |
|
"rwkv.blocks.15.attention.value", |
|
"rwkv.blocks.1.attention.value", |
|
"rwkv.blocks.1.attention.key", |
|
"rwkv.blocks.8.attention.value", |
|
"rwkv.blocks.18.attention.receptance", |
|
"rwkv.blocks.22.attention.receptance", |
|
"rwkv.blocks.6.attention.key", |
|
"rwkv.blocks.27.attention.value", |
|
"rwkv.blocks.12.attention.key", |
|
"rwkv.blocks.19.attention.value" |
|
], |
|
"task_type": "QUESTION_ANSWERING", |
|
"use_dora": false, |
|
"use_rslora": false |
|
} |