|
Trained using https://github.com/tloen/alpaca-lora |
|
with removing the lines |
|
|
|
``` |
|
old_state_dict = model.state_dict |
|
model.state_dict = ( |
|
lambda self, *_, **__: get_peft_model_state_dict( |
|
self, old_state_dict() |
|
) |
|
).__get__(model, type(model)) |
|
``` |
|
|
|
causing problem. |
|
|
|
base_model: yahma/llama-7b-hf |
|
data_path: prognosis/medical_qa_alpaca |
|
output_dir: ./lora-alpaca |
|
batch_size: 128 |
|
micro_batch_size: 8 |
|
num_epochs: 5 |
|
learning_rate: 0.0003 |
|
cutoff_len: 512 |
|
val_set_size: 0.1 |
|
lora_r: 16 |
|
lora_alpha: 16 |
|
lora_dropout: 0.05 |
|
lora_target_modules: ['q_proj', 'k_proj', 'v_proj', 'o_proj'] |
|
train_on_inputs: True |
|
add_eos_token: False |
|
group_by_length: True |
|
wandb_project: medical_alpaca_hf |
|
wandb_run_name: run_3 |
|
wandb_watch: |
|
wandb_log_model: |
|
resume_from_checkpoint: False |
|
prompt template: alpaca |
|
|
|
|
|
### Command used |
|
|
|
|
|
Finetuning |
|
``` |
|
python finetune.py --base_model 'yahma/llama-7b-hf' --data_path 'prognosis/medical_qa_alpaca' --output_dir './lora-alpaca' --wandb_project 'medical_alpaca_hf' --wandb_run_name 'run_3' --lora_target_modules '[q_proj,k_proj,v_proj,o_proj]' --num_epochs 5 --cutoff_len 512 --group_by_length --val_set_size 0.1 --lora_r=16 --micro_batch_size=8 |
|
``` |
|
|
|
Generating |
|
|
|
``` |
|
python generate.py \ |
|
--load_8bit \ |
|
--base_model 'yahma/llama-7b-hf' \ |
|
--lora_weights 'eswardivi/medical_qa_alpaca' \ |
|
--share_gradio |
|
``` |
|
|
|
git lfs |
|
|
|
``` |
|
curl -s https://packagecloud.io/install/repositories/github/git-lfs/script.deb.sh | sudo bash |
|
sudo apt-get install git-lfs |
|
``` |