|
--- |
|
title: "Idefics2-Chatty LoRA Adapter" |
|
description: "A pretrained Idefics2-chatty LoRA adapter trained on 'dawoz/frozenlake_prompts_dataset'." |
|
--- |
|
|
|
# Idefics2-Chatty LoRA Adapter |
|
|
|
This is a pretrained Idefics2-chatty LoRA adapter trained on 'dawoz/frozenlake_prompts_dataset'. the checkpoint save here is the one after 1 epoch of training. |
|
|
|
## Trainer Hyperparameters |
|
|
|
```yaml |
|
num_train_epochs: 3 |
|
per_device_train_batch_size: 8 |
|
per_device_eval_batch_size: 4 |
|
gradient_accumulation_steps: 4 |
|
gradient_checkpointing: true |
|
optim: adamw_bnb_8bit |
|
warmup_steps: 50 |
|
learning_rate: 1e-4 |
|
weight_decay: 0.01 |
|
logging_steps: 25 |
|
output_dir: "outputs" |
|
run_name: "idefics2-frozenlake" |
|
save_strategy: "epoch" |
|
save_steps: 250 |
|
save_total_limit: 4 |
|
eval_strategy: "steps" |
|
eval_steps: 250 |
|
bf16: true |
|
push_to_hub: false |
|
hub_model_id: "idefics2-frozenlake" |
|
remove_unused_columns: false |
|
report_to: "wandb" |