| model_family: llama2-7b | |
| LoRA: | |
| r: 0 | |
| alpha: 32 | |
| dropout: 0.05 | |
| data_path: locuslab/TOFU | |
| split: full | |
| batch_size: 4 | |
| gradient_accumulation_steps: 4 | |
| num_epochs: 5 | |
| lr: 1.0e-05 | |
| save_dir: paper_models/final_ft_noLORA_5_epochs_inst_lr${lr}_${model_family}_${split} | |
| weight_decay: 0.01 | |