Update README.md
Browse files
README.md
CHANGED
|
@@ -32,25 +32,34 @@ This model and its related LoRA was fine-tuned on [https://huggingface.co/failsp
|
|
| 32 |
|
| 33 |
### - PEFT Parameters
|
| 34 |
|
| 35 |
-
- lora_alpha=64
|
| 36 |
-
- lora_dropout=0.05
|
| 37 |
-
- r=128
|
| 38 |
-
- bias="none"
|
| 39 |
|
| 40 |
### - Training Arguments
|
| 41 |
|
| 42 |
-
- num_train_epochs=1
|
| 43 |
-
- per_device_train_batch_size=1
|
| 44 |
-
- gradient_accumulation_steps=4
|
| 45 |
-
- optim="adamw_bnb_8bit"
|
| 46 |
-
- save_steps=25
|
| 47 |
-
- logging_steps=25
|
| 48 |
-
- learning_rate=2e-4
|
| 49 |
-
- weight_decay=0.001
|
| 50 |
-
- fp16=False
|
| 51 |
-
- bf16=False
|
| 52 |
-
- max_grad_norm=0.3
|
| 53 |
-
- max_steps=-1
|
| 54 |
-
- warmup_ratio=0.03
|
| 55 |
-
- group_by_length=True
|
| 56 |
-
- lr_scheduler_type="constant"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 32 |
|
| 33 |
### - PEFT Parameters
|
| 34 |
|
| 35 |
+
- lora_alpha=64
|
| 36 |
+
- lora_dropout=0.05
|
| 37 |
+
- r=128
|
| 38 |
+
- bias="none"
|
| 39 |
|
| 40 |
### - Training Arguments
|
| 41 |
|
| 42 |
+
- num_train_epochs=1
|
| 43 |
+
- per_device_train_batch_size=1
|
| 44 |
+
- gradient_accumulation_steps=4
|
| 45 |
+
- optim="adamw_bnb_8bit"
|
| 46 |
+
- save_steps=25
|
| 47 |
+
- logging_steps=25
|
| 48 |
+
- learning_rate=2e-4
|
| 49 |
+
- weight_decay=0.001
|
| 50 |
+
- fp16=False
|
| 51 |
+
- bf16=False
|
| 52 |
+
- max_grad_norm=0.3
|
| 53 |
+
- max_steps=-1
|
| 54 |
+
- warmup_ratio=0.03
|
| 55 |
+
- group_by_length=True
|
| 56 |
+
- lr_scheduler_type="constant"
|
| 57 |
+
|
| 58 |
+
## Credits
|
| 59 |
+
|
| 60 |
+
- Meta ([https://huggingface.co/meta-llama](https://huggingface.co/meta-llama)): for the original Llama-3;
|
| 61 |
+
- failspy ([https://huggingface.co/failspy](https://huggingface.co/failspy)): for the base model and the orthogonalization implementation;
|
| 62 |
+
- NobodyExistsOnTheInternet ([https://huggingface.co/NobodyExistsOnTheInternet](https://huggingface.co/NobodyExistsOnTheInternet)): for the incredible dataset;
|
| 63 |
+
- Undi95 ([https://huggingface.co/Undi95](https://huggingface.co/Undi95)) and Sao10k ([https://huggingface.co/Sao10K](https://huggingface.co/Sao10K)): my main inspirations for doing these models =]
|
| 64 |
+
|
| 65 |
+
A huge thank you to all of them ☺️
|