beamaia commited on
Commit
72f6d70
·
verified ·
1 Parent(s): dbaf33a

Training in progress, epoch 0, checkpoint

Browse files
checkpoint-10/adapter_config.json CHANGED
@@ -9,13 +9,13 @@
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
12
- "lora_alpha": 8,
13
  "lora_dropout": 0.05,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
16
  "modules_to_save": null,
17
  "peft_type": "LORA",
18
- "r": 8,
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
12
+ "lora_alpha": 16,
13
  "lora_dropout": 0.05,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
16
  "modules_to_save": null,
17
  "peft_type": "LORA",
18
+ "r": 16,
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
checkpoint-10/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b9b2d576eceb9eb6d2d084808671f66357fe38db8508ae6def2045d2e86c4681
3
- size 13648432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41b0915dbabe732a22a164ddd732f0bc39f270aabadd91d0af928396b0fd8cda
3
+ size 27280152
checkpoint-10/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4683358d3ecdb4f1484131e255d12b2c508a88328d6b29e71db7aeee7c48619f
3
- size 27338682
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fef419d6c9ebc6eff293f10d20b8d172f6c661b12092e1e0f2c11c32a2cad3be
3
+ size 54602042
checkpoint-10/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dd0704cf53e73fdd3790d041fcd30775fdd963463e54c18c6c10f750649e8839
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e119516d1e041105639f70ebe026fb0f629d45578e5ab3a60aa8f90c988f825
3
  size 14244
checkpoint-10/trainer_state.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "best_metric": 0.5767254829406738,
3
  "best_model_checkpoint": "./Zephyr/12-01-24-HuggingFaceH4-zephyr-7b-beta_Treino-GPTQ-2_epochs-5_batch_1/checkpoints/checkpoint-10",
4
- "epoch": 0.06938421509106678,
5
  "eval_steps": 500,
6
  "global_step": 10,
7
  "is_hyper_param_search": false,
@@ -9,11 +9,11 @@
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.07,
13
- "eval_loss": 0.5767254829406738,
14
- "eval_runtime": 374.7385,
15
- "eval_samples_per_second": 0.342,
16
- "eval_steps_per_second": 0.171,
17
  "step": 10
18
  }
19
  ],
@@ -21,7 +21,7 @@
21
  "max_steps": 10,
22
  "num_train_epochs": 1,
23
  "save_steps": 500,
24
- "total_flos": 2.797365350105088e+16,
25
  "trial_name": null,
26
  "trial_params": null
27
  }
 
1
  {
2
+ "best_metric": 0.5605378746986389,
3
  "best_model_checkpoint": "./Zephyr/12-01-24-HuggingFaceH4-zephyr-7b-beta_Treino-GPTQ-2_epochs-5_batch_1/checkpoints/checkpoint-10",
4
+ "epoch": 0.2775368603642671,
5
  "eval_steps": 500,
6
  "global_step": 10,
7
  "is_hyper_param_search": false,
 
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.28,
13
+ "eval_loss": 0.5605378746986389,
14
+ "eval_runtime": 375.7067,
15
+ "eval_samples_per_second": 0.341,
16
+ "eval_steps_per_second": 0.17,
17
  "step": 10
18
  }
19
  ],
 
21
  "max_steps": 10,
22
  "num_train_epochs": 1,
23
  "save_steps": 500,
24
+ "total_flos": 1.119482151960576e+17,
25
  "trial_name": null,
26
  "trial_params": null
27
  }
checkpoint-10/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf8ad3c191012642f1f5613c45fec6745c7e550e2a431878ff850248c0965e9f
3
  size 4856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf37fa4263bd42407013461e43287124efa1cc33f6427e440772692788437888
3
  size 4856