kakielekea commited on
Commit
cb7cc15
·
verified ·
1 Parent(s): a0c76f9

End of training

Browse files
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: Gensyn/Qwen2.5-0.5B-Instruct
3
  library_name: transformers
4
  model_name: Qwen2.5-0.5B-Instruct-Gensyn-Swarm-gilded_shrewd_gecko
5
  tags:
@@ -14,7 +14,7 @@ licence: license
14
 
15
  # Model Card for Qwen2.5-0.5B-Instruct-Gensyn-Swarm-gilded_shrewd_gecko
16
 
17
- This model is a fine-tuned version of [Gensyn/Qwen2.5-0.5B-Instruct](https://huggingface.co/Gensyn/Qwen2.5-0.5B-Instruct).
18
  It has been trained using [TRL](https://github.com/huggingface/trl).
19
 
20
  ## Quick start
 
1
  ---
2
+ base_model: unsloth/Qwen2.5-0.5B-Instruct
3
  library_name: transformers
4
  model_name: Qwen2.5-0.5B-Instruct-Gensyn-Swarm-gilded_shrewd_gecko
5
  tags:
 
14
 
15
  # Model Card for Qwen2.5-0.5B-Instruct-Gensyn-Swarm-gilded_shrewd_gecko
16
 
17
+ This model is a fine-tuned version of [unsloth/Qwen2.5-0.5B-Instruct](https://huggingface.co/unsloth/Qwen2.5-0.5B-Instruct).
18
  It has been trained using [TRL](https://github.com/huggingface/trl).
19
 
20
  ## Quick start
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 2.334083110326901e-05,
4
- "train_runtime": 17913.3066,
5
- "train_samples": 17,
6
- "train_samples_per_second": 0.009,
7
- "train_steps_per_second": 0.001
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 5.349516868591308e-07,
4
+ "train_runtime": 33763.7805,
5
+ "train_samples": 5,
6
+ "train_samples_per_second": 0.001,
7
+ "train_steps_per_second": 0.0
8
  }
config.json CHANGED
@@ -3,7 +3,6 @@
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
- "bos_token_id": 151643,
7
  "eos_token_id": 151645,
8
  "hidden_act": "silu",
9
  "hidden_size": 896,
@@ -15,13 +14,15 @@
15
  "num_attention_heads": 14,
16
  "num_hidden_layers": 24,
17
  "num_key_value_heads": 2,
 
18
  "rms_norm_eps": 1e-06,
19
  "rope_scaling": null,
20
  "rope_theta": 1000000.0,
21
- "sliding_window": 32768,
22
  "tie_word_embeddings": true,
23
  "torch_dtype": "float32",
24
  "transformers_version": "4.51.2",
 
25
  "use_cache": true,
26
  "use_sliding_window": false,
27
  "vocab_size": 151936
 
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
 
6
  "eos_token_id": 151645,
7
  "hidden_act": "silu",
8
  "hidden_size": 896,
 
14
  "num_attention_heads": 14,
15
  "num_hidden_layers": 24,
16
  "num_key_value_heads": 2,
17
+ "pad_token_id": 151654,
18
  "rms_norm_eps": 1e-06,
19
  "rope_scaling": null,
20
  "rope_theta": 1000000.0,
21
+ "sliding_window": null,
22
  "tie_word_embeddings": true,
23
  "torch_dtype": "float32",
24
  "transformers_version": "4.51.2",
25
+ "unsloth_fixed": true,
26
  "use_cache": true,
27
  "use_sliding_window": false,
28
  "vocab_size": 151936
generation_config.json CHANGED
@@ -5,7 +5,8 @@
5
  151645,
6
  151643
7
  ],
8
- "pad_token_id": 151643,
 
9
  "repetition_penalty": 1.1,
10
  "temperature": 0.7,
11
  "top_k": 20,
 
5
  151645,
6
  151643
7
  ],
8
+ "max_length": 32768,
9
+ "pad_token_id": 151654,
10
  "repetition_penalty": 1.1,
11
  "temperature": 0.7,
12
  "top_k": 20,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:15d8162ff2f5da1973710d93f2b47430e7ce8999fe19ac9f1612de6271afd4b5
3
  size 1976163472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d994cea4e26e67b289f513b952d75f515fa334b90e8ab27ed9834f0453de6a69
3
  size 1976163472
special_tokens_map.json CHANGED
@@ -22,7 +22,7 @@
22
  "single_word": false
23
  },
24
  "pad_token": {
25
- "content": "<|endoftext|>",
26
  "lstrip": false,
27
  "normalized": false,
28
  "rstrip": false,
 
22
  "single_word": false
23
  },
24
  "pad_token": {
25
+ "content": "<|vision_pad|>",
26
  "lstrip": false,
27
  "normalized": false,
28
  "rstrip": false,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5eee858c5123a4279c3e1f7b81247343f356ac767940b2692a928ad929543214
3
- size 11422063
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64e71213db910f5cafa86d35091f37393dcc344b1bbc34091d1b3eed4cca01d5
3
+ size 11422064
tokenizer_config.json CHANGED
@@ -200,8 +200,9 @@
200
  "eos_token": "<|im_end|>",
201
  "errors": "replace",
202
  "extra_special_tokens": {},
203
- "model_max_length": 131072,
204
- "pad_token": "<|endoftext|>",
 
205
  "split_special_tokens": false,
206
  "tokenizer_class": "Qwen2Tokenizer",
207
  "unk_token": null
 
200
  "eos_token": "<|im_end|>",
201
  "errors": "replace",
202
  "extra_special_tokens": {},
203
+ "model_max_length": 32768,
204
+ "pad_token": "<|vision_pad|>",
205
+ "padding_side": "left",
206
  "split_special_tokens": false,
207
  "tokenizer_class": "Qwen2Tokenizer",
208
  "unk_token": null
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 2.334083110326901e-05,
4
- "train_runtime": 17913.3066,
5
- "train_samples": 17,
6
- "train_samples_per_second": 0.009,
7
- "train_steps_per_second": 0.001
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 5.349516868591308e-07,
4
+ "train_runtime": 33763.7805,
5
+ "train_samples": 5,
6
+ "train_samples_per_second": 0.001,
7
+ "train_steps_per_second": 0.0
8
  }
trainer_state.json CHANGED
@@ -2,120 +2,75 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 3.4705882352941178,
6
  "eval_steps": 500,
7
- "global_step": 10,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
- "completion_length": 131.78125,
14
- "epoch": 0.9411764705882353,
15
- "grad_norm": 10.730029106140137,
16
  "kl": 0.0,
17
  "learning_rate": 5e-07,
18
- "loss": -0.0,
19
- "reward": 6.342288956046104,
20
- "reward_std": 0.6563283740542829,
21
- "rewards/concensus_correctness_reward_func": 2.102062474936247,
22
- "rewards/consensus_reward_func": 1.9375,
23
- "rewards/cumulative_reward_2": 0.0,
24
- "rewards/final_correctness_reward_func": 0.25,
25
- "rewards/question_recreation_reward_func": 0.9382578537333757,
26
- "rewards/soft_format_reward_func": 0.0,
27
- "rewards/strict_format_reward_func": 0.15625,
28
- "rewards/xmlcount_reward_func": 0.9582187570631504,
29
- "step": 2
30
- },
31
- {
32
- "completion_length": 93.83333333333333,
33
- "epoch": 1.4705882352941178,
34
- "grad_norm": 7.171228408813477,
35
- "kl": 0.017023057715656858,
36
- "learning_rate": 4.415111107797445e-07,
37
  "loss": 0.0,
38
- "reward": 6.55574173397488,
39
- "reward_std": 0.43832636404679054,
40
- "rewards/concensus_correctness_reward_func": 2.472111072805193,
41
- "rewards/consensus_reward_func": 1.8888888888888888,
42
  "rewards/cumulative_reward_2": 0.0,
43
- "rewards/final_correctness_reward_func": 0.4444444444444444,
44
- "rewards/question_recreation_reward_func": 0.7796862405310903,
45
  "rewards/soft_format_reward_func": 0.0,
46
- "rewards/strict_format_reward_func": 0.1111111111111111,
47
- "rewards/xmlcount_reward_func": 0.8594999843173556,
48
- "step": 4
49
  },
50
  {
51
- "completion_length": 122.61111111111111,
52
  "epoch": 2.0,
53
- "grad_norm": 3.105441237494233e-08,
54
- "kl": 0.02888714972262581,
55
- "learning_rate": 2.934120444167326e-07,
56
- "loss": 0.0,
57
- "reward": 6.286137633853489,
58
- "reward_std": 0.16779031413089898,
59
- "rewards/concensus_correctness_reward_func": 2.134666668044196,
60
- "rewards/consensus_reward_func": 2.0,
61
- "rewards/cumulative_reward_2": 0.0,
62
- "rewards/final_correctness_reward_func": 0.1111111111111111,
63
- "rewards/question_recreation_reward_func": 0.8890821365122166,
64
- "rewards/soft_format_reward_func": 0.0,
65
- "rewards/strict_format_reward_func": 0.1111111111111111,
66
- "rewards/xmlcount_reward_func": 1.0401666694217258,
67
- "step": 6
68
- },
69
- {
70
- "completion_length": 121.0625,
71
- "epoch": 2.9411764705882355,
72
- "grad_norm": 5.204808235168457,
73
- "kl": 0.036964431335945847,
74
- "learning_rate": 1.2500000000000005e-07,
75
  "loss": 0.0,
76
- "reward": 6.445510894060135,
77
- "reward_std": 0.12171002507966477,
78
- "rewards/concensus_correctness_reward_func": 2.28712497651577,
79
- "rewards/consensus_reward_func": 2.0,
80
  "rewards/cumulative_reward_2": 0.0,
81
- "rewards/final_correctness_reward_func": 0.25,
82
- "rewards/question_recreation_reward_func": 0.8760735102987383,
83
  "rewards/soft_format_reward_func": 0.0,
84
- "rewards/strict_format_reward_func": 0.109375,
85
- "rewards/xmlcount_reward_func": 0.9229374881833792,
86
- "step": 8
87
  },
88
  {
89
- "completion_length": 135.33333333333334,
90
- "epoch": 3.4705882352941178,
91
- "grad_norm": 13.752861976623535,
92
- "kl": 0.08299477315611309,
93
- "learning_rate": 1.507684480352292e-08,
94
- "loss": 0.0,
95
- "reward": 6.4182407591078015,
96
- "reward_std": 0.18086219827334085,
97
- "rewards/concensus_correctness_reward_func": 2.3553333150015936,
98
- "rewards/consensus_reward_func": 2.0,
99
  "rewards/cumulative_reward_2": 0.0,
100
- "rewards/final_correctness_reward_func": 0.2222222222222222,
101
- "rewards/question_recreation_reward_func": 0.7799075724453561,
102
  "rewards/soft_format_reward_func": 0.0,
103
- "rewards/strict_format_reward_func": 0.1111111111111111,
104
- "rewards/xmlcount_reward_func": 0.9496666457917955,
105
- "step": 10
106
- },
107
- {
108
- "epoch": 3.4705882352941178,
109
- "step": 10,
110
  "total_flos": 0.0,
111
- "train_loss": 2.334083110326901e-05,
112
- "train_runtime": 17913.3066,
113
- "train_samples_per_second": 0.009,
114
- "train_steps_per_second": 0.001
115
  }
116
  ],
117
  "logging_steps": 2,
118
- "max_steps": 10,
119
  "num_input_tokens_seen": 0,
120
  "num_train_epochs": 5,
121
  "save_steps": 25,
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 2.8,
6
  "eval_steps": 500,
7
+ "global_step": 5,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
+ "completion_length": 191.7,
14
+ "epoch": 1.0,
15
+ "grad_norm": 3.4990100860595703,
16
  "kl": 0.0,
17
  "learning_rate": 5e-07,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  "loss": 0.0,
19
+ "reward": 0.13842474818229675,
20
+ "reward_std": 0.09636008515954017,
21
+ "rewards/concensus_correctness_reward_func": 0.0,
22
+ "rewards/consensus_reward_func": 0.0,
23
  "rewards/cumulative_reward_2": 0.0,
24
+ "rewards/final_correctness_reward_func": 0.0,
25
+ "rewards/question_recreation_reward_func": 0.11342475041747094,
26
  "rewards/soft_format_reward_func": 0.0,
27
+ "rewards/strict_format_reward_func": 0.0,
28
+ "rewards/xmlcount_reward_func": 0.025,
29
+ "step": 2
30
  },
31
  {
32
+ "completion_length": 173.0,
33
  "epoch": 2.0,
34
+ "grad_norm": 2.3786094188690186,
35
+ "kl": 0.0006931740383151919,
36
+ "learning_rate": 2.5e-07,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
37
  "loss": 0.0,
38
+ "reward": 0.17476099207997323,
39
+ "reward_std": 0.059773775190114974,
40
+ "rewards/concensus_correctness_reward_func": 0.0,
41
+ "rewards/consensus_reward_func": 0.0,
42
  "rewards/cumulative_reward_2": 0.0,
43
+ "rewards/final_correctness_reward_func": 0.0,
44
+ "rewards/question_recreation_reward_func": 0.1956609971821308,
45
  "rewards/soft_format_reward_func": 0.0,
46
+ "rewards/strict_format_reward_func": 0.0,
47
+ "rewards/xmlcount_reward_func": -0.020899999886751175,
48
+ "step": 4
49
  },
50
  {
51
+ "completion_length": 135.875,
52
+ "epoch": 2.8,
53
+ "kl": 0.0009647479673731141,
54
+ "reward": 0.09912843629717827,
55
+ "reward_std": 0.05026640766300261,
56
+ "rewards/concensus_correctness_reward_func": 0.0,
57
+ "rewards/consensus_reward_func": 0.0,
 
 
 
58
  "rewards/cumulative_reward_2": 0.0,
59
+ "rewards/final_correctness_reward_func": 0.0,
60
+ "rewards/question_recreation_reward_func": 0.0678784353658557,
61
  "rewards/soft_format_reward_func": 0.0,
62
+ "rewards/strict_format_reward_func": 0.0,
63
+ "rewards/xmlcount_reward_func": 0.03125,
64
+ "step": 5,
 
 
 
 
65
  "total_flos": 0.0,
66
+ "train_loss": 5.349516868591308e-07,
67
+ "train_runtime": 33763.7805,
68
+ "train_samples_per_second": 0.001,
69
+ "train_steps_per_second": 0.0
70
  }
71
  ],
72
  "logging_steps": 2,
73
+ "max_steps": 5,
74
  "num_input_tokens_seen": 0,
75
  "num_train_epochs": 5,
76
  "save_steps": 25,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:852aacd4d84d703d5046f7d86492b82ca4f7b4f2e2b19b3deda376ebc79b9afb
3
  size 5944
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1abc1be2cf80e7a84ea2ff75490a1d1c8184d0360e57b98a7d0375d568a07be
3
  size 5944