Avokado777 commited on
Commit
fd83268
·
verified ·
1 Parent(s): a297e51

rl-swarm: round 4055, agent fast_small_gibbon

Browse files
Files changed (4) hide show
  1. README.md +2 -0
  2. config.json +2 -3
  3. generation_config.json +1 -2
  4. model.safetensors +1 -1
README.md CHANGED
@@ -9,6 +9,8 @@ tags:
9
  - gensyn
10
  - I am fast small gibbon
11
  - trl
 
 
12
  licence: license
13
  ---
14
 
 
9
  - gensyn
10
  - I am fast small gibbon
11
  - trl
12
+ - genrl-swarm
13
+ - I am fast_small_gibbon
14
  licence: license
15
  ---
16
 
config.json CHANGED
@@ -3,6 +3,7 @@
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
 
6
  "eos_token_id": 151645,
7
  "hidden_act": "silu",
8
  "hidden_size": 896,
@@ -14,15 +15,13 @@
14
  "num_attention_heads": 14,
15
  "num_hidden_layers": 24,
16
  "num_key_value_heads": 2,
17
- "pad_token_id": 151654,
18
  "rms_norm_eps": 1e-06,
19
  "rope_scaling": null,
20
  "rope_theta": 1000000.0,
21
- "sliding_window": null,
22
  "tie_word_embeddings": true,
23
  "torch_dtype": "float32",
24
  "transformers_version": "4.51.3",
25
- "unsloth_fixed": true,
26
  "use_cache": true,
27
  "use_sliding_window": false,
28
  "vocab_size": 151936
 
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
  "eos_token_id": 151645,
8
  "hidden_act": "silu",
9
  "hidden_size": 896,
 
15
  "num_attention_heads": 14,
16
  "num_hidden_layers": 24,
17
  "num_key_value_heads": 2,
 
18
  "rms_norm_eps": 1e-06,
19
  "rope_scaling": null,
20
  "rope_theta": 1000000.0,
21
+ "sliding_window": 32768,
22
  "tie_word_embeddings": true,
23
  "torch_dtype": "float32",
24
  "transformers_version": "4.51.3",
 
25
  "use_cache": true,
26
  "use_sliding_window": false,
27
  "vocab_size": 151936
generation_config.json CHANGED
@@ -5,8 +5,7 @@
5
  151645,
6
  151643
7
  ],
8
- "max_length": 32768,
9
- "pad_token_id": 151654,
10
  "repetition_penalty": 1.1,
11
  "temperature": 0.7,
12
  "top_k": 20,
 
5
  151645,
6
  151643
7
  ],
8
+ "pad_token_id": 151643,
 
9
  "repetition_penalty": 1.1,
10
  "temperature": 0.7,
11
  "top_k": 20,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cd74aabc43044b399b3fdbb155c26bd99656d58a8083aa93eee67aa286664299
3
  size 1976163472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd4cc7517ff00d28dc4abd4b72ba6b51f3786b2bc9dab4a1ba0feeb870d696ee
3
  size 1976163472