doyoungkim commited on
Commit
db10502
·
verified ·
1 Parent(s): 1b7ecb5

Add action tokens and update config

Browse files
Files changed (3) hide show
  1. config.json +18 -3
  2. generation_config.json +1 -1
  3. model.safetensors +2 -2
config.json CHANGED
@@ -12,11 +12,16 @@
12
  "architectures": [
13
  "VLlama3ForCausalLM"
14
  ],
 
 
15
  "head_dim": 64,
 
16
  "hidden_size": 960,
 
17
  "intermediate_size": 2560,
18
  "is_llama_config": true,
19
  "max_position_embeddings": 8192,
 
20
  "model_type": "llama",
21
  "neftune_noise_alpha": 0.0,
22
  "num_attention_heads": 15,
@@ -24,7 +29,6 @@
24
  "num_key_value_heads": 5,
25
  "pad_token_id": 2,
26
  "perceiver_config": {
27
- "_attn_implementation_autoset": false,
28
  "_name_or_path": "",
29
  "add_cross_attention": false,
30
  "architectures": null,
@@ -97,17 +101,20 @@
97
  "use_bfloat16": false
98
  },
99
  "pixel_shuffle_factor": 4,
 
100
  "qk_layer_norms": false,
101
  "rms_norm_eps": 1e-05,
102
  "rope_interleaved": false,
 
103
  "rope_theta": 100000,
104
- "torch_dtype": "bfloat16",
105
  "transformers.js_config": {
106
  "kv_cache_dtype": {
107
  "fp16": "float16",
108
  "q4f16": "float16"
109
  }
110
  },
 
111
  "use_resampler": false,
112
  "vocab_size": 49536
113
  },
@@ -119,22 +126,30 @@
119
  "q4f16": "float16"
120
  }
121
  },
122
- "transformers_version": "4.47.1",
123
  "use_cache": false,
124
  "use_reentrant_checkpointing": false,
125
  "vision_config": {
 
 
126
  "hidden_size": 768,
127
  "image_size": 512,
 
 
 
128
  "max_image_size": {
129
  "longest_edge": 512
130
  },
131
  "model_type": "smolvlm_vision",
132
  "num_attention_heads": 12,
 
 
133
  "patch_size": 16,
134
  "size": {
135
  "longest_edge": 512
136
  },
137
  "tie_word_embeddings": false,
 
138
  "use_base_siglip": false
139
  },
140
  "vocab_size": 49536
 
12
  "architectures": [
13
  "VLlama3ForCausalLM"
14
  ],
15
+ "attention_bias": false,
16
+ "attention_dropout": 0.0,
17
  "head_dim": 64,
18
+ "hidden_act": "silu",
19
  "hidden_size": 960,
20
+ "initializer_range": 0.02,
21
  "intermediate_size": 2560,
22
  "is_llama_config": true,
23
  "max_position_embeddings": 8192,
24
+ "mlp_bias": false,
25
  "model_type": "llama",
26
  "neftune_noise_alpha": 0.0,
27
  "num_attention_heads": 15,
 
29
  "num_key_value_heads": 5,
30
  "pad_token_id": 2,
31
  "perceiver_config": {
 
32
  "_name_or_path": "",
33
  "add_cross_attention": false,
34
  "architectures": null,
 
101
  "use_bfloat16": false
102
  },
103
  "pixel_shuffle_factor": 4,
104
+ "pretraining_tp": 1,
105
  "qk_layer_norms": false,
106
  "rms_norm_eps": 1e-05,
107
  "rope_interleaved": false,
108
+ "rope_scaling": null,
109
  "rope_theta": 100000,
110
+ "torch_dtype": "float32",
111
  "transformers.js_config": {
112
  "kv_cache_dtype": {
113
  "fp16": "float16",
114
  "q4f16": "float16"
115
  }
116
  },
117
+ "use_cache": true,
118
  "use_resampler": false,
119
  "vocab_size": 49536
120
  },
 
126
  "q4f16": "float16"
127
  }
128
  },
129
+ "transformers_version": "4.52.4",
130
  "use_cache": false,
131
  "use_reentrant_checkpointing": false,
132
  "vision_config": {
133
+ "attention_dropout": 0.0,
134
+ "hidden_act": "gelu_pytorch_tanh",
135
  "hidden_size": 768,
136
  "image_size": 512,
137
+ "initializer_range": 0.02,
138
+ "intermediate_size": 3072,
139
+ "layer_norm_eps": 1e-06,
140
  "max_image_size": {
141
  "longest_edge": 512
142
  },
143
  "model_type": "smolvlm_vision",
144
  "num_attention_heads": 12,
145
+ "num_channels": 3,
146
+ "num_hidden_layers": 12,
147
  "patch_size": 16,
148
  "size": {
149
  "longest_edge": 512
150
  },
151
  "tie_word_embeddings": false,
152
+ "torch_dtype": "float32",
153
  "use_base_siglip": false
154
  },
155
  "vocab_size": 49536
generation_config.json CHANGED
@@ -3,5 +3,5 @@
3
  "bos_token_id": 0,
4
  "eos_token_id": 49279,
5
  "pad_token_id": 2,
6
- "transformers_version": "4.47.1"
7
  }
 
3
  "bos_token_id": 0,
4
  "eos_token_id": 49279,
5
  "pad_token_id": 2,
6
+ "transformers_version": "4.52.4"
7
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b9bfd456c9472c0acd5719d6e514c4b859891af205ee1a736552fd3497b8b0c3
3
- size 2029990624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7efc851e741bca179a92bb6492b06520df4f98cc3535868619a2ba57ea9105e
3
+ size 2031956704