Luke-Bergen commited on
Commit
708f6a7
·
verified ·
1 Parent(s): 4e5ad0d

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +21 -3
config.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "architectures": [
3
- "MineralNanoForCausalLM"
4
  ],
5
- "model_type": "mineral-nano",
6
  "vocab_size": 32000,
7
  "hidden_size": 768,
8
  "intermediate_size": 3072,
@@ -17,6 +17,7 @@
17
  "pad_token_id": 0,
18
  "bos_token_id": 1,
19
  "eos_token_id": 2,
 
20
  "tie_word_embeddings": false,
21
  "rope_theta": 10000.0,
22
  "rope_scaling": null,
@@ -24,5 +25,22 @@
24
  "attention_dropout": 0.0,
25
  "mlp_bias": false,
26
  "torch_dtype": "bfloat16",
27
- "transformers_version": "4.45.0"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
  }
 
1
  {
2
  "architectures": [
3
+ "MineralNanoForVisionCausalLM"
4
  ],
5
+ "model_type": "mineral-nano-vision",
6
  "vocab_size": 32000,
7
  "hidden_size": 768,
8
  "intermediate_size": 3072,
 
17
  "pad_token_id": 0,
18
  "bos_token_id": 1,
19
  "eos_token_id": 2,
20
+ "image_token_id": 32000,
21
  "tie_word_embeddings": false,
22
  "rope_theta": 10000.0,
23
  "rope_scaling": null,
 
25
  "attention_dropout": 0.0,
26
  "mlp_bias": false,
27
  "torch_dtype": "bfloat16",
28
+ "transformers_version": "4.45.0",
29
+ "vision_config": {
30
+ "hidden_size": 768,
31
+ "image_size": 224,
32
+ "intermediate_size": 3072,
33
+ "num_attention_heads": 12,
34
+ "num_hidden_layers": 12,
35
+ "patch_size": 16,
36
+ "projection_dim": 768,
37
+ "num_channels": 3,
38
+ "layer_norm_eps": 1e-6,
39
+ "dropout": 0.0,
40
+ "attention_dropout": 0.0
41
+ },
42
+ "vision_feature_layer": -2,
43
+ "vision_feature_select_strategy": "default",
44
+ "image_seq_length": 196,
45
+ "num_image_tokens": 196
46
  }