NeoChen1024 commited on
Commit
80b7cca
·
verified ·
1 Parent(s): abc66ab

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -2,22 +2,21 @@
2
  "architectures": [
3
  "LlavaForConditionalGeneration"
4
  ],
 
5
  "image_seq_length": 729,
6
  "image_token_index": 128077,
7
  "model_type": "llava",
8
  "multimodal_projector_bias": true,
9
  "projector_hidden_act": "gelu",
10
  "quantization_config": {
11
- "autoround_version": "0.5.1",
12
  "bits": 4,
13
  "block_name_to_quantize": "model.language_model.layers",
14
- "damp_percent": 0.01,
15
  "data_type": "int",
16
- "desc_act": false,
17
  "group_size": 128,
18
- "quant_method": "gptq",
19
- "sym": true,
20
- "true_sequential": false
21
  },
22
  "text_config": {
23
  "_name_or_path": "meta-llama/Llama-3.1-8B-Instruct",
@@ -27,6 +26,7 @@
27
  "attention_bias": false,
28
  "attention_dropout": 0.0,
29
  "bos_token_id": 128000,
 
30
  "eos_token_id": [
31
  128001,
32
  128008,
@@ -53,18 +53,17 @@
53
  "rope_type": "llama3"
54
  },
55
  "rope_theta": 500000.0,
56
- "torch_dtype": "bfloat16",
57
  "use_cache": true,
58
  "vocab_size": 128256
59
  },
60
- "torch_dtype": "float16",
61
- "transformers_version": "4.52.3",
62
  "vision_config": {
63
  "_name_or_path": "google/siglip2-so400m-patch14-384",
64
  "architectures": [
65
  "SiglipVisionModel"
66
  ],
67
  "attention_dropout": 0.0,
 
68
  "hidden_act": "gelu_pytorch_tanh",
69
  "hidden_size": 1152,
70
  "image_size": 384,
@@ -74,8 +73,7 @@
74
  "num_attention_heads": 16,
75
  "num_channels": 3,
76
  "num_hidden_layers": 27,
77
- "patch_size": 14,
78
- "torch_dtype": "bfloat16"
79
  },
80
  "vision_feature_layer": -2,
81
  "vision_feature_select_strategy": "full"
 
2
  "architectures": [
3
  "LlavaForConditionalGeneration"
4
  ],
5
+ "dtype": "bfloat16",
6
  "image_seq_length": 729,
7
  "image_token_index": 128077,
8
  "model_type": "llava",
9
  "multimodal_projector_bias": true,
10
  "projector_hidden_act": "gelu",
11
  "quantization_config": {
12
+ "autoround_version": "0.7.0",
13
  "bits": 4,
14
  "block_name_to_quantize": "model.language_model.layers",
 
15
  "data_type": "int",
 
16
  "group_size": 128,
17
+ "packing_format": "auto_round:auto_gptq",
18
+ "quant_method": "auto-round",
19
+ "sym": true
20
  },
21
  "text_config": {
22
  "_name_or_path": "meta-llama/Llama-3.1-8B-Instruct",
 
26
  "attention_bias": false,
27
  "attention_dropout": 0.0,
28
  "bos_token_id": 128000,
29
+ "dtype": "bfloat16",
30
  "eos_token_id": [
31
  128001,
32
  128008,
 
53
  "rope_type": "llama3"
54
  },
55
  "rope_theta": 500000.0,
 
56
  "use_cache": true,
57
  "vocab_size": 128256
58
  },
59
+ "transformers_version": "4.56.1",
 
60
  "vision_config": {
61
  "_name_or_path": "google/siglip2-so400m-patch14-384",
62
  "architectures": [
63
  "SiglipVisionModel"
64
  ],
65
  "attention_dropout": 0.0,
66
+ "dtype": "bfloat16",
67
  "hidden_act": "gelu_pytorch_tanh",
68
  "hidden_size": 1152,
69
  "image_size": 384,
 
73
  "num_attention_heads": 16,
74
  "num_channels": 3,
75
  "num_hidden_layers": 27,
76
+ "patch_size": 14
 
77
  },
78
  "vision_feature_layer": -2,
79
  "vision_feature_select_strategy": "full"
generation_config.json CHANGED
@@ -1,10 +1,13 @@
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 128000,
 
4
  "eos_token_id": [
5
  128001,
6
  128008,
7
  128009
8
  ],
9
- "transformers_version": "4.52.3"
 
 
10
  }
 
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 128000,
4
+ "do_sample": true,
5
  "eos_token_id": [
6
  128001,
7
  128008,
8
  128009
9
  ],
10
+ "temperature": 0.6,
11
+ "top_p": 0.9,
12
+ "transformers_version": "4.56.1"
13
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:057d260b474b0e784d4f2e6dde2816fdeff3431fdefa94ea98d3a3ea65bc17a9
3
- size 4983780528
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a615187113e076dfb191e1a97258839889cadfbce156870fe2b205fb35f6519
3
+ size 4979609960
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9b5820b0e324f31de5eadd26cbf980448f3648df296e8e177686c7e72f3a413
3
- size 1648700952
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a8269df045da542a05551d467df4468d5df13d3738f89a93e40647d68954a40
3
+ size 1647861000
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff
 
quantization_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bits": 4,
3
+ "group_size": 128,
4
+ "sym": true,
5
+ "data_type": "int",
6
+ "autoround_version": "0.7.0",
7
+ "block_name_to_quantize": "model.language_model.layers",
8
+ "quant_method": "auto-round",
9
+ "packing_format": "auto_round:auto_gptq"
10
+ }
tokenizer_config.json CHANGED
@@ -2059,5 +2059,5 @@
2059
  ],
2060
  "model_max_length": 131072,
2061
  "processor_class": "LlavaProcessor",
2062
- "tokenizer_class": "PreTrainedTokenizer"
2063
  }
 
2059
  ],
2060
  "model_max_length": 131072,
2061
  "processor_class": "LlavaProcessor",
2062
+ "tokenizer_class": "PreTrainedTokenizerFast"
2063
  }