Mistral-Small-3.2-24B-Instruct-2506-pruned-GGUF / scores /Mistral-Small-3.2-24B-Instruct-2506-Q4_K_M-unsloth.md
eaddario's picture
Add GGUF internal file structure
650665a verified

Mistral-Small-3.2-24B-Instruct-2506-Q4_K_M-unsloth.gguf - GGUF Internal File Dump

  • Endian: LITTLE endian

Key Value Metadata Store

There are 44 key-value pairs in this file

POS TYPE Count Key Value
1 UINT32 1 GGUF.version 3
2 UINT64 1 GGUF.tensor_count 363
3 UINT64 1 GGUF.kv_count 41
4 STRING 1 general.architecture llama
5 STRING 1 general.type model
6 STRING 1 general.name Mistral-Small-3.2-24B-Instruct-2506
7 STRING 1 general.version 2506
8 STRING 1 general.finetune Instruct
9 STRING 1 general.basename Mistral-Small-3.2-24B-Instruct-2506
10 STRING 1 general.quantized_by Unsloth
11 STRING 1 general.size_label 24B
12 STRING 1 general.repo_url https://huggingface.co/unsloth
13 UINT32 1 llama.block_count 40
14 UINT32 1 llama.context_length 131072
15 UINT32 1 llama.embedding_length 5120
16 UINT32 1 llama.feed_forward_length 32768
17 UINT32 1 llama.attention.head_count 32
18 UINT32 1 llama.attention.head_count_kv 8
19 FLOAT32 1 llama.rope.freq_base 1000000000.0
20 FLOAT32 1 llama.attention.layer_norm_rms_epsilon 1e-05
21 UINT32 1 llama.attention.key_length 128
22 UINT32 1 llama.attention.value_length 128
23 UINT32 1 llama.vocab_size 131072
24 UINT32 1 llama.rope.dimension_count 128
25 STRING 1 tokenizer.ggml.model gpt2
26 STRING 1 tokenizer.ggml.pre tekken
27 [STRING] 131072 tokenizer.ggml.tokens [ <unk>, <s>, </s>, [INST], [/INST], ... ]
28 [INT32] 131072 tokenizer.ggml.token_type [ 3, 3, 3, 3, 3, 3, 3, ... ]
29 [STRING] 269443 tokenizer.ggml.merges [ Ġ Ġ, Ġ t, e r, i n, Ġ ĠĠĠ, ... ]
30 UINT32 1 tokenizer.ggml.bos_token_id 1
31 UINT32 1 tokenizer.ggml.eos_token_id 2
32 UINT32 1 tokenizer.ggml.unknown_token_id 0
33 UINT32 1 tokenizer.ggml.padding_token_id 11
34 BOOL 1 tokenizer.ggml.add_bos_token True
35 BOOL 1 tokenizer.ggml.add_sep_token False
36 BOOL 1 tokenizer.ggml.add_eos_token False
37 STRING 1 tokenizer.chat_template {#- Copyright 2025-present the...Version 2.0 (the "License") #}
38 BOOL 1 tokenizer.ggml.add_space_prefix False
39 UINT32 1 general.quantization_version 2
40 UINT32 1 general.file_type 15
41 STRING 1 quantize.imatrix.file Mistral-Small-3.2-24B-Instruct-2506-GGUF/imatrix_unsloth.dat
42 STRING 1 quantize.imatrix.dataset unsloth_calibration_Mistral-Small-3.2-24B-Instruct-2506.txt
43 UINT32 1 quantize.imatrix.entries_count 280
44 UINT32 1 quantize.imatrix.chunks_count 218

Tensors Overview ~24B Elements

Total number of elements in all tensors: 23572403200 Elements

Tensor Data Offset

This table contains the offset and data segment relative to start of file

T_ID Tensor Layer Name Data Offset (B) Data Size (B)
0 output.weight 0x787d80 0x20d00000
1 output_norm.weight 0x21487d80 0x5000
2 token_embd.weight 0x2148cd80 0x16800000
3 blk.0.attn_k.weight 0x37c8cd80 0x2d0000
4 blk.0.attn_norm.weight 0x37f5cd80 0x5000
5 blk.0.attn_output.weight 0x37f61d80 0xb40000
6 blk.0.attn_q.weight 0x38aa1d80 0xb40000
7 blk.0.attn_v.weight 0x395e1d80 0x41a000
8 blk.0.ffn_down.weight 0x399fbd80 0x8340000
9 blk.0.ffn_gate.weight 0x41d3bd80 0x5a00000
10 blk.0.ffn_norm.weight 0x4773bd80 0x5000
11 blk.0.ffn_up.weight 0x47740d80 0x5a00000
12 blk.1.attn_k.weight 0x4d140d80 0x2d0000
13 blk.1.attn_norm.weight 0x4d410d80 0x5000
14 blk.1.attn_output.weight 0x4d415d80 0xb40000
15 blk.1.attn_q.weight 0x4df55d80 0xb40000
16 blk.1.attn_v.weight 0x4ea95d80 0x41a000
17 blk.1.ffn_down.weight 0x4eeafd80 0x8340000
18 blk.1.ffn_gate.weight 0x571efd80 0x5a00000
19 blk.1.ffn_norm.weight 0x5cbefd80 0x5000
20 blk.1.ffn_up.weight 0x5cbf4d80 0x5a00000
21 blk.2.attn_k.weight 0x625f4d80 0x2d0000
22 blk.2.attn_norm.weight 0x628c4d80 0x5000
23 blk.2.attn_output.weight 0x628c9d80 0xb40000
24 blk.2.attn_q.weight 0x63409d80 0xb40000
25 blk.2.attn_v.weight 0x63f49d80 0x41a000
26 blk.2.ffn_down.weight 0x64363d80 0x8340000
27 blk.2.ffn_gate.weight 0x6c6a3d80 0x5a00000
28 blk.2.ffn_norm.weight 0x720a3d80 0x5000
29 blk.2.ffn_up.weight 0x720a8d80 0x5a00000
30 blk.3.attn_k.weight 0x77aa8d80 0x2d0000
31 blk.3.attn_norm.weight 0x77d78d80 0x5000
32 blk.3.attn_output.weight 0x77d7dd80 0xb40000
33 blk.3.attn_q.weight 0x788bdd80 0xb40000
34 blk.3.attn_v.weight 0x793fdd80 0x41a000
35 blk.3.ffn_down.weight 0x79817d80 0x8340000
36 blk.3.ffn_gate.weight 0x81b57d80 0x5a00000
37 blk.3.ffn_norm.weight 0x87557d80 0x5000
38 blk.3.ffn_up.weight 0x8755cd80 0x5a00000
39 blk.4.attn_k.weight 0x8cf5cd80 0x2d0000
40 blk.4.attn_norm.weight 0x8d22cd80 0x5000
41 blk.4.attn_output.weight 0x8d231d80 0xb40000
42 blk.4.attn_q.weight 0x8dd71d80 0xb40000
43 blk.4.attn_v.weight 0x8e8b1d80 0x41a000
44 blk.4.ffn_down.weight 0x8eccbd80 0x8340000
45 blk.4.ffn_gate.weight 0x9700bd80 0x5a00000
46 blk.4.ffn_norm.weight 0x9ca0bd80 0x5000
47 blk.4.ffn_up.weight 0x9ca10d80 0x5a00000
48 blk.5.attn_k.weight 0xa2410d80 0x2d0000
49 blk.5.attn_norm.weight 0xa26e0d80 0x5000
50 blk.5.attn_output.weight 0xa26e5d80 0xb40000
51 blk.5.attn_q.weight 0xa3225d80 0xb40000
52 blk.5.attn_v.weight 0xa3d65d80 0x2d0000
53 blk.5.ffn_down.weight 0xa4035d80 0x5a00000
54 blk.5.ffn_gate.weight 0xa9a35d80 0x5a00000
55 blk.5.ffn_norm.weight 0xaf435d80 0x5000
56 blk.5.ffn_up.weight 0xaf43ad80 0x5a00000
57 blk.6.attn_k.weight 0xb4e3ad80 0x2d0000
58 blk.6.attn_norm.weight 0xb510ad80 0x5000
59 blk.6.attn_output.weight 0xb510fd80 0xb40000
60 blk.6.attn_q.weight 0xb5c4fd80 0xb40000
61 blk.6.attn_v.weight 0xb678fd80 0x2d0000
62 blk.6.ffn_down.weight 0xb6a5fd80 0x5a00000
63 blk.6.ffn_gate.weight 0xbc45fd80 0x5a00000
64 blk.6.ffn_norm.weight 0xc1e5fd80 0x5000
65 blk.6.ffn_up.weight 0xc1e64d80 0x5a00000
66 blk.7.attn_k.weight 0xc7864d80 0x2d0000
67 blk.7.attn_norm.weight 0xc7b34d80 0x5000
68 blk.7.attn_output.weight 0xc7b39d80 0xb40000
69 blk.7.attn_q.weight 0xc8679d80 0xb40000
70 blk.7.attn_v.weight 0xc91b9d80 0x41a000
71 blk.7.ffn_down.weight 0xc95d3d80 0x8340000
72 blk.7.ffn_gate.weight 0xd1913d80 0x5a00000
73 blk.7.ffn_norm.weight 0xd7313d80 0x5000
74 blk.7.ffn_up.weight 0xd7318d80 0x5a00000
75 blk.8.attn_k.weight 0xdcd18d80 0x2d0000
76 blk.8.attn_norm.weight 0xdcfe8d80 0x5000
77 blk.8.attn_output.weight 0xdcfedd80 0xb40000
78 blk.8.attn_q.weight 0xddb2dd80 0xb40000
79 blk.8.attn_v.weight 0xde66dd80 0x2d0000
80 blk.8.ffn_down.weight 0xde93dd80 0x5a00000
81 blk.8.ffn_gate.weight 0xe433dd80 0x5a00000
82 blk.8.ffn_norm.weight 0xe9d3dd80 0x5000
83 blk.8.ffn_up.weight 0xe9d42d80 0x5a00000
84 blk.9.attn_k.weight 0xef742d80 0x2d0000
85 blk.9.attn_norm.weight 0xefa12d80 0x5000
86 blk.9.attn_output.weight 0xefa17d80 0xb40000
87 blk.9.attn_q.weight 0xf0557d80 0xb40000
88 blk.9.attn_v.weight 0xf1097d80 0x2d0000
89 blk.9.ffn_down.weight 0xf1367d80 0x5a00000
90 blk.9.ffn_gate.weight 0xf6d67d80 0x5a00000
91 blk.9.ffn_norm.weight 0xfc767d80 0x5000
92 blk.9.ffn_up.weight 0xfc76cd80 0x5a00000
93 blk.10.attn_k.weight 0x10216cd80 0x2d0000
94 blk.10.attn_norm.weight 0x10243cd80 0x5000
95 blk.10.attn_output.weight 0x102441d80 0xb40000
96 blk.10.attn_q.weight 0x102f81d80 0xb40000
97 blk.10.attn_v.weight 0x103ac1d80 0x41a000
98 blk.10.ffn_down.weight 0x103edbd80 0x8340000
99 blk.10.ffn_gate.weight 0x10c21bd80 0x5a00000
100 blk.10.ffn_norm.weight 0x111c1bd80 0x5000
101 blk.10.ffn_up.weight 0x111c20d80 0x5a00000
102 blk.11.attn_k.weight 0x117620d80 0x2d0000
103 blk.11.attn_norm.weight 0x1178f0d80 0x5000
104 blk.11.attn_output.weight 0x1178f5d80 0xb40000
105 blk.11.attn_q.weight 0x118435d80 0xb40000
106 blk.11.attn_v.weight 0x118f75d80 0x2d0000
107 blk.11.ffn_down.weight 0x119245d80 0x5a00000
108 blk.11.ffn_gate.weight 0x11ec45d80 0x5a00000
109 blk.11.ffn_norm.weight 0x124645d80 0x5000
110 blk.11.ffn_up.weight 0x12464ad80 0x5a00000
111 blk.12.attn_k.weight 0x12a04ad80 0x2d0000
112 blk.12.attn_norm.weight 0x12a31ad80 0x5000
113 blk.12.attn_output.weight 0x12a31fd80 0xb40000
114 blk.12.attn_q.weight 0x12ae5fd80 0xb40000
115 blk.12.attn_v.weight 0x12b99fd80 0x2d0000
116 blk.12.ffn_down.weight 0x12bc6fd80 0x5a00000
117 blk.12.ffn_gate.weight 0x13166fd80 0x5a00000
118 blk.12.ffn_norm.weight 0x13706fd80 0x5000
119 blk.12.ffn_up.weight 0x137074d80 0x5a00000
120 blk.13.attn_k.weight 0x13ca74d80 0x2d0000
121 blk.13.attn_norm.weight 0x13cd44d80 0x5000
122 blk.13.attn_output.weight 0x13cd49d80 0xb40000
123 blk.13.attn_q.weight 0x13d889d80 0xb40000
124 blk.13.attn_v.weight 0x13e3c9d80 0x41a000
125 blk.13.ffn_down.weight 0x13e7e3d80 0x8340000
126 blk.13.ffn_gate.weight 0x146b23d80 0x5a00000
127 blk.13.ffn_norm.weight 0x14c523d80 0x5000
128 blk.13.ffn_up.weight 0x14c528d80 0x5a00000
129 blk.14.attn_k.weight 0x151f28d80 0x2d0000
130 blk.14.attn_norm.weight 0x1521f8d80 0x5000
131 blk.14.attn_output.weight 0x1521fdd80 0xb40000
132 blk.14.attn_q.weight 0x152d3dd80 0xb40000
133 blk.14.attn_v.weight 0x15387dd80 0x2d0000
134 blk.14.ffn_down.weight 0x153b4dd80 0x5a00000
135 blk.14.ffn_gate.weight 0x15954dd80 0x5a00000
136 blk.14.ffn_norm.weight 0x15ef4dd80 0x5000
137 blk.14.ffn_up.weight 0x15ef52d80 0x5a00000
138 blk.15.attn_k.weight 0x164952d80 0x2d0000
139 blk.15.attn_norm.weight 0x164c22d80 0x5000
140 blk.15.attn_output.weight 0x164c27d80 0xb40000
141 blk.15.attn_q.weight 0x165767d80 0xb40000
142 blk.15.attn_v.weight 0x1662a7d80 0x2d0000
143 blk.15.ffn_down.weight 0x166577d80 0x5a00000
144 blk.15.ffn_gate.weight 0x16bf77d80 0x5a00000
145 blk.15.ffn_norm.weight 0x171977d80 0x5000
146 blk.15.ffn_up.weight 0x17197cd80 0x5a00000
147 blk.16.attn_k.weight 0x17737cd80 0x2d0000
148 blk.16.attn_norm.weight 0x17764cd80 0x5000
149 blk.16.attn_output.weight 0x177651d80 0xb40000
150 blk.16.attn_q.weight 0x178191d80 0xb40000
151 blk.16.attn_v.weight 0x178cd1d80 0x41a000
152 blk.16.ffn_down.weight 0x1790ebd80 0x8340000
153 blk.16.ffn_gate.weight 0x18142bd80 0x5a00000
154 blk.16.ffn_norm.weight 0x186e2bd80 0x5000
155 blk.16.ffn_up.weight 0x186e30d80 0x5a00000
156 blk.17.attn_k.weight 0x18c830d80 0x2d0000
157 blk.17.attn_norm.weight 0x18cb00d80 0x5000
158 blk.17.attn_output.weight 0x18cb05d80 0xb40000
159 blk.17.attn_q.weight 0x18d645d80 0xb40000
160 blk.17.attn_v.weight 0x18e185d80 0x2d0000
161 blk.17.ffn_down.weight 0x18e455d80 0x5a00000
162 blk.17.ffn_gate.weight 0x193e55d80 0x5a00000
163 blk.17.ffn_norm.weight 0x199855d80 0x5000
164 blk.17.ffn_up.weight 0x19985ad80 0x5a00000
165 blk.18.attn_k.weight 0x19f25ad80 0x2d0000
166 blk.18.attn_norm.weight 0x19f52ad80 0x5000
167 blk.18.attn_output.weight 0x19f52fd80 0xb40000
168 blk.18.attn_q.weight 0x1a006fd80 0xb40000
169 blk.18.attn_v.weight 0x1a0bafd80 0x2d0000
170 blk.18.ffn_down.weight 0x1a0e7fd80 0x5a00000
171 blk.18.ffn_gate.weight 0x1a687fd80 0x5a00000
172 blk.18.ffn_norm.weight 0x1ac27fd80 0x5000
173 blk.18.ffn_up.weight 0x1ac284d80 0x5a00000
174 blk.19.attn_k.weight 0x1b1c84d80 0x2d0000
175 blk.19.attn_norm.weight 0x1b1f54d80 0x5000
176 blk.19.attn_output.weight 0x1b1f59d80 0xb40000
177 blk.19.attn_q.weight 0x1b2a99d80 0xb40000
178 blk.19.attn_v.weight 0x1b35d9d80 0x41a000
179 blk.19.ffn_down.weight 0x1b39f3d80 0x8340000
180 blk.19.ffn_gate.weight 0x1bbd33d80 0x5a00000
181 blk.19.ffn_norm.weight 0x1c1733d80 0x5000
182 blk.19.ffn_up.weight 0x1c1738d80 0x5a00000
183 blk.20.attn_k.weight 0x1c7138d80 0x2d0000
184 blk.20.attn_norm.weight 0x1c7408d80 0x5000
185 blk.20.attn_output.weight 0x1c740dd80 0xb40000
186 blk.20.attn_q.weight 0x1c7f4dd80 0xb40000
187 blk.20.attn_v.weight 0x1c8a8dd80 0x2d0000
188 blk.20.ffn_down.weight 0x1c8d5dd80 0x5a00000
189 blk.20.ffn_gate.weight 0x1ce75dd80 0x5a00000
190 blk.20.ffn_norm.weight 0x1d415dd80 0x5000
191 blk.20.ffn_up.weight 0x1d4162d80 0x5a00000
192 blk.21.attn_k.weight 0x1d9b62d80 0x2d0000
193 blk.21.attn_norm.weight 0x1d9e32d80 0x5000
194 blk.21.attn_output.weight 0x1d9e37d80 0xb40000
195 blk.21.attn_q.weight 0x1da977d80 0xb40000
196 blk.21.attn_v.weight 0x1db4b7d80 0x2d0000
197 blk.21.ffn_down.weight 0x1db787d80 0x5a00000
198 blk.21.ffn_gate.weight 0x1e1187d80 0x5a00000
199 blk.21.ffn_norm.weight 0x1e6b87d80 0x5000
200 blk.21.ffn_up.weight 0x1e6b8cd80 0x5a00000
201 blk.22.attn_k.weight 0x1ec58cd80 0x2d0000
202 blk.22.attn_norm.weight 0x1ec85cd80 0x5000
203 blk.22.attn_output.weight 0x1ec861d80 0xb40000
204 blk.22.attn_q.weight 0x1ed3a1d80 0xb40000
205 blk.22.attn_v.weight 0x1edee1d80 0x41a000
206 blk.22.ffn_down.weight 0x1ee2fbd80 0x8340000
207 blk.22.ffn_gate.weight 0x1f663bd80 0x5a00000
208 blk.22.ffn_norm.weight 0x1fc03bd80 0x5000
209 blk.22.ffn_up.weight 0x1fc040d80 0x5a00000
210 blk.23.attn_k.weight 0x201a40d80 0x2d0000
211 blk.23.attn_norm.weight 0x201d10d80 0x5000
212 blk.23.attn_output.weight 0x201d15d80 0xb40000
213 blk.23.attn_q.weight 0x202855d80 0xb40000
214 blk.23.attn_v.weight 0x203395d80 0x2d0000
215 blk.23.ffn_down.weight 0x203665d80 0x5a00000
216 blk.23.ffn_gate.weight 0x209065d80 0x5a00000
217 blk.23.ffn_norm.weight 0x20ea65d80 0x5000
218 blk.23.ffn_up.weight 0x20ea6ad80 0x5a00000
219 blk.24.attn_k.weight 0x21446ad80 0x2d0000
220 blk.24.attn_norm.weight 0x21473ad80 0x5000
221 blk.24.attn_output.weight 0x21473fd80 0xb40000
222 blk.24.attn_q.weight 0x21527fd80 0xb40000
223 blk.24.attn_v.weight 0x215dbfd80 0x2d0000
224 blk.24.ffn_down.weight 0x21608fd80 0x5a00000
225 blk.24.ffn_gate.weight 0x21ba8fd80 0x5a00000
226 blk.24.ffn_norm.weight 0x22148fd80 0x5000
227 blk.24.ffn_up.weight 0x221494d80 0x5a00000
228 blk.25.attn_k.weight 0x226e94d80 0x2d0000
229 blk.25.attn_norm.weight 0x227164d80 0x5000
230 blk.25.attn_output.weight 0x227169d80 0xb40000
231 blk.25.attn_q.weight 0x227ca9d80 0xb40000
232 blk.25.attn_v.weight 0x2287e9d80 0x41a000
233 blk.25.ffn_down.weight 0x228c03d80 0x8340000
234 blk.25.ffn_gate.weight 0x230f43d80 0x5a00000
235 blk.25.ffn_norm.weight 0x236943d80 0x5000
236 blk.25.ffn_up.weight 0x236948d80 0x5a00000
237 blk.26.attn_k.weight 0x23c348d80 0x2d0000
238 blk.26.attn_norm.weight 0x23c618d80 0x5000
239 blk.26.attn_output.weight 0x23c61dd80 0xb40000
240 blk.26.attn_q.weight 0x23d15dd80 0xb40000
241 blk.26.attn_v.weight 0x23dc9dd80 0x2d0000
242 blk.26.ffn_down.weight 0x23df6dd80 0x5a00000
243 blk.26.ffn_gate.weight 0x24396dd80 0x5a00000
244 blk.26.ffn_norm.weight 0x24936dd80 0x5000
245 blk.26.ffn_up.weight 0x249372d80 0x5a00000
246 blk.27.attn_k.weight 0x24ed72d80 0x2d0000
247 blk.27.attn_norm.weight 0x24f042d80 0x5000
248 blk.27.attn_output.weight 0x24f047d80 0xb40000
249 blk.27.attn_q.weight 0x24fb87d80 0xb40000
250 blk.27.attn_v.weight 0x2506c7d80 0x2d0000
251 blk.27.ffn_down.weight 0x250997d80 0x5a00000
252 blk.27.ffn_gate.weight 0x256397d80 0x5a00000
253 blk.27.ffn_norm.weight 0x25bd97d80 0x5000
254 blk.27.ffn_up.weight 0x25bd9cd80 0x5a00000
255 blk.28.attn_k.weight 0x26179cd80 0x2d0000
256 blk.28.attn_norm.weight 0x261a6cd80 0x5000
257 blk.28.attn_output.weight 0x261a71d80 0xb40000
258 blk.28.attn_q.weight 0x2625b1d80 0xb40000
259 blk.28.attn_v.weight 0x2630f1d80 0x41a000
260 blk.28.ffn_down.weight 0x26350bd80 0x8340000
261 blk.28.ffn_gate.weight 0x26b84bd80 0x5a00000
262 blk.28.ffn_norm.weight 0x27124bd80 0x5000
263 blk.28.ffn_up.weight 0x271250d80 0x5a00000
264 blk.29.attn_k.weight 0x276c50d80 0x2d0000
265 blk.29.attn_norm.weight 0x276f20d80 0x5000
266 blk.29.attn_output.weight 0x276f25d80 0xb40000
267 blk.29.attn_q.weight 0x277a65d80 0xb40000
268 blk.29.attn_v.weight 0x2785a5d80 0x2d0000
269 blk.29.ffn_down.weight 0x278875d80 0x5a00000
270 blk.29.ffn_gate.weight 0x27e275d80 0x5a00000
271 blk.29.ffn_norm.weight 0x283c75d80 0x5000
272 blk.29.ffn_up.weight 0x283c7ad80 0x5a00000
273 blk.30.attn_k.weight 0x28967ad80 0x2d0000
274 blk.30.attn_norm.weight 0x28994ad80 0x5000
275 blk.30.attn_output.weight 0x28994fd80 0xb40000
276 blk.30.attn_q.weight 0x28a48fd80 0xb40000
277 blk.30.attn_v.weight 0x28afcfd80 0x2d0000
278 blk.30.ffn_down.weight 0x28b29fd80 0x5a00000
279 blk.30.ffn_gate.weight 0x290c9fd80 0x5a00000
280 blk.30.ffn_norm.weight 0x29669fd80 0x5000
281 blk.30.ffn_up.weight 0x2966a4d80 0x5a00000
282 blk.31.attn_k.weight 0x29c0a4d80 0x2d0000
283 blk.31.attn_norm.weight 0x29c374d80 0x5000
284 blk.31.attn_output.weight 0x29c379d80 0xb40000
285 blk.31.attn_q.weight 0x29ceb9d80 0xb40000
286 blk.31.attn_v.weight 0x29d9f9d80 0x41a000
287 blk.31.ffn_down.weight 0x29de13d80 0x8340000
288 blk.31.ffn_gate.weight 0x2a6153d80 0x5a00000
289 blk.31.ffn_norm.weight 0x2abb53d80 0x5000
290 blk.31.ffn_up.weight 0x2abb58d80 0x5a00000
291 blk.32.attn_k.weight 0x2b1558d80 0x2d0000
292 blk.32.attn_norm.weight 0x2b1828d80 0x5000
293 blk.32.attn_output.weight 0x2b182dd80 0xb40000
294 blk.32.attn_q.weight 0x2b236dd80 0xb40000
295 blk.32.attn_v.weight 0x2b2eadd80 0x2d0000
296 blk.32.ffn_down.weight 0x2b317dd80 0x5a00000
297 blk.32.ffn_gate.weight 0x2b8b7dd80 0x5a00000
298 blk.32.ffn_norm.weight 0x2be57dd80 0x5000
299 blk.32.ffn_up.weight 0x2be582d80 0x5a00000
300 blk.33.attn_k.weight 0x2c3f82d80 0x2d0000
301 blk.33.attn_norm.weight 0x2c4252d80 0x5000
302 blk.33.attn_output.weight 0x2c4257d80 0xb40000
303 blk.33.attn_q.weight 0x2c4d97d80 0xb40000
304 blk.33.attn_v.weight 0x2c58d7d80 0x2d0000
305 blk.33.ffn_down.weight 0x2c5ba7d80 0x5a00000
306 blk.33.ffn_gate.weight 0x2cb5a7d80 0x5a00000
307 blk.33.ffn_norm.weight 0x2d0fa7d80 0x5000
308 blk.33.ffn_up.weight 0x2d0facd80 0x5a00000
309 blk.34.attn_k.weight 0x2d69acd80 0x2d0000
310 blk.34.attn_norm.weight 0x2d6c7cd80 0x5000
311 blk.34.attn_output.weight 0x2d6c81d80 0xb40000
312 blk.34.attn_q.weight 0x2d77c1d80 0xb40000
313 blk.34.attn_v.weight 0x2d8301d80 0x41a000
314 blk.34.ffn_down.weight 0x2d871bd80 0x8340000
315 blk.34.ffn_gate.weight 0x2e0a5bd80 0x5a00000
316 blk.34.ffn_norm.weight 0x2e645bd80 0x5000
317 blk.34.ffn_up.weight 0x2e6460d80 0x5a00000
318 blk.35.attn_k.weight 0x2ebe60d80 0x2d0000
319 blk.35.attn_norm.weight 0x2ec130d80 0x5000
320 blk.35.attn_output.weight 0x2ec135d80 0xb40000
321 blk.35.attn_q.weight 0x2ecc75d80 0xb40000
322 blk.35.attn_v.weight 0x2ed7b5d80 0x41a000
323 blk.35.ffn_down.weight 0x2edbcfd80 0x8340000
324 blk.35.ffn_gate.weight 0x2f5f0fd80 0x5a00000
325 blk.35.ffn_norm.weight 0x2fb90fd80 0x5000
326 blk.35.ffn_up.weight 0x2fb914d80 0x5a00000
327 blk.36.attn_k.weight 0x301314d80 0x2d0000
328 blk.36.attn_norm.weight 0x3015e4d80 0x5000
329 blk.36.attn_output.weight 0x3015e9d80 0xb40000
330 blk.36.attn_q.weight 0x302129d80 0xb40000
331 blk.36.attn_v.weight 0x302c69d80 0x41a000
332 blk.36.ffn_down.weight 0x303083d80 0x8340000
333 blk.36.ffn_gate.weight 0x30b3c3d80 0x5a00000
334 blk.36.ffn_norm.weight 0x310dc3d80 0x5000
335 blk.36.ffn_up.weight 0x310dc8d80 0x5a00000
336 blk.37.attn_k.weight 0x3167c8d80 0x2d0000
337 blk.37.attn_norm.weight 0x316a98d80 0x5000
338 blk.37.attn_output.weight 0x316a9dd80 0xb40000
339 blk.37.attn_q.weight 0x3175ddd80 0xb40000
340 blk.37.attn_v.weight 0x31811dd80 0x41a000
341 blk.37.ffn_down.weight 0x318537d80 0x8340000
342 blk.37.ffn_gate.weight 0x320877d80 0x5a00000
343 blk.37.ffn_norm.weight 0x326277d80 0x5000
344 blk.37.ffn_up.weight 0x32627cd80 0x5a00000
345 blk.38.attn_k.weight 0x32bc7cd80 0x2d0000
346 blk.38.attn_norm.weight 0x32bf4cd80 0x5000
347 blk.38.attn_output.weight 0x32bf51d80 0xb40000
348 blk.38.attn_q.weight 0x32ca91d80 0xb40000
349 blk.38.attn_v.weight 0x32d5d1d80 0x41a000
350 blk.38.ffn_down.weight 0x32d9ebd80 0x8340000
351 blk.38.ffn_gate.weight 0x335d2bd80 0x5a00000
352 blk.38.ffn_norm.weight 0x33b72bd80 0x5000
353 blk.38.ffn_up.weight 0x33b730d80 0x5a00000
354 blk.39.attn_k.weight 0x341130d80 0x2d0000
355 blk.39.attn_norm.weight 0x341400d80 0x5000
356 blk.39.attn_output.weight 0x341405d80 0xb40000
357 blk.39.attn_q.weight 0x341f45d80 0xb40000
358 blk.39.attn_v.weight 0x342a85d80 0x41a000
359 blk.39.ffn_down.weight 0x342e9fd80 0x8340000
360 blk.39.ffn_gate.weight 0x34b1dfd80 0x5a00000
361 blk.39.ffn_norm.weight 0x350bdfd80 0x5000
362 blk.39.ffn_up.weight 0x350be4d80 0x5a00000

Base Tensor Group : ~1B Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
0 output.weight Output (W) (~671M) 671088640 5120 x 131072 x 1 x 1 Q6_K 6.5625
1 output_norm.weight Output Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
2 token_embd.weight Token Embedding (W) (~671M) 671088640 5120 x 131072 x 1 x 1 Q4_K 4.5000
  • Total elements in base: ( ~1B) 1342182400
  • Percentage of total elements: 5.69%
  • Bits per Weight (BPW) for base: 5.5314 bits

Block 0 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
3 blk.0.attn_k.weight Block 0 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
4 blk.0.attn_norm.weight Block 0 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
5 blk.0.attn_output.weight Block 0 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
6 blk.0.attn_q.weight Block 0 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
7 blk.0.attn_v.weight Block 0 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
8 blk.0.ffn_down.weight Block 0 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
9 blk.0.ffn_gate.weight Block 0 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
10 blk.0.ffn_norm.weight Block 0 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
11 blk.0.ffn_up.weight Block 0 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.0: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.0: 5.1426 bits

Block 1 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
12 blk.1.attn_k.weight Block 1 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
13 blk.1.attn_norm.weight Block 1 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
14 blk.1.attn_output.weight Block 1 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
15 blk.1.attn_q.weight Block 1 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
16 blk.1.attn_v.weight Block 1 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
17 blk.1.ffn_down.weight Block 1 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
18 blk.1.ffn_gate.weight Block 1 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
19 blk.1.ffn_norm.weight Block 1 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
20 blk.1.ffn_up.weight Block 1 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.1: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.1: 5.1426 bits

Block 2 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
21 blk.2.attn_k.weight Block 2 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
22 blk.2.attn_norm.weight Block 2 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
23 blk.2.attn_output.weight Block 2 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
24 blk.2.attn_q.weight Block 2 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
25 blk.2.attn_v.weight Block 2 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
26 blk.2.ffn_down.weight Block 2 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
27 blk.2.ffn_gate.weight Block 2 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
28 blk.2.ffn_norm.weight Block 2 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
29 blk.2.ffn_up.weight Block 2 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.2: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.2: 5.1426 bits

Block 3 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
30 blk.3.attn_k.weight Block 3 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
31 blk.3.attn_norm.weight Block 3 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
32 blk.3.attn_output.weight Block 3 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
33 blk.3.attn_q.weight Block 3 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
34 blk.3.attn_v.weight Block 3 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
35 blk.3.ffn_down.weight Block 3 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
36 blk.3.ffn_gate.weight Block 3 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
37 blk.3.ffn_norm.weight Block 3 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
38 blk.3.ffn_up.weight Block 3 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.3: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.3: 5.1426 bits

Block 4 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
39 blk.4.attn_k.weight Block 4 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
40 blk.4.attn_norm.weight Block 4 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
41 blk.4.attn_output.weight Block 4 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
42 blk.4.attn_q.weight Block 4 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
43 blk.4.attn_v.weight Block 4 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
44 blk.4.ffn_down.weight Block 4 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
45 blk.4.ffn_gate.weight Block 4 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
46 blk.4.ffn_norm.weight Block 4 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
47 blk.4.ffn_up.weight Block 4 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.4: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.4: 5.1426 bits

Block 5 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
48 blk.5.attn_k.weight Block 5 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
49 blk.5.attn_norm.weight Block 5 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
50 blk.5.attn_output.weight Block 5 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
51 blk.5.attn_q.weight Block 5 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
52 blk.5.attn_v.weight Block 5 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
53 blk.5.ffn_down.weight Block 5 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
54 blk.5.ffn_gate.weight Block 5 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
55 blk.5.ffn_norm.weight Block 5 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
56 blk.5.ffn_up.weight Block 5 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.5: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.5: 4.5005 bits

Block 6 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
57 blk.6.attn_k.weight Block 6 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
58 blk.6.attn_norm.weight Block 6 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
59 blk.6.attn_output.weight Block 6 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
60 blk.6.attn_q.weight Block 6 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
61 blk.6.attn_v.weight Block 6 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
62 blk.6.ffn_down.weight Block 6 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
63 blk.6.ffn_gate.weight Block 6 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
64 blk.6.ffn_norm.weight Block 6 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
65 blk.6.ffn_up.weight Block 6 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.6: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.6: 4.5005 bits

Block 7 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
66 blk.7.attn_k.weight Block 7 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
67 blk.7.attn_norm.weight Block 7 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
68 blk.7.attn_output.weight Block 7 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
69 blk.7.attn_q.weight Block 7 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
70 blk.7.attn_v.weight Block 7 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
71 blk.7.ffn_down.weight Block 7 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
72 blk.7.ffn_gate.weight Block 7 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
73 blk.7.ffn_norm.weight Block 7 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
74 blk.7.ffn_up.weight Block 7 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.7: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.7: 5.1426 bits

Block 8 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
75 blk.8.attn_k.weight Block 8 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
76 blk.8.attn_norm.weight Block 8 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
77 blk.8.attn_output.weight Block 8 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
78 blk.8.attn_q.weight Block 8 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
79 blk.8.attn_v.weight Block 8 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
80 blk.8.ffn_down.weight Block 8 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
81 blk.8.ffn_gate.weight Block 8 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
82 blk.8.ffn_norm.weight Block 8 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
83 blk.8.ffn_up.weight Block 8 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.8: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.8: 4.5005 bits

Block 9 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
84 blk.9.attn_k.weight Block 9 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
85 blk.9.attn_norm.weight Block 9 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
86 blk.9.attn_output.weight Block 9 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
87 blk.9.attn_q.weight Block 9 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
88 blk.9.attn_v.weight Block 9 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
89 blk.9.ffn_down.weight Block 9 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
90 blk.9.ffn_gate.weight Block 9 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
91 blk.9.ffn_norm.weight Block 9 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
92 blk.9.ffn_up.weight Block 9 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.9: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.9: 4.5005 bits

Block 10 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
93 blk.10.attn_k.weight Block 10 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
94 blk.10.attn_norm.weight Block 10 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
95 blk.10.attn_output.weight Block 10 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
96 blk.10.attn_q.weight Block 10 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
97 blk.10.attn_v.weight Block 10 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
98 blk.10.ffn_down.weight Block 10 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
99 blk.10.ffn_gate.weight Block 10 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
100 blk.10.ffn_norm.weight Block 10 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
101 blk.10.ffn_up.weight Block 10 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.10: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.10: 5.1426 bits

Block 11 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
102 blk.11.attn_k.weight Block 11 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
103 blk.11.attn_norm.weight Block 11 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
104 blk.11.attn_output.weight Block 11 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
105 blk.11.attn_q.weight Block 11 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
106 blk.11.attn_v.weight Block 11 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
107 blk.11.ffn_down.weight Block 11 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
108 blk.11.ffn_gate.weight Block 11 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
109 blk.11.ffn_norm.weight Block 11 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
110 blk.11.ffn_up.weight Block 11 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.11: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.11: 4.5005 bits

Block 12 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
111 blk.12.attn_k.weight Block 12 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
112 blk.12.attn_norm.weight Block 12 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
113 blk.12.attn_output.weight Block 12 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
114 blk.12.attn_q.weight Block 12 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
115 blk.12.attn_v.weight Block 12 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
116 blk.12.ffn_down.weight Block 12 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
117 blk.12.ffn_gate.weight Block 12 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
118 blk.12.ffn_norm.weight Block 12 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
119 blk.12.ffn_up.weight Block 12 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.12: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.12: 4.5005 bits

Block 13 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
120 blk.13.attn_k.weight Block 13 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
121 blk.13.attn_norm.weight Block 13 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
122 blk.13.attn_output.weight Block 13 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
123 blk.13.attn_q.weight Block 13 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
124 blk.13.attn_v.weight Block 13 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
125 blk.13.ffn_down.weight Block 13 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
126 blk.13.ffn_gate.weight Block 13 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
127 blk.13.ffn_norm.weight Block 13 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
128 blk.13.ffn_up.weight Block 13 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.13: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.13: 5.1426 bits

Block 14 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
129 blk.14.attn_k.weight Block 14 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
130 blk.14.attn_norm.weight Block 14 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
131 blk.14.attn_output.weight Block 14 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
132 blk.14.attn_q.weight Block 14 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
133 blk.14.attn_v.weight Block 14 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
134 blk.14.ffn_down.weight Block 14 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
135 blk.14.ffn_gate.weight Block 14 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
136 blk.14.ffn_norm.weight Block 14 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
137 blk.14.ffn_up.weight Block 14 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.14: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.14: 4.5005 bits

Block 15 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
138 blk.15.attn_k.weight Block 15 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
139 blk.15.attn_norm.weight Block 15 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
140 blk.15.attn_output.weight Block 15 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
141 blk.15.attn_q.weight Block 15 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
142 blk.15.attn_v.weight Block 15 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
143 blk.15.ffn_down.weight Block 15 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
144 blk.15.ffn_gate.weight Block 15 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
145 blk.15.ffn_norm.weight Block 15 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
146 blk.15.ffn_up.weight Block 15 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.15: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.15: 4.5005 bits

Block 16 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
147 blk.16.attn_k.weight Block 16 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
148 blk.16.attn_norm.weight Block 16 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
149 blk.16.attn_output.weight Block 16 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
150 blk.16.attn_q.weight Block 16 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
151 blk.16.attn_v.weight Block 16 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
152 blk.16.ffn_down.weight Block 16 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
153 blk.16.ffn_gate.weight Block 16 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
154 blk.16.ffn_norm.weight Block 16 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
155 blk.16.ffn_up.weight Block 16 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.16: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.16: 5.1426 bits

Block 17 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
156 blk.17.attn_k.weight Block 17 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
157 blk.17.attn_norm.weight Block 17 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
158 blk.17.attn_output.weight Block 17 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
159 blk.17.attn_q.weight Block 17 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
160 blk.17.attn_v.weight Block 17 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
161 blk.17.ffn_down.weight Block 17 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
162 blk.17.ffn_gate.weight Block 17 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
163 blk.17.ffn_norm.weight Block 17 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
164 blk.17.ffn_up.weight Block 17 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.17: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.17: 4.5005 bits

Block 18 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
165 blk.18.attn_k.weight Block 18 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
166 blk.18.attn_norm.weight Block 18 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
167 blk.18.attn_output.weight Block 18 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
168 blk.18.attn_q.weight Block 18 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
169 blk.18.attn_v.weight Block 18 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
170 blk.18.ffn_down.weight Block 18 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
171 blk.18.ffn_gate.weight Block 18 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
172 blk.18.ffn_norm.weight Block 18 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
173 blk.18.ffn_up.weight Block 18 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.18: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.18: 4.5005 bits

Block 19 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
174 blk.19.attn_k.weight Block 19 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
175 blk.19.attn_norm.weight Block 19 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
176 blk.19.attn_output.weight Block 19 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
177 blk.19.attn_q.weight Block 19 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
178 blk.19.attn_v.weight Block 19 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
179 blk.19.ffn_down.weight Block 19 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
180 blk.19.ffn_gate.weight Block 19 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
181 blk.19.ffn_norm.weight Block 19 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
182 blk.19.ffn_up.weight Block 19 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.19: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.19: 5.1426 bits

Block 20 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
183 blk.20.attn_k.weight Block 20 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
184 blk.20.attn_norm.weight Block 20 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
185 blk.20.attn_output.weight Block 20 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
186 blk.20.attn_q.weight Block 20 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
187 blk.20.attn_v.weight Block 20 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
188 blk.20.ffn_down.weight Block 20 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
189 blk.20.ffn_gate.weight Block 20 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
190 blk.20.ffn_norm.weight Block 20 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
191 blk.20.ffn_up.weight Block 20 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.20: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.20: 4.5005 bits

Block 21 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
192 blk.21.attn_k.weight Block 21 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
193 blk.21.attn_norm.weight Block 21 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
194 blk.21.attn_output.weight Block 21 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
195 blk.21.attn_q.weight Block 21 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
196 blk.21.attn_v.weight Block 21 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
197 blk.21.ffn_down.weight Block 21 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
198 blk.21.ffn_gate.weight Block 21 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
199 blk.21.ffn_norm.weight Block 21 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
200 blk.21.ffn_up.weight Block 21 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.21: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.21: 4.5005 bits

Block 22 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
201 blk.22.attn_k.weight Block 22 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
202 blk.22.attn_norm.weight Block 22 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
203 blk.22.attn_output.weight Block 22 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
204 blk.22.attn_q.weight Block 22 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
205 blk.22.attn_v.weight Block 22 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
206 blk.22.ffn_down.weight Block 22 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
207 blk.22.ffn_gate.weight Block 22 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
208 blk.22.ffn_norm.weight Block 22 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
209 blk.22.ffn_up.weight Block 22 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.22: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.22: 5.1426 bits

Block 23 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
210 blk.23.attn_k.weight Block 23 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
211 blk.23.attn_norm.weight Block 23 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
212 blk.23.attn_output.weight Block 23 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
213 blk.23.attn_q.weight Block 23 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
214 blk.23.attn_v.weight Block 23 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
215 blk.23.ffn_down.weight Block 23 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
216 blk.23.ffn_gate.weight Block 23 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
217 blk.23.ffn_norm.weight Block 23 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
218 blk.23.ffn_up.weight Block 23 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.23: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.23: 4.5005 bits

Block 24 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
219 blk.24.attn_k.weight Block 24 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
220 blk.24.attn_norm.weight Block 24 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
221 blk.24.attn_output.weight Block 24 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
222 blk.24.attn_q.weight Block 24 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
223 blk.24.attn_v.weight Block 24 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
224 blk.24.ffn_down.weight Block 24 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
225 blk.24.ffn_gate.weight Block 24 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
226 blk.24.ffn_norm.weight Block 24 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
227 blk.24.ffn_up.weight Block 24 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.24: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.24: 4.5005 bits

Block 25 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
228 blk.25.attn_k.weight Block 25 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
229 blk.25.attn_norm.weight Block 25 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
230 blk.25.attn_output.weight Block 25 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
231 blk.25.attn_q.weight Block 25 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
232 blk.25.attn_v.weight Block 25 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
233 blk.25.ffn_down.weight Block 25 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
234 blk.25.ffn_gate.weight Block 25 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
235 blk.25.ffn_norm.weight Block 25 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
236 blk.25.ffn_up.weight Block 25 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.25: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.25: 5.1426 bits

Block 26 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
237 blk.26.attn_k.weight Block 26 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
238 blk.26.attn_norm.weight Block 26 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
239 blk.26.attn_output.weight Block 26 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
240 blk.26.attn_q.weight Block 26 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
241 blk.26.attn_v.weight Block 26 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
242 blk.26.ffn_down.weight Block 26 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
243 blk.26.ffn_gate.weight Block 26 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
244 blk.26.ffn_norm.weight Block 26 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
245 blk.26.ffn_up.weight Block 26 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.26: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.26: 4.5005 bits

Block 27 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
246 blk.27.attn_k.weight Block 27 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
247 blk.27.attn_norm.weight Block 27 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
248 blk.27.attn_output.weight Block 27 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
249 blk.27.attn_q.weight Block 27 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
250 blk.27.attn_v.weight Block 27 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
251 blk.27.ffn_down.weight Block 27 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
252 blk.27.ffn_gate.weight Block 27 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
253 blk.27.ffn_norm.weight Block 27 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
254 blk.27.ffn_up.weight Block 27 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.27: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.27: 4.5005 bits

Block 28 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
255 blk.28.attn_k.weight Block 28 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
256 blk.28.attn_norm.weight Block 28 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
257 blk.28.attn_output.weight Block 28 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
258 blk.28.attn_q.weight Block 28 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
259 blk.28.attn_v.weight Block 28 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
260 blk.28.ffn_down.weight Block 28 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
261 blk.28.ffn_gate.weight Block 28 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
262 blk.28.ffn_norm.weight Block 28 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
263 blk.28.ffn_up.weight Block 28 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.28: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.28: 5.1426 bits

Block 29 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
264 blk.29.attn_k.weight Block 29 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
265 blk.29.attn_norm.weight Block 29 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
266 blk.29.attn_output.weight Block 29 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
267 blk.29.attn_q.weight Block 29 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
268 blk.29.attn_v.weight Block 29 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
269 blk.29.ffn_down.weight Block 29 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
270 blk.29.ffn_gate.weight Block 29 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
271 blk.29.ffn_norm.weight Block 29 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
272 blk.29.ffn_up.weight Block 29 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.29: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.29: 4.5005 bits

Block 30 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
273 blk.30.attn_k.weight Block 30 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
274 blk.30.attn_norm.weight Block 30 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
275 blk.30.attn_output.weight Block 30 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
276 blk.30.attn_q.weight Block 30 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
277 blk.30.attn_v.weight Block 30 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
278 blk.30.ffn_down.weight Block 30 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
279 blk.30.ffn_gate.weight Block 30 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
280 blk.30.ffn_norm.weight Block 30 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
281 blk.30.ffn_up.weight Block 30 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.30: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.30: 4.5005 bits

Block 31 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
282 blk.31.attn_k.weight Block 31 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
283 blk.31.attn_norm.weight Block 31 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
284 blk.31.attn_output.weight Block 31 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
285 blk.31.attn_q.weight Block 31 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
286 blk.31.attn_v.weight Block 31 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
287 blk.31.ffn_down.weight Block 31 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
288 blk.31.ffn_gate.weight Block 31 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
289 blk.31.ffn_norm.weight Block 31 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
290 blk.31.ffn_up.weight Block 31 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.31: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.31: 5.1426 bits

Block 32 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
291 blk.32.attn_k.weight Block 32 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
292 blk.32.attn_norm.weight Block 32 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
293 blk.32.attn_output.weight Block 32 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
294 blk.32.attn_q.weight Block 32 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
295 blk.32.attn_v.weight Block 32 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
296 blk.32.ffn_down.weight Block 32 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
297 blk.32.ffn_gate.weight Block 32 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
298 blk.32.ffn_norm.weight Block 32 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
299 blk.32.ffn_up.weight Block 32 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.32: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.32: 4.5005 bits

Block 33 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
300 blk.33.attn_k.weight Block 33 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
301 blk.33.attn_norm.weight Block 33 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
302 blk.33.attn_output.weight Block 33 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
303 blk.33.attn_q.weight Block 33 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
304 blk.33.attn_v.weight Block 33 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
305 blk.33.ffn_down.weight Block 33 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q4_K 4.5000
306 blk.33.ffn_gate.weight Block 33 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
307 blk.33.ffn_norm.weight Block 33 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
308 blk.33.ffn_up.weight Block 33 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.33: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.33: 4.5005 bits

Block 34 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
309 blk.34.attn_k.weight Block 34 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
310 blk.34.attn_norm.weight Block 34 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
311 blk.34.attn_output.weight Block 34 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
312 blk.34.attn_q.weight Block 34 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
313 blk.34.attn_v.weight Block 34 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
314 blk.34.ffn_down.weight Block 34 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
315 blk.34.ffn_gate.weight Block 34 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
316 blk.34.ffn_norm.weight Block 34 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
317 blk.34.ffn_up.weight Block 34 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.34: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.34: 5.1426 bits

Block 35 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
318 blk.35.attn_k.weight Block 35 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
319 blk.35.attn_norm.weight Block 35 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
320 blk.35.attn_output.weight Block 35 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
321 blk.35.attn_q.weight Block 35 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
322 blk.35.attn_v.weight Block 35 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
323 blk.35.ffn_down.weight Block 35 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
324 blk.35.ffn_gate.weight Block 35 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
325 blk.35.ffn_norm.weight Block 35 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
326 blk.35.ffn_up.weight Block 35 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.35: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.35: 5.1426 bits

Block 36 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
327 blk.36.attn_k.weight Block 36 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
328 blk.36.attn_norm.weight Block 36 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
329 blk.36.attn_output.weight Block 36 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
330 blk.36.attn_q.weight Block 36 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
331 blk.36.attn_v.weight Block 36 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
332 blk.36.ffn_down.weight Block 36 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
333 blk.36.ffn_gate.weight Block 36 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
334 blk.36.ffn_norm.weight Block 36 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
335 blk.36.ffn_up.weight Block 36 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.36: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.36: 5.1426 bits

Block 37 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
336 blk.37.attn_k.weight Block 37 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
337 blk.37.attn_norm.weight Block 37 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
338 blk.37.attn_output.weight Block 37 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
339 blk.37.attn_q.weight Block 37 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
340 blk.37.attn_v.weight Block 37 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
341 blk.37.ffn_down.weight Block 37 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
342 blk.37.ffn_gate.weight Block 37 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
343 blk.37.ffn_norm.weight Block 37 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
344 blk.37.ffn_up.weight Block 37 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.37: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.37: 5.1426 bits

Block 38 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
345 blk.38.attn_k.weight Block 38 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
346 blk.38.attn_norm.weight Block 38 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
347 blk.38.attn_output.weight Block 38 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
348 blk.38.attn_q.weight Block 38 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
349 blk.38.attn_v.weight Block 38 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
350 blk.38.ffn_down.weight Block 38 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
351 blk.38.ffn_gate.weight Block 38 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
352 blk.38.ffn_norm.weight Block 38 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
353 blk.38.ffn_up.weight Block 38 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.38: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.38: 5.1426 bits

Block 39 Tensor Group : ~556M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type BPW
354 blk.39.attn_k.weight Block 39 Attention Key (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q4_K 4.5000
355 blk.39.attn_norm.weight Block 39 Attention Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
356 blk.39.attn_output.weight Block 39 Attention Output (W) ( ~21M) 20971520 4096 x 5120 x 1 x 1 Q4_K 4.5000
357 blk.39.attn_q.weight Block 39 Attention Query (W) ( ~21M) 20971520 5120 x 4096 x 1 x 1 Q4_K 4.5000
358 blk.39.attn_v.weight Block 39 Attention Value (W) ( ~5M) 5242880 5120 x 1024 x 1 x 1 Q6_K 6.5625
359 blk.39.ffn_down.weight Block 39 Feed-Forward Network "Down" (W) (~168M) 167772160 32768 x 5120 x 1 x 1 Q6_K 6.5625
360 blk.39.ffn_gate.weight Block 39 Feed-Forward Network "Gate" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
361 blk.39.ffn_norm.weight Block 39 Feed-Forward Network Normalization (W) ( ~5K) 5120 5120 x 1 x 1 x 1 F32 32.0000
362 blk.39.ffn_up.weight Block 39 Feed-Forward Network "Up" (W) (~168M) 167772160 5120 x 32768 x 1 x 1 Q4_K 4.5000
  • Total elements in blk.39: (~556M) 555755520
  • Percentage of total elements: 2.36%
  • Bits per Weight (BPW) for blk.39: 5.1426 bits

Total BPW for Mistral-Small-3.2-24B-Instruct-2506-Q4_K_M-unsloth.gguf: 4.8620 bits