Update config.json
Browse files- config.json +224 -224
config.json
CHANGED
|
@@ -42,7 +42,7 @@
|
|
| 42 |
"model.layers.0.mlp.down_proj": {
|
| 43 |
"bias": false,
|
| 44 |
"enable_norm": true,
|
| 45 |
-
"enable_perm":
|
| 46 |
"group_num": 1,
|
| 47 |
"group_size": 14336,
|
| 48 |
"in_features": 14336,
|
|
@@ -66,7 +66,7 @@
|
|
| 66 |
"model.layers.0.mlp.gate_proj": {
|
| 67 |
"bias": false,
|
| 68 |
"enable_norm": true,
|
| 69 |
-
"enable_perm":
|
| 70 |
"group_num": 1,
|
| 71 |
"group_size": 4096,
|
| 72 |
"in_features": 4096,
|
|
@@ -90,7 +90,7 @@
|
|
| 90 |
"model.layers.0.mlp.up_proj": {
|
| 91 |
"bias": false,
|
| 92 |
"enable_norm": true,
|
| 93 |
-
"enable_perm":
|
| 94 |
"group_num": 1,
|
| 95 |
"group_size": 4096,
|
| 96 |
"in_features": 4096,
|
|
@@ -114,7 +114,7 @@
|
|
| 114 |
"model.layers.0.self_attn.k_proj": {
|
| 115 |
"bias": false,
|
| 116 |
"enable_norm": true,
|
| 117 |
-
"enable_perm":
|
| 118 |
"group_num": 1,
|
| 119 |
"group_size": 4096,
|
| 120 |
"in_features": 4096,
|
|
@@ -138,7 +138,7 @@
|
|
| 138 |
"model.layers.0.self_attn.o_proj": {
|
| 139 |
"bias": false,
|
| 140 |
"enable_norm": true,
|
| 141 |
-
"enable_perm":
|
| 142 |
"group_num": 1,
|
| 143 |
"group_size": 4096,
|
| 144 |
"in_features": 4096,
|
|
@@ -162,7 +162,7 @@
|
|
| 162 |
"model.layers.0.self_attn.q_proj": {
|
| 163 |
"bias": false,
|
| 164 |
"enable_norm": true,
|
| 165 |
-
"enable_perm":
|
| 166 |
"group_num": 1,
|
| 167 |
"group_size": 4096,
|
| 168 |
"in_features": 4096,
|
|
@@ -186,7 +186,7 @@
|
|
| 186 |
"model.layers.0.self_attn.v_proj": {
|
| 187 |
"bias": false,
|
| 188 |
"enable_norm": true,
|
| 189 |
-
"enable_perm":
|
| 190 |
"group_num": 1,
|
| 191 |
"group_size": 4096,
|
| 192 |
"in_features": 4096,
|
|
@@ -210,7 +210,7 @@
|
|
| 210 |
"model.layers.1.mlp.down_proj": {
|
| 211 |
"bias": false,
|
| 212 |
"enable_norm": true,
|
| 213 |
-
"enable_perm":
|
| 214 |
"group_num": 1,
|
| 215 |
"group_size": 14336,
|
| 216 |
"in_features": 14336,
|
|
@@ -234,7 +234,7 @@
|
|
| 234 |
"model.layers.1.mlp.gate_proj": {
|
| 235 |
"bias": false,
|
| 236 |
"enable_norm": true,
|
| 237 |
-
"enable_perm":
|
| 238 |
"group_num": 1,
|
| 239 |
"group_size": 4096,
|
| 240 |
"in_features": 4096,
|
|
@@ -258,7 +258,7 @@
|
|
| 258 |
"model.layers.1.mlp.up_proj": {
|
| 259 |
"bias": false,
|
| 260 |
"enable_norm": true,
|
| 261 |
-
"enable_perm":
|
| 262 |
"group_num": 1,
|
| 263 |
"group_size": 4096,
|
| 264 |
"in_features": 4096,
|
|
@@ -282,7 +282,7 @@
|
|
| 282 |
"model.layers.1.self_attn.k_proj": {
|
| 283 |
"bias": false,
|
| 284 |
"enable_norm": true,
|
| 285 |
-
"enable_perm":
|
| 286 |
"group_num": 1,
|
| 287 |
"group_size": 4096,
|
| 288 |
"in_features": 4096,
|
|
@@ -306,7 +306,7 @@
|
|
| 306 |
"model.layers.1.self_attn.o_proj": {
|
| 307 |
"bias": false,
|
| 308 |
"enable_norm": true,
|
| 309 |
-
"enable_perm":
|
| 310 |
"group_num": 1,
|
| 311 |
"group_size": 4096,
|
| 312 |
"in_features": 4096,
|
|
@@ -330,7 +330,7 @@
|
|
| 330 |
"model.layers.1.self_attn.q_proj": {
|
| 331 |
"bias": false,
|
| 332 |
"enable_norm": true,
|
| 333 |
-
"enable_perm":
|
| 334 |
"group_num": 1,
|
| 335 |
"group_size": 4096,
|
| 336 |
"in_features": 4096,
|
|
@@ -354,7 +354,7 @@
|
|
| 354 |
"model.layers.1.self_attn.v_proj": {
|
| 355 |
"bias": false,
|
| 356 |
"enable_norm": true,
|
| 357 |
-
"enable_perm":
|
| 358 |
"group_num": 1,
|
| 359 |
"group_size": 4096,
|
| 360 |
"in_features": 4096,
|
|
@@ -378,7 +378,7 @@
|
|
| 378 |
"model.layers.10.mlp.down_proj": {
|
| 379 |
"bias": false,
|
| 380 |
"enable_norm": true,
|
| 381 |
-
"enable_perm":
|
| 382 |
"group_num": 1,
|
| 383 |
"group_size": 14336,
|
| 384 |
"in_features": 14336,
|
|
@@ -402,7 +402,7 @@
|
|
| 402 |
"model.layers.10.mlp.gate_proj": {
|
| 403 |
"bias": false,
|
| 404 |
"enable_norm": true,
|
| 405 |
-
"enable_perm":
|
| 406 |
"group_num": 1,
|
| 407 |
"group_size": 4096,
|
| 408 |
"in_features": 4096,
|
|
@@ -426,7 +426,7 @@
|
|
| 426 |
"model.layers.10.mlp.up_proj": {
|
| 427 |
"bias": false,
|
| 428 |
"enable_norm": true,
|
| 429 |
-
"enable_perm":
|
| 430 |
"group_num": 1,
|
| 431 |
"group_size": 4096,
|
| 432 |
"in_features": 4096,
|
|
@@ -450,7 +450,7 @@
|
|
| 450 |
"model.layers.10.self_attn.k_proj": {
|
| 451 |
"bias": false,
|
| 452 |
"enable_norm": true,
|
| 453 |
-
"enable_perm":
|
| 454 |
"group_num": 1,
|
| 455 |
"group_size": 4096,
|
| 456 |
"in_features": 4096,
|
|
@@ -474,7 +474,7 @@
|
|
| 474 |
"model.layers.10.self_attn.o_proj": {
|
| 475 |
"bias": false,
|
| 476 |
"enable_norm": true,
|
| 477 |
-
"enable_perm":
|
| 478 |
"group_num": 1,
|
| 479 |
"group_size": 4096,
|
| 480 |
"in_features": 4096,
|
|
@@ -498,7 +498,7 @@
|
|
| 498 |
"model.layers.10.self_attn.q_proj": {
|
| 499 |
"bias": false,
|
| 500 |
"enable_norm": true,
|
| 501 |
-
"enable_perm":
|
| 502 |
"group_num": 1,
|
| 503 |
"group_size": 4096,
|
| 504 |
"in_features": 4096,
|
|
@@ -522,7 +522,7 @@
|
|
| 522 |
"model.layers.10.self_attn.v_proj": {
|
| 523 |
"bias": false,
|
| 524 |
"enable_norm": true,
|
| 525 |
-
"enable_perm":
|
| 526 |
"group_num": 1,
|
| 527 |
"group_size": 4096,
|
| 528 |
"in_features": 4096,
|
|
@@ -546,7 +546,7 @@
|
|
| 546 |
"model.layers.11.mlp.down_proj": {
|
| 547 |
"bias": false,
|
| 548 |
"enable_norm": true,
|
| 549 |
-
"enable_perm":
|
| 550 |
"group_num": 1,
|
| 551 |
"group_size": 14336,
|
| 552 |
"in_features": 14336,
|
|
@@ -570,7 +570,7 @@
|
|
| 570 |
"model.layers.11.mlp.gate_proj": {
|
| 571 |
"bias": false,
|
| 572 |
"enable_norm": true,
|
| 573 |
-
"enable_perm":
|
| 574 |
"group_num": 1,
|
| 575 |
"group_size": 4096,
|
| 576 |
"in_features": 4096,
|
|
@@ -594,7 +594,7 @@
|
|
| 594 |
"model.layers.11.mlp.up_proj": {
|
| 595 |
"bias": false,
|
| 596 |
"enable_norm": true,
|
| 597 |
-
"enable_perm":
|
| 598 |
"group_num": 1,
|
| 599 |
"group_size": 4096,
|
| 600 |
"in_features": 4096,
|
|
@@ -618,7 +618,7 @@
|
|
| 618 |
"model.layers.11.self_attn.k_proj": {
|
| 619 |
"bias": false,
|
| 620 |
"enable_norm": true,
|
| 621 |
-
"enable_perm":
|
| 622 |
"group_num": 1,
|
| 623 |
"group_size": 4096,
|
| 624 |
"in_features": 4096,
|
|
@@ -642,7 +642,7 @@
|
|
| 642 |
"model.layers.11.self_attn.o_proj": {
|
| 643 |
"bias": false,
|
| 644 |
"enable_norm": true,
|
| 645 |
-
"enable_perm":
|
| 646 |
"group_num": 1,
|
| 647 |
"group_size": 4096,
|
| 648 |
"in_features": 4096,
|
|
@@ -666,7 +666,7 @@
|
|
| 666 |
"model.layers.11.self_attn.q_proj": {
|
| 667 |
"bias": false,
|
| 668 |
"enable_norm": true,
|
| 669 |
-
"enable_perm":
|
| 670 |
"group_num": 1,
|
| 671 |
"group_size": 4096,
|
| 672 |
"in_features": 4096,
|
|
@@ -690,7 +690,7 @@
|
|
| 690 |
"model.layers.11.self_attn.v_proj": {
|
| 691 |
"bias": false,
|
| 692 |
"enable_norm": true,
|
| 693 |
-
"enable_perm":
|
| 694 |
"group_num": 1,
|
| 695 |
"group_size": 4096,
|
| 696 |
"in_features": 4096,
|
|
@@ -714,7 +714,7 @@
|
|
| 714 |
"model.layers.12.mlp.down_proj": {
|
| 715 |
"bias": false,
|
| 716 |
"enable_norm": true,
|
| 717 |
-
"enable_perm":
|
| 718 |
"group_num": 1,
|
| 719 |
"group_size": 14336,
|
| 720 |
"in_features": 14336,
|
|
@@ -738,7 +738,7 @@
|
|
| 738 |
"model.layers.12.mlp.gate_proj": {
|
| 739 |
"bias": false,
|
| 740 |
"enable_norm": true,
|
| 741 |
-
"enable_perm":
|
| 742 |
"group_num": 1,
|
| 743 |
"group_size": 4096,
|
| 744 |
"in_features": 4096,
|
|
@@ -762,7 +762,7 @@
|
|
| 762 |
"model.layers.12.mlp.up_proj": {
|
| 763 |
"bias": false,
|
| 764 |
"enable_norm": true,
|
| 765 |
-
"enable_perm":
|
| 766 |
"group_num": 1,
|
| 767 |
"group_size": 4096,
|
| 768 |
"in_features": 4096,
|
|
@@ -786,7 +786,7 @@
|
|
| 786 |
"model.layers.12.self_attn.k_proj": {
|
| 787 |
"bias": false,
|
| 788 |
"enable_norm": true,
|
| 789 |
-
"enable_perm":
|
| 790 |
"group_num": 1,
|
| 791 |
"group_size": 4096,
|
| 792 |
"in_features": 4096,
|
|
@@ -810,7 +810,7 @@
|
|
| 810 |
"model.layers.12.self_attn.o_proj": {
|
| 811 |
"bias": false,
|
| 812 |
"enable_norm": true,
|
| 813 |
-
"enable_perm":
|
| 814 |
"group_num": 1,
|
| 815 |
"group_size": 4096,
|
| 816 |
"in_features": 4096,
|
|
@@ -834,7 +834,7 @@
|
|
| 834 |
"model.layers.12.self_attn.q_proj": {
|
| 835 |
"bias": false,
|
| 836 |
"enable_norm": true,
|
| 837 |
-
"enable_perm":
|
| 838 |
"group_num": 1,
|
| 839 |
"group_size": 4096,
|
| 840 |
"in_features": 4096,
|
|
@@ -858,7 +858,7 @@
|
|
| 858 |
"model.layers.12.self_attn.v_proj": {
|
| 859 |
"bias": false,
|
| 860 |
"enable_norm": true,
|
| 861 |
-
"enable_perm":
|
| 862 |
"group_num": 1,
|
| 863 |
"group_size": 4096,
|
| 864 |
"in_features": 4096,
|
|
@@ -882,7 +882,7 @@
|
|
| 882 |
"model.layers.13.mlp.down_proj": {
|
| 883 |
"bias": false,
|
| 884 |
"enable_norm": true,
|
| 885 |
-
"enable_perm":
|
| 886 |
"group_num": 1,
|
| 887 |
"group_size": 14336,
|
| 888 |
"in_features": 14336,
|
|
@@ -906,7 +906,7 @@
|
|
| 906 |
"model.layers.13.mlp.gate_proj": {
|
| 907 |
"bias": false,
|
| 908 |
"enable_norm": true,
|
| 909 |
-
"enable_perm":
|
| 910 |
"group_num": 1,
|
| 911 |
"group_size": 4096,
|
| 912 |
"in_features": 4096,
|
|
@@ -930,7 +930,7 @@
|
|
| 930 |
"model.layers.13.mlp.up_proj": {
|
| 931 |
"bias": false,
|
| 932 |
"enable_norm": true,
|
| 933 |
-
"enable_perm":
|
| 934 |
"group_num": 1,
|
| 935 |
"group_size": 4096,
|
| 936 |
"in_features": 4096,
|
|
@@ -954,7 +954,7 @@
|
|
| 954 |
"model.layers.13.self_attn.k_proj": {
|
| 955 |
"bias": false,
|
| 956 |
"enable_norm": true,
|
| 957 |
-
"enable_perm":
|
| 958 |
"group_num": 1,
|
| 959 |
"group_size": 4096,
|
| 960 |
"in_features": 4096,
|
|
@@ -978,7 +978,7 @@
|
|
| 978 |
"model.layers.13.self_attn.o_proj": {
|
| 979 |
"bias": false,
|
| 980 |
"enable_norm": true,
|
| 981 |
-
"enable_perm":
|
| 982 |
"group_num": 1,
|
| 983 |
"group_size": 4096,
|
| 984 |
"in_features": 4096,
|
|
@@ -1002,7 +1002,7 @@
|
|
| 1002 |
"model.layers.13.self_attn.q_proj": {
|
| 1003 |
"bias": false,
|
| 1004 |
"enable_norm": true,
|
| 1005 |
-
"enable_perm":
|
| 1006 |
"group_num": 1,
|
| 1007 |
"group_size": 4096,
|
| 1008 |
"in_features": 4096,
|
|
@@ -1026,7 +1026,7 @@
|
|
| 1026 |
"model.layers.13.self_attn.v_proj": {
|
| 1027 |
"bias": false,
|
| 1028 |
"enable_norm": true,
|
| 1029 |
-
"enable_perm":
|
| 1030 |
"group_num": 1,
|
| 1031 |
"group_size": 4096,
|
| 1032 |
"in_features": 4096,
|
|
@@ -1050,7 +1050,7 @@
|
|
| 1050 |
"model.layers.14.mlp.down_proj": {
|
| 1051 |
"bias": false,
|
| 1052 |
"enable_norm": true,
|
| 1053 |
-
"enable_perm":
|
| 1054 |
"group_num": 1,
|
| 1055 |
"group_size": 14336,
|
| 1056 |
"in_features": 14336,
|
|
@@ -1074,7 +1074,7 @@
|
|
| 1074 |
"model.layers.14.mlp.gate_proj": {
|
| 1075 |
"bias": false,
|
| 1076 |
"enable_norm": true,
|
| 1077 |
-
"enable_perm":
|
| 1078 |
"group_num": 1,
|
| 1079 |
"group_size": 4096,
|
| 1080 |
"in_features": 4096,
|
|
@@ -1098,7 +1098,7 @@
|
|
| 1098 |
"model.layers.14.mlp.up_proj": {
|
| 1099 |
"bias": false,
|
| 1100 |
"enable_norm": true,
|
| 1101 |
-
"enable_perm":
|
| 1102 |
"group_num": 1,
|
| 1103 |
"group_size": 4096,
|
| 1104 |
"in_features": 4096,
|
|
@@ -1122,7 +1122,7 @@
|
|
| 1122 |
"model.layers.14.self_attn.k_proj": {
|
| 1123 |
"bias": false,
|
| 1124 |
"enable_norm": true,
|
| 1125 |
-
"enable_perm":
|
| 1126 |
"group_num": 1,
|
| 1127 |
"group_size": 4096,
|
| 1128 |
"in_features": 4096,
|
|
@@ -1146,7 +1146,7 @@
|
|
| 1146 |
"model.layers.14.self_attn.o_proj": {
|
| 1147 |
"bias": false,
|
| 1148 |
"enable_norm": true,
|
| 1149 |
-
"enable_perm":
|
| 1150 |
"group_num": 1,
|
| 1151 |
"group_size": 4096,
|
| 1152 |
"in_features": 4096,
|
|
@@ -1170,7 +1170,7 @@
|
|
| 1170 |
"model.layers.14.self_attn.q_proj": {
|
| 1171 |
"bias": false,
|
| 1172 |
"enable_norm": true,
|
| 1173 |
-
"enable_perm":
|
| 1174 |
"group_num": 1,
|
| 1175 |
"group_size": 4096,
|
| 1176 |
"in_features": 4096,
|
|
@@ -1194,7 +1194,7 @@
|
|
| 1194 |
"model.layers.14.self_attn.v_proj": {
|
| 1195 |
"bias": false,
|
| 1196 |
"enable_norm": true,
|
| 1197 |
-
"enable_perm":
|
| 1198 |
"group_num": 1,
|
| 1199 |
"group_size": 4096,
|
| 1200 |
"in_features": 4096,
|
|
@@ -1218,7 +1218,7 @@
|
|
| 1218 |
"model.layers.15.mlp.down_proj": {
|
| 1219 |
"bias": false,
|
| 1220 |
"enable_norm": true,
|
| 1221 |
-
"enable_perm":
|
| 1222 |
"group_num": 1,
|
| 1223 |
"group_size": 14336,
|
| 1224 |
"in_features": 14336,
|
|
@@ -1242,7 +1242,7 @@
|
|
| 1242 |
"model.layers.15.mlp.gate_proj": {
|
| 1243 |
"bias": false,
|
| 1244 |
"enable_norm": true,
|
| 1245 |
-
"enable_perm":
|
| 1246 |
"group_num": 1,
|
| 1247 |
"group_size": 4096,
|
| 1248 |
"in_features": 4096,
|
|
@@ -1266,7 +1266,7 @@
|
|
| 1266 |
"model.layers.15.mlp.up_proj": {
|
| 1267 |
"bias": false,
|
| 1268 |
"enable_norm": true,
|
| 1269 |
-
"enable_perm":
|
| 1270 |
"group_num": 1,
|
| 1271 |
"group_size": 4096,
|
| 1272 |
"in_features": 4096,
|
|
@@ -1290,7 +1290,7 @@
|
|
| 1290 |
"model.layers.15.self_attn.k_proj": {
|
| 1291 |
"bias": false,
|
| 1292 |
"enable_norm": true,
|
| 1293 |
-
"enable_perm":
|
| 1294 |
"group_num": 1,
|
| 1295 |
"group_size": 4096,
|
| 1296 |
"in_features": 4096,
|
|
@@ -1314,7 +1314,7 @@
|
|
| 1314 |
"model.layers.15.self_attn.o_proj": {
|
| 1315 |
"bias": false,
|
| 1316 |
"enable_norm": true,
|
| 1317 |
-
"enable_perm":
|
| 1318 |
"group_num": 1,
|
| 1319 |
"group_size": 4096,
|
| 1320 |
"in_features": 4096,
|
|
@@ -1338,7 +1338,7 @@
|
|
| 1338 |
"model.layers.15.self_attn.q_proj": {
|
| 1339 |
"bias": false,
|
| 1340 |
"enable_norm": true,
|
| 1341 |
-
"enable_perm":
|
| 1342 |
"group_num": 1,
|
| 1343 |
"group_size": 4096,
|
| 1344 |
"in_features": 4096,
|
|
@@ -1362,7 +1362,7 @@
|
|
| 1362 |
"model.layers.15.self_attn.v_proj": {
|
| 1363 |
"bias": false,
|
| 1364 |
"enable_norm": true,
|
| 1365 |
-
"enable_perm":
|
| 1366 |
"group_num": 1,
|
| 1367 |
"group_size": 4096,
|
| 1368 |
"in_features": 4096,
|
|
@@ -1386,7 +1386,7 @@
|
|
| 1386 |
"model.layers.16.mlp.down_proj": {
|
| 1387 |
"bias": false,
|
| 1388 |
"enable_norm": true,
|
| 1389 |
-
"enable_perm":
|
| 1390 |
"group_num": 1,
|
| 1391 |
"group_size": 14336,
|
| 1392 |
"in_features": 14336,
|
|
@@ -1410,7 +1410,7 @@
|
|
| 1410 |
"model.layers.16.mlp.gate_proj": {
|
| 1411 |
"bias": false,
|
| 1412 |
"enable_norm": true,
|
| 1413 |
-
"enable_perm":
|
| 1414 |
"group_num": 1,
|
| 1415 |
"group_size": 4096,
|
| 1416 |
"in_features": 4096,
|
|
@@ -1434,7 +1434,7 @@
|
|
| 1434 |
"model.layers.16.mlp.up_proj": {
|
| 1435 |
"bias": false,
|
| 1436 |
"enable_norm": true,
|
| 1437 |
-
"enable_perm":
|
| 1438 |
"group_num": 1,
|
| 1439 |
"group_size": 4096,
|
| 1440 |
"in_features": 4096,
|
|
@@ -1458,7 +1458,7 @@
|
|
| 1458 |
"model.layers.16.self_attn.k_proj": {
|
| 1459 |
"bias": false,
|
| 1460 |
"enable_norm": true,
|
| 1461 |
-
"enable_perm":
|
| 1462 |
"group_num": 1,
|
| 1463 |
"group_size": 4096,
|
| 1464 |
"in_features": 4096,
|
|
@@ -1482,7 +1482,7 @@
|
|
| 1482 |
"model.layers.16.self_attn.o_proj": {
|
| 1483 |
"bias": false,
|
| 1484 |
"enable_norm": true,
|
| 1485 |
-
"enable_perm":
|
| 1486 |
"group_num": 1,
|
| 1487 |
"group_size": 4096,
|
| 1488 |
"in_features": 4096,
|
|
@@ -1506,7 +1506,7 @@
|
|
| 1506 |
"model.layers.16.self_attn.q_proj": {
|
| 1507 |
"bias": false,
|
| 1508 |
"enable_norm": true,
|
| 1509 |
-
"enable_perm":
|
| 1510 |
"group_num": 1,
|
| 1511 |
"group_size": 4096,
|
| 1512 |
"in_features": 4096,
|
|
@@ -1530,7 +1530,7 @@
|
|
| 1530 |
"model.layers.16.self_attn.v_proj": {
|
| 1531 |
"bias": false,
|
| 1532 |
"enable_norm": true,
|
| 1533 |
-
"enable_perm":
|
| 1534 |
"group_num": 1,
|
| 1535 |
"group_size": 4096,
|
| 1536 |
"in_features": 4096,
|
|
@@ -1554,7 +1554,7 @@
|
|
| 1554 |
"model.layers.17.mlp.down_proj": {
|
| 1555 |
"bias": false,
|
| 1556 |
"enable_norm": true,
|
| 1557 |
-
"enable_perm":
|
| 1558 |
"group_num": 1,
|
| 1559 |
"group_size": 14336,
|
| 1560 |
"in_features": 14336,
|
|
@@ -1578,7 +1578,7 @@
|
|
| 1578 |
"model.layers.17.mlp.gate_proj": {
|
| 1579 |
"bias": false,
|
| 1580 |
"enable_norm": true,
|
| 1581 |
-
"enable_perm":
|
| 1582 |
"group_num": 1,
|
| 1583 |
"group_size": 4096,
|
| 1584 |
"in_features": 4096,
|
|
@@ -1602,7 +1602,7 @@
|
|
| 1602 |
"model.layers.17.mlp.up_proj": {
|
| 1603 |
"bias": false,
|
| 1604 |
"enable_norm": true,
|
| 1605 |
-
"enable_perm":
|
| 1606 |
"group_num": 1,
|
| 1607 |
"group_size": 4096,
|
| 1608 |
"in_features": 4096,
|
|
@@ -1626,7 +1626,7 @@
|
|
| 1626 |
"model.layers.17.self_attn.k_proj": {
|
| 1627 |
"bias": false,
|
| 1628 |
"enable_norm": true,
|
| 1629 |
-
"enable_perm":
|
| 1630 |
"group_num": 1,
|
| 1631 |
"group_size": 4096,
|
| 1632 |
"in_features": 4096,
|
|
@@ -1650,7 +1650,7 @@
|
|
| 1650 |
"model.layers.17.self_attn.o_proj": {
|
| 1651 |
"bias": false,
|
| 1652 |
"enable_norm": true,
|
| 1653 |
-
"enable_perm":
|
| 1654 |
"group_num": 1,
|
| 1655 |
"group_size": 4096,
|
| 1656 |
"in_features": 4096,
|
|
@@ -1674,7 +1674,7 @@
|
|
| 1674 |
"model.layers.17.self_attn.q_proj": {
|
| 1675 |
"bias": false,
|
| 1676 |
"enable_norm": true,
|
| 1677 |
-
"enable_perm":
|
| 1678 |
"group_num": 1,
|
| 1679 |
"group_size": 4096,
|
| 1680 |
"in_features": 4096,
|
|
@@ -1698,7 +1698,7 @@
|
|
| 1698 |
"model.layers.17.self_attn.v_proj": {
|
| 1699 |
"bias": false,
|
| 1700 |
"enable_norm": true,
|
| 1701 |
-
"enable_perm":
|
| 1702 |
"group_num": 1,
|
| 1703 |
"group_size": 4096,
|
| 1704 |
"in_features": 4096,
|
|
@@ -1722,7 +1722,7 @@
|
|
| 1722 |
"model.layers.18.mlp.down_proj": {
|
| 1723 |
"bias": false,
|
| 1724 |
"enable_norm": true,
|
| 1725 |
-
"enable_perm":
|
| 1726 |
"group_num": 1,
|
| 1727 |
"group_size": 14336,
|
| 1728 |
"in_features": 14336,
|
|
@@ -1746,7 +1746,7 @@
|
|
| 1746 |
"model.layers.18.mlp.gate_proj": {
|
| 1747 |
"bias": false,
|
| 1748 |
"enable_norm": true,
|
| 1749 |
-
"enable_perm":
|
| 1750 |
"group_num": 1,
|
| 1751 |
"group_size": 4096,
|
| 1752 |
"in_features": 4096,
|
|
@@ -1770,7 +1770,7 @@
|
|
| 1770 |
"model.layers.18.mlp.up_proj": {
|
| 1771 |
"bias": false,
|
| 1772 |
"enable_norm": true,
|
| 1773 |
-
"enable_perm":
|
| 1774 |
"group_num": 1,
|
| 1775 |
"group_size": 4096,
|
| 1776 |
"in_features": 4096,
|
|
@@ -1794,7 +1794,7 @@
|
|
| 1794 |
"model.layers.18.self_attn.k_proj": {
|
| 1795 |
"bias": false,
|
| 1796 |
"enable_norm": true,
|
| 1797 |
-
"enable_perm":
|
| 1798 |
"group_num": 1,
|
| 1799 |
"group_size": 4096,
|
| 1800 |
"in_features": 4096,
|
|
@@ -1818,7 +1818,7 @@
|
|
| 1818 |
"model.layers.18.self_attn.o_proj": {
|
| 1819 |
"bias": false,
|
| 1820 |
"enable_norm": true,
|
| 1821 |
-
"enable_perm":
|
| 1822 |
"group_num": 1,
|
| 1823 |
"group_size": 4096,
|
| 1824 |
"in_features": 4096,
|
|
@@ -1842,7 +1842,7 @@
|
|
| 1842 |
"model.layers.18.self_attn.q_proj": {
|
| 1843 |
"bias": false,
|
| 1844 |
"enable_norm": true,
|
| 1845 |
-
"enable_perm":
|
| 1846 |
"group_num": 1,
|
| 1847 |
"group_size": 4096,
|
| 1848 |
"in_features": 4096,
|
|
@@ -1866,7 +1866,7 @@
|
|
| 1866 |
"model.layers.18.self_attn.v_proj": {
|
| 1867 |
"bias": false,
|
| 1868 |
"enable_norm": true,
|
| 1869 |
-
"enable_perm":
|
| 1870 |
"group_num": 1,
|
| 1871 |
"group_size": 4096,
|
| 1872 |
"in_features": 4096,
|
|
@@ -1890,7 +1890,7 @@
|
|
| 1890 |
"model.layers.19.mlp.down_proj": {
|
| 1891 |
"bias": false,
|
| 1892 |
"enable_norm": true,
|
| 1893 |
-
"enable_perm":
|
| 1894 |
"group_num": 1,
|
| 1895 |
"group_size": 14336,
|
| 1896 |
"in_features": 14336,
|
|
@@ -1914,7 +1914,7 @@
|
|
| 1914 |
"model.layers.19.mlp.gate_proj": {
|
| 1915 |
"bias": false,
|
| 1916 |
"enable_norm": true,
|
| 1917 |
-
"enable_perm":
|
| 1918 |
"group_num": 1,
|
| 1919 |
"group_size": 4096,
|
| 1920 |
"in_features": 4096,
|
|
@@ -1938,7 +1938,7 @@
|
|
| 1938 |
"model.layers.19.mlp.up_proj": {
|
| 1939 |
"bias": false,
|
| 1940 |
"enable_norm": true,
|
| 1941 |
-
"enable_perm":
|
| 1942 |
"group_num": 1,
|
| 1943 |
"group_size": 4096,
|
| 1944 |
"in_features": 4096,
|
|
@@ -1962,7 +1962,7 @@
|
|
| 1962 |
"model.layers.19.self_attn.k_proj": {
|
| 1963 |
"bias": false,
|
| 1964 |
"enable_norm": true,
|
| 1965 |
-
"enable_perm":
|
| 1966 |
"group_num": 1,
|
| 1967 |
"group_size": 4096,
|
| 1968 |
"in_features": 4096,
|
|
@@ -1986,7 +1986,7 @@
|
|
| 1986 |
"model.layers.19.self_attn.o_proj": {
|
| 1987 |
"bias": false,
|
| 1988 |
"enable_norm": true,
|
| 1989 |
-
"enable_perm":
|
| 1990 |
"group_num": 1,
|
| 1991 |
"group_size": 4096,
|
| 1992 |
"in_features": 4096,
|
|
@@ -2010,7 +2010,7 @@
|
|
| 2010 |
"model.layers.19.self_attn.q_proj": {
|
| 2011 |
"bias": false,
|
| 2012 |
"enable_norm": true,
|
| 2013 |
-
"enable_perm":
|
| 2014 |
"group_num": 1,
|
| 2015 |
"group_size": 4096,
|
| 2016 |
"in_features": 4096,
|
|
@@ -2034,7 +2034,7 @@
|
|
| 2034 |
"model.layers.19.self_attn.v_proj": {
|
| 2035 |
"bias": false,
|
| 2036 |
"enable_norm": true,
|
| 2037 |
-
"enable_perm":
|
| 2038 |
"group_num": 1,
|
| 2039 |
"group_size": 4096,
|
| 2040 |
"in_features": 4096,
|
|
@@ -2058,7 +2058,7 @@
|
|
| 2058 |
"model.layers.2.mlp.down_proj": {
|
| 2059 |
"bias": false,
|
| 2060 |
"enable_norm": true,
|
| 2061 |
-
"enable_perm":
|
| 2062 |
"group_num": 1,
|
| 2063 |
"group_size": 14336,
|
| 2064 |
"in_features": 14336,
|
|
@@ -2082,7 +2082,7 @@
|
|
| 2082 |
"model.layers.2.mlp.gate_proj": {
|
| 2083 |
"bias": false,
|
| 2084 |
"enable_norm": true,
|
| 2085 |
-
"enable_perm":
|
| 2086 |
"group_num": 1,
|
| 2087 |
"group_size": 4096,
|
| 2088 |
"in_features": 4096,
|
|
@@ -2106,7 +2106,7 @@
|
|
| 2106 |
"model.layers.2.mlp.up_proj": {
|
| 2107 |
"bias": false,
|
| 2108 |
"enable_norm": true,
|
| 2109 |
-
"enable_perm":
|
| 2110 |
"group_num": 1,
|
| 2111 |
"group_size": 4096,
|
| 2112 |
"in_features": 4096,
|
|
@@ -2130,7 +2130,7 @@
|
|
| 2130 |
"model.layers.2.self_attn.k_proj": {
|
| 2131 |
"bias": false,
|
| 2132 |
"enable_norm": true,
|
| 2133 |
-
"enable_perm":
|
| 2134 |
"group_num": 1,
|
| 2135 |
"group_size": 4096,
|
| 2136 |
"in_features": 4096,
|
|
@@ -2154,7 +2154,7 @@
|
|
| 2154 |
"model.layers.2.self_attn.o_proj": {
|
| 2155 |
"bias": false,
|
| 2156 |
"enable_norm": true,
|
| 2157 |
-
"enable_perm":
|
| 2158 |
"group_num": 1,
|
| 2159 |
"group_size": 4096,
|
| 2160 |
"in_features": 4096,
|
|
@@ -2178,7 +2178,7 @@
|
|
| 2178 |
"model.layers.2.self_attn.q_proj": {
|
| 2179 |
"bias": false,
|
| 2180 |
"enable_norm": true,
|
| 2181 |
-
"enable_perm":
|
| 2182 |
"group_num": 1,
|
| 2183 |
"group_size": 4096,
|
| 2184 |
"in_features": 4096,
|
|
@@ -2202,7 +2202,7 @@
|
|
| 2202 |
"model.layers.2.self_attn.v_proj": {
|
| 2203 |
"bias": false,
|
| 2204 |
"enable_norm": true,
|
| 2205 |
-
"enable_perm":
|
| 2206 |
"group_num": 1,
|
| 2207 |
"group_size": 4096,
|
| 2208 |
"in_features": 4096,
|
|
@@ -2226,7 +2226,7 @@
|
|
| 2226 |
"model.layers.20.mlp.down_proj": {
|
| 2227 |
"bias": false,
|
| 2228 |
"enable_norm": true,
|
| 2229 |
-
"enable_perm":
|
| 2230 |
"group_num": 1,
|
| 2231 |
"group_size": 14336,
|
| 2232 |
"in_features": 14336,
|
|
@@ -2250,7 +2250,7 @@
|
|
| 2250 |
"model.layers.20.mlp.gate_proj": {
|
| 2251 |
"bias": false,
|
| 2252 |
"enable_norm": true,
|
| 2253 |
-
"enable_perm":
|
| 2254 |
"group_num": 1,
|
| 2255 |
"group_size": 4096,
|
| 2256 |
"in_features": 4096,
|
|
@@ -2274,7 +2274,7 @@
|
|
| 2274 |
"model.layers.20.mlp.up_proj": {
|
| 2275 |
"bias": false,
|
| 2276 |
"enable_norm": true,
|
| 2277 |
-
"enable_perm":
|
| 2278 |
"group_num": 1,
|
| 2279 |
"group_size": 4096,
|
| 2280 |
"in_features": 4096,
|
|
@@ -2298,7 +2298,7 @@
|
|
| 2298 |
"model.layers.20.self_attn.k_proj": {
|
| 2299 |
"bias": false,
|
| 2300 |
"enable_norm": true,
|
| 2301 |
-
"enable_perm":
|
| 2302 |
"group_num": 1,
|
| 2303 |
"group_size": 4096,
|
| 2304 |
"in_features": 4096,
|
|
@@ -2322,7 +2322,7 @@
|
|
| 2322 |
"model.layers.20.self_attn.o_proj": {
|
| 2323 |
"bias": false,
|
| 2324 |
"enable_norm": true,
|
| 2325 |
-
"enable_perm":
|
| 2326 |
"group_num": 1,
|
| 2327 |
"group_size": 4096,
|
| 2328 |
"in_features": 4096,
|
|
@@ -2346,7 +2346,7 @@
|
|
| 2346 |
"model.layers.20.self_attn.q_proj": {
|
| 2347 |
"bias": false,
|
| 2348 |
"enable_norm": true,
|
| 2349 |
-
"enable_perm":
|
| 2350 |
"group_num": 1,
|
| 2351 |
"group_size": 4096,
|
| 2352 |
"in_features": 4096,
|
|
@@ -2370,7 +2370,7 @@
|
|
| 2370 |
"model.layers.20.self_attn.v_proj": {
|
| 2371 |
"bias": false,
|
| 2372 |
"enable_norm": true,
|
| 2373 |
-
"enable_perm":
|
| 2374 |
"group_num": 1,
|
| 2375 |
"group_size": 4096,
|
| 2376 |
"in_features": 4096,
|
|
@@ -2394,7 +2394,7 @@
|
|
| 2394 |
"model.layers.21.mlp.down_proj": {
|
| 2395 |
"bias": false,
|
| 2396 |
"enable_norm": true,
|
| 2397 |
-
"enable_perm":
|
| 2398 |
"group_num": 1,
|
| 2399 |
"group_size": 14336,
|
| 2400 |
"in_features": 14336,
|
|
@@ -2418,7 +2418,7 @@
|
|
| 2418 |
"model.layers.21.mlp.gate_proj": {
|
| 2419 |
"bias": false,
|
| 2420 |
"enable_norm": true,
|
| 2421 |
-
"enable_perm":
|
| 2422 |
"group_num": 1,
|
| 2423 |
"group_size": 4096,
|
| 2424 |
"in_features": 4096,
|
|
@@ -2442,7 +2442,7 @@
|
|
| 2442 |
"model.layers.21.mlp.up_proj": {
|
| 2443 |
"bias": false,
|
| 2444 |
"enable_norm": true,
|
| 2445 |
-
"enable_perm":
|
| 2446 |
"group_num": 1,
|
| 2447 |
"group_size": 4096,
|
| 2448 |
"in_features": 4096,
|
|
@@ -2466,7 +2466,7 @@
|
|
| 2466 |
"model.layers.21.self_attn.k_proj": {
|
| 2467 |
"bias": false,
|
| 2468 |
"enable_norm": true,
|
| 2469 |
-
"enable_perm":
|
| 2470 |
"group_num": 1,
|
| 2471 |
"group_size": 4096,
|
| 2472 |
"in_features": 4096,
|
|
@@ -2490,7 +2490,7 @@
|
|
| 2490 |
"model.layers.21.self_attn.o_proj": {
|
| 2491 |
"bias": false,
|
| 2492 |
"enable_norm": true,
|
| 2493 |
-
"enable_perm":
|
| 2494 |
"group_num": 1,
|
| 2495 |
"group_size": 4096,
|
| 2496 |
"in_features": 4096,
|
|
@@ -2514,7 +2514,7 @@
|
|
| 2514 |
"model.layers.21.self_attn.q_proj": {
|
| 2515 |
"bias": false,
|
| 2516 |
"enable_norm": true,
|
| 2517 |
-
"enable_perm":
|
| 2518 |
"group_num": 1,
|
| 2519 |
"group_size": 4096,
|
| 2520 |
"in_features": 4096,
|
|
@@ -2538,7 +2538,7 @@
|
|
| 2538 |
"model.layers.21.self_attn.v_proj": {
|
| 2539 |
"bias": false,
|
| 2540 |
"enable_norm": true,
|
| 2541 |
-
"enable_perm":
|
| 2542 |
"group_num": 1,
|
| 2543 |
"group_size": 4096,
|
| 2544 |
"in_features": 4096,
|
|
@@ -2562,7 +2562,7 @@
|
|
| 2562 |
"model.layers.22.mlp.down_proj": {
|
| 2563 |
"bias": false,
|
| 2564 |
"enable_norm": true,
|
| 2565 |
-
"enable_perm":
|
| 2566 |
"group_num": 1,
|
| 2567 |
"group_size": 14336,
|
| 2568 |
"in_features": 14336,
|
|
@@ -2586,7 +2586,7 @@
|
|
| 2586 |
"model.layers.22.mlp.gate_proj": {
|
| 2587 |
"bias": false,
|
| 2588 |
"enable_norm": true,
|
| 2589 |
-
"enable_perm":
|
| 2590 |
"group_num": 1,
|
| 2591 |
"group_size": 4096,
|
| 2592 |
"in_features": 4096,
|
|
@@ -2610,7 +2610,7 @@
|
|
| 2610 |
"model.layers.22.mlp.up_proj": {
|
| 2611 |
"bias": false,
|
| 2612 |
"enable_norm": true,
|
| 2613 |
-
"enable_perm":
|
| 2614 |
"group_num": 1,
|
| 2615 |
"group_size": 4096,
|
| 2616 |
"in_features": 4096,
|
|
@@ -2634,7 +2634,7 @@
|
|
| 2634 |
"model.layers.22.self_attn.k_proj": {
|
| 2635 |
"bias": false,
|
| 2636 |
"enable_norm": true,
|
| 2637 |
-
"enable_perm":
|
| 2638 |
"group_num": 1,
|
| 2639 |
"group_size": 4096,
|
| 2640 |
"in_features": 4096,
|
|
@@ -2658,7 +2658,7 @@
|
|
| 2658 |
"model.layers.22.self_attn.o_proj": {
|
| 2659 |
"bias": false,
|
| 2660 |
"enable_norm": true,
|
| 2661 |
-
"enable_perm":
|
| 2662 |
"group_num": 1,
|
| 2663 |
"group_size": 4096,
|
| 2664 |
"in_features": 4096,
|
|
@@ -2682,7 +2682,7 @@
|
|
| 2682 |
"model.layers.22.self_attn.q_proj": {
|
| 2683 |
"bias": false,
|
| 2684 |
"enable_norm": true,
|
| 2685 |
-
"enable_perm":
|
| 2686 |
"group_num": 1,
|
| 2687 |
"group_size": 4096,
|
| 2688 |
"in_features": 4096,
|
|
@@ -2706,7 +2706,7 @@
|
|
| 2706 |
"model.layers.22.self_attn.v_proj": {
|
| 2707 |
"bias": false,
|
| 2708 |
"enable_norm": true,
|
| 2709 |
-
"enable_perm":
|
| 2710 |
"group_num": 1,
|
| 2711 |
"group_size": 4096,
|
| 2712 |
"in_features": 4096,
|
|
@@ -2730,7 +2730,7 @@
|
|
| 2730 |
"model.layers.23.mlp.down_proj": {
|
| 2731 |
"bias": false,
|
| 2732 |
"enable_norm": true,
|
| 2733 |
-
"enable_perm":
|
| 2734 |
"group_num": 1,
|
| 2735 |
"group_size": 14336,
|
| 2736 |
"in_features": 14336,
|
|
@@ -2754,7 +2754,7 @@
|
|
| 2754 |
"model.layers.23.mlp.gate_proj": {
|
| 2755 |
"bias": false,
|
| 2756 |
"enable_norm": true,
|
| 2757 |
-
"enable_perm":
|
| 2758 |
"group_num": 1,
|
| 2759 |
"group_size": 4096,
|
| 2760 |
"in_features": 4096,
|
|
@@ -2778,7 +2778,7 @@
|
|
| 2778 |
"model.layers.23.mlp.up_proj": {
|
| 2779 |
"bias": false,
|
| 2780 |
"enable_norm": true,
|
| 2781 |
-
"enable_perm":
|
| 2782 |
"group_num": 1,
|
| 2783 |
"group_size": 4096,
|
| 2784 |
"in_features": 4096,
|
|
@@ -2802,7 +2802,7 @@
|
|
| 2802 |
"model.layers.23.self_attn.k_proj": {
|
| 2803 |
"bias": false,
|
| 2804 |
"enable_norm": true,
|
| 2805 |
-
"enable_perm":
|
| 2806 |
"group_num": 1,
|
| 2807 |
"group_size": 4096,
|
| 2808 |
"in_features": 4096,
|
|
@@ -2826,7 +2826,7 @@
|
|
| 2826 |
"model.layers.23.self_attn.o_proj": {
|
| 2827 |
"bias": false,
|
| 2828 |
"enable_norm": true,
|
| 2829 |
-
"enable_perm":
|
| 2830 |
"group_num": 1,
|
| 2831 |
"group_size": 4096,
|
| 2832 |
"in_features": 4096,
|
|
@@ -2850,7 +2850,7 @@
|
|
| 2850 |
"model.layers.23.self_attn.q_proj": {
|
| 2851 |
"bias": false,
|
| 2852 |
"enable_norm": true,
|
| 2853 |
-
"enable_perm":
|
| 2854 |
"group_num": 1,
|
| 2855 |
"group_size": 4096,
|
| 2856 |
"in_features": 4096,
|
|
@@ -2874,7 +2874,7 @@
|
|
| 2874 |
"model.layers.23.self_attn.v_proj": {
|
| 2875 |
"bias": false,
|
| 2876 |
"enable_norm": true,
|
| 2877 |
-
"enable_perm":
|
| 2878 |
"group_num": 1,
|
| 2879 |
"group_size": 4096,
|
| 2880 |
"in_features": 4096,
|
|
@@ -2898,7 +2898,7 @@
|
|
| 2898 |
"model.layers.24.mlp.down_proj": {
|
| 2899 |
"bias": false,
|
| 2900 |
"enable_norm": true,
|
| 2901 |
-
"enable_perm":
|
| 2902 |
"group_num": 1,
|
| 2903 |
"group_size": 14336,
|
| 2904 |
"in_features": 14336,
|
|
@@ -2922,7 +2922,7 @@
|
|
| 2922 |
"model.layers.24.mlp.gate_proj": {
|
| 2923 |
"bias": false,
|
| 2924 |
"enable_norm": true,
|
| 2925 |
-
"enable_perm":
|
| 2926 |
"group_num": 1,
|
| 2927 |
"group_size": 4096,
|
| 2928 |
"in_features": 4096,
|
|
@@ -2946,7 +2946,7 @@
|
|
| 2946 |
"model.layers.24.mlp.up_proj": {
|
| 2947 |
"bias": false,
|
| 2948 |
"enable_norm": true,
|
| 2949 |
-
"enable_perm":
|
| 2950 |
"group_num": 1,
|
| 2951 |
"group_size": 4096,
|
| 2952 |
"in_features": 4096,
|
|
@@ -2970,7 +2970,7 @@
|
|
| 2970 |
"model.layers.24.self_attn.k_proj": {
|
| 2971 |
"bias": false,
|
| 2972 |
"enable_norm": true,
|
| 2973 |
-
"enable_perm":
|
| 2974 |
"group_num": 1,
|
| 2975 |
"group_size": 4096,
|
| 2976 |
"in_features": 4096,
|
|
@@ -2994,7 +2994,7 @@
|
|
| 2994 |
"model.layers.24.self_attn.o_proj": {
|
| 2995 |
"bias": false,
|
| 2996 |
"enable_norm": true,
|
| 2997 |
-
"enable_perm":
|
| 2998 |
"group_num": 1,
|
| 2999 |
"group_size": 4096,
|
| 3000 |
"in_features": 4096,
|
|
@@ -3018,7 +3018,7 @@
|
|
| 3018 |
"model.layers.24.self_attn.q_proj": {
|
| 3019 |
"bias": false,
|
| 3020 |
"enable_norm": true,
|
| 3021 |
-
"enable_perm":
|
| 3022 |
"group_num": 1,
|
| 3023 |
"group_size": 4096,
|
| 3024 |
"in_features": 4096,
|
|
@@ -3042,7 +3042,7 @@
|
|
| 3042 |
"model.layers.24.self_attn.v_proj": {
|
| 3043 |
"bias": false,
|
| 3044 |
"enable_norm": true,
|
| 3045 |
-
"enable_perm":
|
| 3046 |
"group_num": 1,
|
| 3047 |
"group_size": 4096,
|
| 3048 |
"in_features": 4096,
|
|
@@ -3066,7 +3066,7 @@
|
|
| 3066 |
"model.layers.25.mlp.down_proj": {
|
| 3067 |
"bias": false,
|
| 3068 |
"enable_norm": true,
|
| 3069 |
-
"enable_perm":
|
| 3070 |
"group_num": 1,
|
| 3071 |
"group_size": 14336,
|
| 3072 |
"in_features": 14336,
|
|
@@ -3090,7 +3090,7 @@
|
|
| 3090 |
"model.layers.25.mlp.gate_proj": {
|
| 3091 |
"bias": false,
|
| 3092 |
"enable_norm": true,
|
| 3093 |
-
"enable_perm":
|
| 3094 |
"group_num": 1,
|
| 3095 |
"group_size": 4096,
|
| 3096 |
"in_features": 4096,
|
|
@@ -3114,7 +3114,7 @@
|
|
| 3114 |
"model.layers.25.mlp.up_proj": {
|
| 3115 |
"bias": false,
|
| 3116 |
"enable_norm": true,
|
| 3117 |
-
"enable_perm":
|
| 3118 |
"group_num": 1,
|
| 3119 |
"group_size": 4096,
|
| 3120 |
"in_features": 4096,
|
|
@@ -3138,7 +3138,7 @@
|
|
| 3138 |
"model.layers.25.self_attn.k_proj": {
|
| 3139 |
"bias": false,
|
| 3140 |
"enable_norm": true,
|
| 3141 |
-
"enable_perm":
|
| 3142 |
"group_num": 1,
|
| 3143 |
"group_size": 4096,
|
| 3144 |
"in_features": 4096,
|
|
@@ -3162,7 +3162,7 @@
|
|
| 3162 |
"model.layers.25.self_attn.o_proj": {
|
| 3163 |
"bias": false,
|
| 3164 |
"enable_norm": true,
|
| 3165 |
-
"enable_perm":
|
| 3166 |
"group_num": 1,
|
| 3167 |
"group_size": 4096,
|
| 3168 |
"in_features": 4096,
|
|
@@ -3186,7 +3186,7 @@
|
|
| 3186 |
"model.layers.25.self_attn.q_proj": {
|
| 3187 |
"bias": false,
|
| 3188 |
"enable_norm": true,
|
| 3189 |
-
"enable_perm":
|
| 3190 |
"group_num": 1,
|
| 3191 |
"group_size": 4096,
|
| 3192 |
"in_features": 4096,
|
|
@@ -3210,7 +3210,7 @@
|
|
| 3210 |
"model.layers.25.self_attn.v_proj": {
|
| 3211 |
"bias": false,
|
| 3212 |
"enable_norm": true,
|
| 3213 |
-
"enable_perm":
|
| 3214 |
"group_num": 1,
|
| 3215 |
"group_size": 4096,
|
| 3216 |
"in_features": 4096,
|
|
@@ -3234,7 +3234,7 @@
|
|
| 3234 |
"model.layers.26.mlp.down_proj": {
|
| 3235 |
"bias": false,
|
| 3236 |
"enable_norm": true,
|
| 3237 |
-
"enable_perm":
|
| 3238 |
"group_num": 1,
|
| 3239 |
"group_size": 14336,
|
| 3240 |
"in_features": 14336,
|
|
@@ -3258,7 +3258,7 @@
|
|
| 3258 |
"model.layers.26.mlp.gate_proj": {
|
| 3259 |
"bias": false,
|
| 3260 |
"enable_norm": true,
|
| 3261 |
-
"enable_perm":
|
| 3262 |
"group_num": 1,
|
| 3263 |
"group_size": 4096,
|
| 3264 |
"in_features": 4096,
|
|
@@ -3282,7 +3282,7 @@
|
|
| 3282 |
"model.layers.26.mlp.up_proj": {
|
| 3283 |
"bias": false,
|
| 3284 |
"enable_norm": true,
|
| 3285 |
-
"enable_perm":
|
| 3286 |
"group_num": 1,
|
| 3287 |
"group_size": 4096,
|
| 3288 |
"in_features": 4096,
|
|
@@ -3306,7 +3306,7 @@
|
|
| 3306 |
"model.layers.26.self_attn.k_proj": {
|
| 3307 |
"bias": false,
|
| 3308 |
"enable_norm": true,
|
| 3309 |
-
"enable_perm":
|
| 3310 |
"group_num": 1,
|
| 3311 |
"group_size": 4096,
|
| 3312 |
"in_features": 4096,
|
|
@@ -3330,7 +3330,7 @@
|
|
| 3330 |
"model.layers.26.self_attn.o_proj": {
|
| 3331 |
"bias": false,
|
| 3332 |
"enable_norm": true,
|
| 3333 |
-
"enable_perm":
|
| 3334 |
"group_num": 1,
|
| 3335 |
"group_size": 4096,
|
| 3336 |
"in_features": 4096,
|
|
@@ -3354,7 +3354,7 @@
|
|
| 3354 |
"model.layers.26.self_attn.q_proj": {
|
| 3355 |
"bias": false,
|
| 3356 |
"enable_norm": true,
|
| 3357 |
-
"enable_perm":
|
| 3358 |
"group_num": 1,
|
| 3359 |
"group_size": 4096,
|
| 3360 |
"in_features": 4096,
|
|
@@ -3378,7 +3378,7 @@
|
|
| 3378 |
"model.layers.26.self_attn.v_proj": {
|
| 3379 |
"bias": false,
|
| 3380 |
"enable_norm": true,
|
| 3381 |
-
"enable_perm":
|
| 3382 |
"group_num": 1,
|
| 3383 |
"group_size": 4096,
|
| 3384 |
"in_features": 4096,
|
|
@@ -3402,7 +3402,7 @@
|
|
| 3402 |
"model.layers.27.mlp.down_proj": {
|
| 3403 |
"bias": false,
|
| 3404 |
"enable_norm": true,
|
| 3405 |
-
"enable_perm":
|
| 3406 |
"group_num": 1,
|
| 3407 |
"group_size": 14336,
|
| 3408 |
"in_features": 14336,
|
|
@@ -3426,7 +3426,7 @@
|
|
| 3426 |
"model.layers.27.mlp.gate_proj": {
|
| 3427 |
"bias": false,
|
| 3428 |
"enable_norm": true,
|
| 3429 |
-
"enable_perm":
|
| 3430 |
"group_num": 1,
|
| 3431 |
"group_size": 4096,
|
| 3432 |
"in_features": 4096,
|
|
@@ -3450,7 +3450,7 @@
|
|
| 3450 |
"model.layers.27.mlp.up_proj": {
|
| 3451 |
"bias": false,
|
| 3452 |
"enable_norm": true,
|
| 3453 |
-
"enable_perm":
|
| 3454 |
"group_num": 1,
|
| 3455 |
"group_size": 4096,
|
| 3456 |
"in_features": 4096,
|
|
@@ -3474,7 +3474,7 @@
|
|
| 3474 |
"model.layers.27.self_attn.k_proj": {
|
| 3475 |
"bias": false,
|
| 3476 |
"enable_norm": true,
|
| 3477 |
-
"enable_perm":
|
| 3478 |
"group_num": 1,
|
| 3479 |
"group_size": 4096,
|
| 3480 |
"in_features": 4096,
|
|
@@ -3498,7 +3498,7 @@
|
|
| 3498 |
"model.layers.27.self_attn.o_proj": {
|
| 3499 |
"bias": false,
|
| 3500 |
"enable_norm": true,
|
| 3501 |
-
"enable_perm":
|
| 3502 |
"group_num": 1,
|
| 3503 |
"group_size": 4096,
|
| 3504 |
"in_features": 4096,
|
|
@@ -3522,7 +3522,7 @@
|
|
| 3522 |
"model.layers.27.self_attn.q_proj": {
|
| 3523 |
"bias": false,
|
| 3524 |
"enable_norm": true,
|
| 3525 |
-
"enable_perm":
|
| 3526 |
"group_num": 1,
|
| 3527 |
"group_size": 4096,
|
| 3528 |
"in_features": 4096,
|
|
@@ -3546,7 +3546,7 @@
|
|
| 3546 |
"model.layers.27.self_attn.v_proj": {
|
| 3547 |
"bias": false,
|
| 3548 |
"enable_norm": true,
|
| 3549 |
-
"enable_perm":
|
| 3550 |
"group_num": 1,
|
| 3551 |
"group_size": 4096,
|
| 3552 |
"in_features": 4096,
|
|
@@ -3570,7 +3570,7 @@
|
|
| 3570 |
"model.layers.28.mlp.down_proj": {
|
| 3571 |
"bias": false,
|
| 3572 |
"enable_norm": true,
|
| 3573 |
-
"enable_perm":
|
| 3574 |
"group_num": 1,
|
| 3575 |
"group_size": 14336,
|
| 3576 |
"in_features": 14336,
|
|
@@ -3594,7 +3594,7 @@
|
|
| 3594 |
"model.layers.28.mlp.gate_proj": {
|
| 3595 |
"bias": false,
|
| 3596 |
"enable_norm": true,
|
| 3597 |
-
"enable_perm":
|
| 3598 |
"group_num": 1,
|
| 3599 |
"group_size": 4096,
|
| 3600 |
"in_features": 4096,
|
|
@@ -3618,7 +3618,7 @@
|
|
| 3618 |
"model.layers.28.mlp.up_proj": {
|
| 3619 |
"bias": false,
|
| 3620 |
"enable_norm": true,
|
| 3621 |
-
"enable_perm":
|
| 3622 |
"group_num": 1,
|
| 3623 |
"group_size": 4096,
|
| 3624 |
"in_features": 4096,
|
|
@@ -3642,7 +3642,7 @@
|
|
| 3642 |
"model.layers.28.self_attn.k_proj": {
|
| 3643 |
"bias": false,
|
| 3644 |
"enable_norm": true,
|
| 3645 |
-
"enable_perm":
|
| 3646 |
"group_num": 1,
|
| 3647 |
"group_size": 4096,
|
| 3648 |
"in_features": 4096,
|
|
@@ -3666,7 +3666,7 @@
|
|
| 3666 |
"model.layers.28.self_attn.o_proj": {
|
| 3667 |
"bias": false,
|
| 3668 |
"enable_norm": true,
|
| 3669 |
-
"enable_perm":
|
| 3670 |
"group_num": 1,
|
| 3671 |
"group_size": 4096,
|
| 3672 |
"in_features": 4096,
|
|
@@ -3690,7 +3690,7 @@
|
|
| 3690 |
"model.layers.28.self_attn.q_proj": {
|
| 3691 |
"bias": false,
|
| 3692 |
"enable_norm": true,
|
| 3693 |
-
"enable_perm":
|
| 3694 |
"group_num": 1,
|
| 3695 |
"group_size": 4096,
|
| 3696 |
"in_features": 4096,
|
|
@@ -3714,7 +3714,7 @@
|
|
| 3714 |
"model.layers.28.self_attn.v_proj": {
|
| 3715 |
"bias": false,
|
| 3716 |
"enable_norm": true,
|
| 3717 |
-
"enable_perm":
|
| 3718 |
"group_num": 1,
|
| 3719 |
"group_size": 4096,
|
| 3720 |
"in_features": 4096,
|
|
@@ -3738,7 +3738,7 @@
|
|
| 3738 |
"model.layers.29.mlp.down_proj": {
|
| 3739 |
"bias": false,
|
| 3740 |
"enable_norm": true,
|
| 3741 |
-
"enable_perm":
|
| 3742 |
"group_num": 1,
|
| 3743 |
"group_size": 14336,
|
| 3744 |
"in_features": 14336,
|
|
@@ -3762,7 +3762,7 @@
|
|
| 3762 |
"model.layers.29.mlp.gate_proj": {
|
| 3763 |
"bias": false,
|
| 3764 |
"enable_norm": true,
|
| 3765 |
-
"enable_perm":
|
| 3766 |
"group_num": 1,
|
| 3767 |
"group_size": 4096,
|
| 3768 |
"in_features": 4096,
|
|
@@ -3786,7 +3786,7 @@
|
|
| 3786 |
"model.layers.29.mlp.up_proj": {
|
| 3787 |
"bias": false,
|
| 3788 |
"enable_norm": true,
|
| 3789 |
-
"enable_perm":
|
| 3790 |
"group_num": 1,
|
| 3791 |
"group_size": 4096,
|
| 3792 |
"in_features": 4096,
|
|
@@ -3810,7 +3810,7 @@
|
|
| 3810 |
"model.layers.29.self_attn.k_proj": {
|
| 3811 |
"bias": false,
|
| 3812 |
"enable_norm": true,
|
| 3813 |
-
"enable_perm":
|
| 3814 |
"group_num": 1,
|
| 3815 |
"group_size": 4096,
|
| 3816 |
"in_features": 4096,
|
|
@@ -3834,7 +3834,7 @@
|
|
| 3834 |
"model.layers.29.self_attn.o_proj": {
|
| 3835 |
"bias": false,
|
| 3836 |
"enable_norm": true,
|
| 3837 |
-
"enable_perm":
|
| 3838 |
"group_num": 1,
|
| 3839 |
"group_size": 4096,
|
| 3840 |
"in_features": 4096,
|
|
@@ -3858,7 +3858,7 @@
|
|
| 3858 |
"model.layers.29.self_attn.q_proj": {
|
| 3859 |
"bias": false,
|
| 3860 |
"enable_norm": true,
|
| 3861 |
-
"enable_perm":
|
| 3862 |
"group_num": 1,
|
| 3863 |
"group_size": 4096,
|
| 3864 |
"in_features": 4096,
|
|
@@ -3882,7 +3882,7 @@
|
|
| 3882 |
"model.layers.29.self_attn.v_proj": {
|
| 3883 |
"bias": false,
|
| 3884 |
"enable_norm": true,
|
| 3885 |
-
"enable_perm":
|
| 3886 |
"group_num": 1,
|
| 3887 |
"group_size": 4096,
|
| 3888 |
"in_features": 4096,
|
|
@@ -3906,7 +3906,7 @@
|
|
| 3906 |
"model.layers.3.mlp.down_proj": {
|
| 3907 |
"bias": false,
|
| 3908 |
"enable_norm": true,
|
| 3909 |
-
"enable_perm":
|
| 3910 |
"group_num": 1,
|
| 3911 |
"group_size": 14336,
|
| 3912 |
"in_features": 14336,
|
|
@@ -3930,7 +3930,7 @@
|
|
| 3930 |
"model.layers.3.mlp.gate_proj": {
|
| 3931 |
"bias": false,
|
| 3932 |
"enable_norm": true,
|
| 3933 |
-
"enable_perm":
|
| 3934 |
"group_num": 1,
|
| 3935 |
"group_size": 4096,
|
| 3936 |
"in_features": 4096,
|
|
@@ -3954,7 +3954,7 @@
|
|
| 3954 |
"model.layers.3.mlp.up_proj": {
|
| 3955 |
"bias": false,
|
| 3956 |
"enable_norm": true,
|
| 3957 |
-
"enable_perm":
|
| 3958 |
"group_num": 1,
|
| 3959 |
"group_size": 4096,
|
| 3960 |
"in_features": 4096,
|
|
@@ -3978,7 +3978,7 @@
|
|
| 3978 |
"model.layers.3.self_attn.k_proj": {
|
| 3979 |
"bias": false,
|
| 3980 |
"enable_norm": true,
|
| 3981 |
-
"enable_perm":
|
| 3982 |
"group_num": 1,
|
| 3983 |
"group_size": 4096,
|
| 3984 |
"in_features": 4096,
|
|
@@ -4002,7 +4002,7 @@
|
|
| 4002 |
"model.layers.3.self_attn.o_proj": {
|
| 4003 |
"bias": false,
|
| 4004 |
"enable_norm": true,
|
| 4005 |
-
"enable_perm":
|
| 4006 |
"group_num": 1,
|
| 4007 |
"group_size": 4096,
|
| 4008 |
"in_features": 4096,
|
|
@@ -4026,7 +4026,7 @@
|
|
| 4026 |
"model.layers.3.self_attn.q_proj": {
|
| 4027 |
"bias": false,
|
| 4028 |
"enable_norm": true,
|
| 4029 |
-
"enable_perm":
|
| 4030 |
"group_num": 1,
|
| 4031 |
"group_size": 4096,
|
| 4032 |
"in_features": 4096,
|
|
@@ -4050,7 +4050,7 @@
|
|
| 4050 |
"model.layers.3.self_attn.v_proj": {
|
| 4051 |
"bias": false,
|
| 4052 |
"enable_norm": true,
|
| 4053 |
-
"enable_perm":
|
| 4054 |
"group_num": 1,
|
| 4055 |
"group_size": 4096,
|
| 4056 |
"in_features": 4096,
|
|
@@ -4074,7 +4074,7 @@
|
|
| 4074 |
"model.layers.30.mlp.down_proj": {
|
| 4075 |
"bias": false,
|
| 4076 |
"enable_norm": true,
|
| 4077 |
-
"enable_perm":
|
| 4078 |
"group_num": 1,
|
| 4079 |
"group_size": 14336,
|
| 4080 |
"in_features": 14336,
|
|
@@ -4098,7 +4098,7 @@
|
|
| 4098 |
"model.layers.30.mlp.gate_proj": {
|
| 4099 |
"bias": false,
|
| 4100 |
"enable_norm": true,
|
| 4101 |
-
"enable_perm":
|
| 4102 |
"group_num": 1,
|
| 4103 |
"group_size": 4096,
|
| 4104 |
"in_features": 4096,
|
|
@@ -4122,7 +4122,7 @@
|
|
| 4122 |
"model.layers.30.mlp.up_proj": {
|
| 4123 |
"bias": false,
|
| 4124 |
"enable_norm": true,
|
| 4125 |
-
"enable_perm":
|
| 4126 |
"group_num": 1,
|
| 4127 |
"group_size": 4096,
|
| 4128 |
"in_features": 4096,
|
|
@@ -4146,7 +4146,7 @@
|
|
| 4146 |
"model.layers.30.self_attn.k_proj": {
|
| 4147 |
"bias": false,
|
| 4148 |
"enable_norm": true,
|
| 4149 |
-
"enable_perm":
|
| 4150 |
"group_num": 1,
|
| 4151 |
"group_size": 4096,
|
| 4152 |
"in_features": 4096,
|
|
@@ -4170,7 +4170,7 @@
|
|
| 4170 |
"model.layers.30.self_attn.o_proj": {
|
| 4171 |
"bias": false,
|
| 4172 |
"enable_norm": true,
|
| 4173 |
-
"enable_perm":
|
| 4174 |
"group_num": 1,
|
| 4175 |
"group_size": 4096,
|
| 4176 |
"in_features": 4096,
|
|
@@ -4194,7 +4194,7 @@
|
|
| 4194 |
"model.layers.30.self_attn.q_proj": {
|
| 4195 |
"bias": false,
|
| 4196 |
"enable_norm": true,
|
| 4197 |
-
"enable_perm":
|
| 4198 |
"group_num": 1,
|
| 4199 |
"group_size": 4096,
|
| 4200 |
"in_features": 4096,
|
|
@@ -4218,7 +4218,7 @@
|
|
| 4218 |
"model.layers.30.self_attn.v_proj": {
|
| 4219 |
"bias": false,
|
| 4220 |
"enable_norm": true,
|
| 4221 |
-
"enable_perm":
|
| 4222 |
"group_num": 1,
|
| 4223 |
"group_size": 4096,
|
| 4224 |
"in_features": 4096,
|
|
@@ -4242,7 +4242,7 @@
|
|
| 4242 |
"model.layers.31.mlp.down_proj": {
|
| 4243 |
"bias": false,
|
| 4244 |
"enable_norm": true,
|
| 4245 |
-
"enable_perm":
|
| 4246 |
"group_num": 1,
|
| 4247 |
"group_size": 14336,
|
| 4248 |
"in_features": 14336,
|
|
@@ -4266,7 +4266,7 @@
|
|
| 4266 |
"model.layers.31.mlp.gate_proj": {
|
| 4267 |
"bias": false,
|
| 4268 |
"enable_norm": true,
|
| 4269 |
-
"enable_perm":
|
| 4270 |
"group_num": 1,
|
| 4271 |
"group_size": 4096,
|
| 4272 |
"in_features": 4096,
|
|
@@ -4290,7 +4290,7 @@
|
|
| 4290 |
"model.layers.31.mlp.up_proj": {
|
| 4291 |
"bias": false,
|
| 4292 |
"enable_norm": true,
|
| 4293 |
-
"enable_perm":
|
| 4294 |
"group_num": 1,
|
| 4295 |
"group_size": 4096,
|
| 4296 |
"in_features": 4096,
|
|
@@ -4314,7 +4314,7 @@
|
|
| 4314 |
"model.layers.31.self_attn.k_proj": {
|
| 4315 |
"bias": false,
|
| 4316 |
"enable_norm": true,
|
| 4317 |
-
"enable_perm":
|
| 4318 |
"group_num": 1,
|
| 4319 |
"group_size": 4096,
|
| 4320 |
"in_features": 4096,
|
|
@@ -4338,7 +4338,7 @@
|
|
| 4338 |
"model.layers.31.self_attn.o_proj": {
|
| 4339 |
"bias": false,
|
| 4340 |
"enable_norm": true,
|
| 4341 |
-
"enable_perm":
|
| 4342 |
"group_num": 1,
|
| 4343 |
"group_size": 4096,
|
| 4344 |
"in_features": 4096,
|
|
@@ -4362,7 +4362,7 @@
|
|
| 4362 |
"model.layers.31.self_attn.q_proj": {
|
| 4363 |
"bias": false,
|
| 4364 |
"enable_norm": true,
|
| 4365 |
-
"enable_perm":
|
| 4366 |
"group_num": 1,
|
| 4367 |
"group_size": 4096,
|
| 4368 |
"in_features": 4096,
|
|
@@ -4386,7 +4386,7 @@
|
|
| 4386 |
"model.layers.31.self_attn.v_proj": {
|
| 4387 |
"bias": false,
|
| 4388 |
"enable_norm": true,
|
| 4389 |
-
"enable_perm":
|
| 4390 |
"group_num": 1,
|
| 4391 |
"group_size": 4096,
|
| 4392 |
"in_features": 4096,
|
|
@@ -4410,7 +4410,7 @@
|
|
| 4410 |
"model.layers.4.mlp.down_proj": {
|
| 4411 |
"bias": false,
|
| 4412 |
"enable_norm": true,
|
| 4413 |
-
"enable_perm":
|
| 4414 |
"group_num": 1,
|
| 4415 |
"group_size": 14336,
|
| 4416 |
"in_features": 14336,
|
|
@@ -4434,7 +4434,7 @@
|
|
| 4434 |
"model.layers.4.mlp.gate_proj": {
|
| 4435 |
"bias": false,
|
| 4436 |
"enable_norm": true,
|
| 4437 |
-
"enable_perm":
|
| 4438 |
"group_num": 1,
|
| 4439 |
"group_size": 4096,
|
| 4440 |
"in_features": 4096,
|
|
@@ -4458,7 +4458,7 @@
|
|
| 4458 |
"model.layers.4.mlp.up_proj": {
|
| 4459 |
"bias": false,
|
| 4460 |
"enable_norm": true,
|
| 4461 |
-
"enable_perm":
|
| 4462 |
"group_num": 1,
|
| 4463 |
"group_size": 4096,
|
| 4464 |
"in_features": 4096,
|
|
@@ -4482,7 +4482,7 @@
|
|
| 4482 |
"model.layers.4.self_attn.k_proj": {
|
| 4483 |
"bias": false,
|
| 4484 |
"enable_norm": true,
|
| 4485 |
-
"enable_perm":
|
| 4486 |
"group_num": 1,
|
| 4487 |
"group_size": 4096,
|
| 4488 |
"in_features": 4096,
|
|
@@ -4506,7 +4506,7 @@
|
|
| 4506 |
"model.layers.4.self_attn.o_proj": {
|
| 4507 |
"bias": false,
|
| 4508 |
"enable_norm": true,
|
| 4509 |
-
"enable_perm":
|
| 4510 |
"group_num": 1,
|
| 4511 |
"group_size": 4096,
|
| 4512 |
"in_features": 4096,
|
|
@@ -4530,7 +4530,7 @@
|
|
| 4530 |
"model.layers.4.self_attn.q_proj": {
|
| 4531 |
"bias": false,
|
| 4532 |
"enable_norm": true,
|
| 4533 |
-
"enable_perm":
|
| 4534 |
"group_num": 1,
|
| 4535 |
"group_size": 4096,
|
| 4536 |
"in_features": 4096,
|
|
@@ -4554,7 +4554,7 @@
|
|
| 4554 |
"model.layers.4.self_attn.v_proj": {
|
| 4555 |
"bias": false,
|
| 4556 |
"enable_norm": true,
|
| 4557 |
-
"enable_perm":
|
| 4558 |
"group_num": 1,
|
| 4559 |
"group_size": 4096,
|
| 4560 |
"in_features": 4096,
|
|
@@ -4578,7 +4578,7 @@
|
|
| 4578 |
"model.layers.5.mlp.down_proj": {
|
| 4579 |
"bias": false,
|
| 4580 |
"enable_norm": true,
|
| 4581 |
-
"enable_perm":
|
| 4582 |
"group_num": 1,
|
| 4583 |
"group_size": 14336,
|
| 4584 |
"in_features": 14336,
|
|
@@ -4602,7 +4602,7 @@
|
|
| 4602 |
"model.layers.5.mlp.gate_proj": {
|
| 4603 |
"bias": false,
|
| 4604 |
"enable_norm": true,
|
| 4605 |
-
"enable_perm":
|
| 4606 |
"group_num": 1,
|
| 4607 |
"group_size": 4096,
|
| 4608 |
"in_features": 4096,
|
|
@@ -4626,7 +4626,7 @@
|
|
| 4626 |
"model.layers.5.mlp.up_proj": {
|
| 4627 |
"bias": false,
|
| 4628 |
"enable_norm": true,
|
| 4629 |
-
"enable_perm":
|
| 4630 |
"group_num": 1,
|
| 4631 |
"group_size": 4096,
|
| 4632 |
"in_features": 4096,
|
|
@@ -4650,7 +4650,7 @@
|
|
| 4650 |
"model.layers.5.self_attn.k_proj": {
|
| 4651 |
"bias": false,
|
| 4652 |
"enable_norm": true,
|
| 4653 |
-
"enable_perm":
|
| 4654 |
"group_num": 1,
|
| 4655 |
"group_size": 4096,
|
| 4656 |
"in_features": 4096,
|
|
@@ -4674,7 +4674,7 @@
|
|
| 4674 |
"model.layers.5.self_attn.o_proj": {
|
| 4675 |
"bias": false,
|
| 4676 |
"enable_norm": true,
|
| 4677 |
-
"enable_perm":
|
| 4678 |
"group_num": 1,
|
| 4679 |
"group_size": 4096,
|
| 4680 |
"in_features": 4096,
|
|
@@ -4698,7 +4698,7 @@
|
|
| 4698 |
"model.layers.5.self_attn.q_proj": {
|
| 4699 |
"bias": false,
|
| 4700 |
"enable_norm": true,
|
| 4701 |
-
"enable_perm":
|
| 4702 |
"group_num": 1,
|
| 4703 |
"group_size": 4096,
|
| 4704 |
"in_features": 4096,
|
|
@@ -4722,7 +4722,7 @@
|
|
| 4722 |
"model.layers.5.self_attn.v_proj": {
|
| 4723 |
"bias": false,
|
| 4724 |
"enable_norm": true,
|
| 4725 |
-
"enable_perm":
|
| 4726 |
"group_num": 1,
|
| 4727 |
"group_size": 4096,
|
| 4728 |
"in_features": 4096,
|
|
@@ -4746,7 +4746,7 @@
|
|
| 4746 |
"model.layers.6.mlp.down_proj": {
|
| 4747 |
"bias": false,
|
| 4748 |
"enable_norm": true,
|
| 4749 |
-
"enable_perm":
|
| 4750 |
"group_num": 1,
|
| 4751 |
"group_size": 14336,
|
| 4752 |
"in_features": 14336,
|
|
@@ -4770,7 +4770,7 @@
|
|
| 4770 |
"model.layers.6.mlp.gate_proj": {
|
| 4771 |
"bias": false,
|
| 4772 |
"enable_norm": true,
|
| 4773 |
-
"enable_perm":
|
| 4774 |
"group_num": 1,
|
| 4775 |
"group_size": 4096,
|
| 4776 |
"in_features": 4096,
|
|
@@ -4794,7 +4794,7 @@
|
|
| 4794 |
"model.layers.6.mlp.up_proj": {
|
| 4795 |
"bias": false,
|
| 4796 |
"enable_norm": true,
|
| 4797 |
-
"enable_perm":
|
| 4798 |
"group_num": 1,
|
| 4799 |
"group_size": 4096,
|
| 4800 |
"in_features": 4096,
|
|
@@ -4818,7 +4818,7 @@
|
|
| 4818 |
"model.layers.6.self_attn.k_proj": {
|
| 4819 |
"bias": false,
|
| 4820 |
"enable_norm": true,
|
| 4821 |
-
"enable_perm":
|
| 4822 |
"group_num": 1,
|
| 4823 |
"group_size": 4096,
|
| 4824 |
"in_features": 4096,
|
|
@@ -4842,7 +4842,7 @@
|
|
| 4842 |
"model.layers.6.self_attn.o_proj": {
|
| 4843 |
"bias": false,
|
| 4844 |
"enable_norm": true,
|
| 4845 |
-
"enable_perm":
|
| 4846 |
"group_num": 1,
|
| 4847 |
"group_size": 4096,
|
| 4848 |
"in_features": 4096,
|
|
@@ -4866,7 +4866,7 @@
|
|
| 4866 |
"model.layers.6.self_attn.q_proj": {
|
| 4867 |
"bias": false,
|
| 4868 |
"enable_norm": true,
|
| 4869 |
-
"enable_perm":
|
| 4870 |
"group_num": 1,
|
| 4871 |
"group_size": 4096,
|
| 4872 |
"in_features": 4096,
|
|
@@ -4890,7 +4890,7 @@
|
|
| 4890 |
"model.layers.6.self_attn.v_proj": {
|
| 4891 |
"bias": false,
|
| 4892 |
"enable_norm": true,
|
| 4893 |
-
"enable_perm":
|
| 4894 |
"group_num": 1,
|
| 4895 |
"group_size": 4096,
|
| 4896 |
"in_features": 4096,
|
|
@@ -4914,7 +4914,7 @@
|
|
| 4914 |
"model.layers.7.mlp.down_proj": {
|
| 4915 |
"bias": false,
|
| 4916 |
"enable_norm": true,
|
| 4917 |
-
"enable_perm":
|
| 4918 |
"group_num": 1,
|
| 4919 |
"group_size": 14336,
|
| 4920 |
"in_features": 14336,
|
|
@@ -4938,7 +4938,7 @@
|
|
| 4938 |
"model.layers.7.mlp.gate_proj": {
|
| 4939 |
"bias": false,
|
| 4940 |
"enable_norm": true,
|
| 4941 |
-
"enable_perm":
|
| 4942 |
"group_num": 1,
|
| 4943 |
"group_size": 4096,
|
| 4944 |
"in_features": 4096,
|
|
@@ -4962,7 +4962,7 @@
|
|
| 4962 |
"model.layers.7.mlp.up_proj": {
|
| 4963 |
"bias": false,
|
| 4964 |
"enable_norm": true,
|
| 4965 |
-
"enable_perm":
|
| 4966 |
"group_num": 1,
|
| 4967 |
"group_size": 4096,
|
| 4968 |
"in_features": 4096,
|
|
@@ -4986,7 +4986,7 @@
|
|
| 4986 |
"model.layers.7.self_attn.k_proj": {
|
| 4987 |
"bias": false,
|
| 4988 |
"enable_norm": true,
|
| 4989 |
-
"enable_perm":
|
| 4990 |
"group_num": 1,
|
| 4991 |
"group_size": 4096,
|
| 4992 |
"in_features": 4096,
|
|
@@ -5010,7 +5010,7 @@
|
|
| 5010 |
"model.layers.7.self_attn.o_proj": {
|
| 5011 |
"bias": false,
|
| 5012 |
"enable_norm": true,
|
| 5013 |
-
"enable_perm":
|
| 5014 |
"group_num": 1,
|
| 5015 |
"group_size": 4096,
|
| 5016 |
"in_features": 4096,
|
|
@@ -5034,7 +5034,7 @@
|
|
| 5034 |
"model.layers.7.self_attn.q_proj": {
|
| 5035 |
"bias": false,
|
| 5036 |
"enable_norm": true,
|
| 5037 |
-
"enable_perm":
|
| 5038 |
"group_num": 1,
|
| 5039 |
"group_size": 4096,
|
| 5040 |
"in_features": 4096,
|
|
@@ -5058,7 +5058,7 @@
|
|
| 5058 |
"model.layers.7.self_attn.v_proj": {
|
| 5059 |
"bias": false,
|
| 5060 |
"enable_norm": true,
|
| 5061 |
-
"enable_perm":
|
| 5062 |
"group_num": 1,
|
| 5063 |
"group_size": 4096,
|
| 5064 |
"in_features": 4096,
|
|
@@ -5082,7 +5082,7 @@
|
|
| 5082 |
"model.layers.8.mlp.down_proj": {
|
| 5083 |
"bias": false,
|
| 5084 |
"enable_norm": true,
|
| 5085 |
-
"enable_perm":
|
| 5086 |
"group_num": 1,
|
| 5087 |
"group_size": 14336,
|
| 5088 |
"in_features": 14336,
|
|
@@ -5106,7 +5106,7 @@
|
|
| 5106 |
"model.layers.8.mlp.gate_proj": {
|
| 5107 |
"bias": false,
|
| 5108 |
"enable_norm": true,
|
| 5109 |
-
"enable_perm":
|
| 5110 |
"group_num": 1,
|
| 5111 |
"group_size": 4096,
|
| 5112 |
"in_features": 4096,
|
|
@@ -5130,7 +5130,7 @@
|
|
| 5130 |
"model.layers.8.mlp.up_proj": {
|
| 5131 |
"bias": false,
|
| 5132 |
"enable_norm": true,
|
| 5133 |
-
"enable_perm":
|
| 5134 |
"group_num": 1,
|
| 5135 |
"group_size": 4096,
|
| 5136 |
"in_features": 4096,
|
|
@@ -5154,7 +5154,7 @@
|
|
| 5154 |
"model.layers.8.self_attn.k_proj": {
|
| 5155 |
"bias": false,
|
| 5156 |
"enable_norm": true,
|
| 5157 |
-
"enable_perm":
|
| 5158 |
"group_num": 1,
|
| 5159 |
"group_size": 4096,
|
| 5160 |
"in_features": 4096,
|
|
@@ -5178,7 +5178,7 @@
|
|
| 5178 |
"model.layers.8.self_attn.o_proj": {
|
| 5179 |
"bias": false,
|
| 5180 |
"enable_norm": true,
|
| 5181 |
-
"enable_perm":
|
| 5182 |
"group_num": 1,
|
| 5183 |
"group_size": 4096,
|
| 5184 |
"in_features": 4096,
|
|
@@ -5202,7 +5202,7 @@
|
|
| 5202 |
"model.layers.8.self_attn.q_proj": {
|
| 5203 |
"bias": false,
|
| 5204 |
"enable_norm": true,
|
| 5205 |
-
"enable_perm":
|
| 5206 |
"group_num": 1,
|
| 5207 |
"group_size": 4096,
|
| 5208 |
"in_features": 4096,
|
|
@@ -5226,7 +5226,7 @@
|
|
| 5226 |
"model.layers.8.self_attn.v_proj": {
|
| 5227 |
"bias": false,
|
| 5228 |
"enable_norm": true,
|
| 5229 |
-
"enable_perm":
|
| 5230 |
"group_num": 1,
|
| 5231 |
"group_size": 4096,
|
| 5232 |
"in_features": 4096,
|
|
@@ -5250,7 +5250,7 @@
|
|
| 5250 |
"model.layers.9.mlp.down_proj": {
|
| 5251 |
"bias": false,
|
| 5252 |
"enable_norm": true,
|
| 5253 |
-
"enable_perm":
|
| 5254 |
"group_num": 1,
|
| 5255 |
"group_size": 14336,
|
| 5256 |
"in_features": 14336,
|
|
@@ -5274,7 +5274,7 @@
|
|
| 5274 |
"model.layers.9.mlp.gate_proj": {
|
| 5275 |
"bias": false,
|
| 5276 |
"enable_norm": true,
|
| 5277 |
-
"enable_perm":
|
| 5278 |
"group_num": 1,
|
| 5279 |
"group_size": 4096,
|
| 5280 |
"in_features": 4096,
|
|
@@ -5298,7 +5298,7 @@
|
|
| 5298 |
"model.layers.9.mlp.up_proj": {
|
| 5299 |
"bias": false,
|
| 5300 |
"enable_norm": true,
|
| 5301 |
-
"enable_perm":
|
| 5302 |
"group_num": 1,
|
| 5303 |
"group_size": 4096,
|
| 5304 |
"in_features": 4096,
|
|
@@ -5322,7 +5322,7 @@
|
|
| 5322 |
"model.layers.9.self_attn.k_proj": {
|
| 5323 |
"bias": false,
|
| 5324 |
"enable_norm": true,
|
| 5325 |
-
"enable_perm":
|
| 5326 |
"group_num": 1,
|
| 5327 |
"group_size": 4096,
|
| 5328 |
"in_features": 4096,
|
|
@@ -5346,7 +5346,7 @@
|
|
| 5346 |
"model.layers.9.self_attn.o_proj": {
|
| 5347 |
"bias": false,
|
| 5348 |
"enable_norm": true,
|
| 5349 |
-
"enable_perm":
|
| 5350 |
"group_num": 1,
|
| 5351 |
"group_size": 4096,
|
| 5352 |
"in_features": 4096,
|
|
@@ -5370,7 +5370,7 @@
|
|
| 5370 |
"model.layers.9.self_attn.q_proj": {
|
| 5371 |
"bias": false,
|
| 5372 |
"enable_norm": true,
|
| 5373 |
-
"enable_perm":
|
| 5374 |
"group_num": 1,
|
| 5375 |
"group_size": 4096,
|
| 5376 |
"in_features": 4096,
|
|
@@ -5394,7 +5394,7 @@
|
|
| 5394 |
"model.layers.9.self_attn.v_proj": {
|
| 5395 |
"bias": false,
|
| 5396 |
"enable_norm": true,
|
| 5397 |
-
"enable_perm":
|
| 5398 |
"group_num": 1,
|
| 5399 |
"group_size": 4096,
|
| 5400 |
"in_features": 4096,
|
|
|
|
| 42 |
"model.layers.0.mlp.down_proj": {
|
| 43 |
"bias": false,
|
| 44 |
"enable_norm": true,
|
| 45 |
+
"enable_perm": false,
|
| 46 |
"group_num": 1,
|
| 47 |
"group_size": 14336,
|
| 48 |
"in_features": 14336,
|
|
|
|
| 66 |
"model.layers.0.mlp.gate_proj": {
|
| 67 |
"bias": false,
|
| 68 |
"enable_norm": true,
|
| 69 |
+
"enable_perm": false,
|
| 70 |
"group_num": 1,
|
| 71 |
"group_size": 4096,
|
| 72 |
"in_features": 4096,
|
|
|
|
| 90 |
"model.layers.0.mlp.up_proj": {
|
| 91 |
"bias": false,
|
| 92 |
"enable_norm": true,
|
| 93 |
+
"enable_perm": false,
|
| 94 |
"group_num": 1,
|
| 95 |
"group_size": 4096,
|
| 96 |
"in_features": 4096,
|
|
|
|
| 114 |
"model.layers.0.self_attn.k_proj": {
|
| 115 |
"bias": false,
|
| 116 |
"enable_norm": true,
|
| 117 |
+
"enable_perm": false,
|
| 118 |
"group_num": 1,
|
| 119 |
"group_size": 4096,
|
| 120 |
"in_features": 4096,
|
|
|
|
| 138 |
"model.layers.0.self_attn.o_proj": {
|
| 139 |
"bias": false,
|
| 140 |
"enable_norm": true,
|
| 141 |
+
"enable_perm": false,
|
| 142 |
"group_num": 1,
|
| 143 |
"group_size": 4096,
|
| 144 |
"in_features": 4096,
|
|
|
|
| 162 |
"model.layers.0.self_attn.q_proj": {
|
| 163 |
"bias": false,
|
| 164 |
"enable_norm": true,
|
| 165 |
+
"enable_perm": false,
|
| 166 |
"group_num": 1,
|
| 167 |
"group_size": 4096,
|
| 168 |
"in_features": 4096,
|
|
|
|
| 186 |
"model.layers.0.self_attn.v_proj": {
|
| 187 |
"bias": false,
|
| 188 |
"enable_norm": true,
|
| 189 |
+
"enable_perm": false,
|
| 190 |
"group_num": 1,
|
| 191 |
"group_size": 4096,
|
| 192 |
"in_features": 4096,
|
|
|
|
| 210 |
"model.layers.1.mlp.down_proj": {
|
| 211 |
"bias": false,
|
| 212 |
"enable_norm": true,
|
| 213 |
+
"enable_perm": false,
|
| 214 |
"group_num": 1,
|
| 215 |
"group_size": 14336,
|
| 216 |
"in_features": 14336,
|
|
|
|
| 234 |
"model.layers.1.mlp.gate_proj": {
|
| 235 |
"bias": false,
|
| 236 |
"enable_norm": true,
|
| 237 |
+
"enable_perm": false,
|
| 238 |
"group_num": 1,
|
| 239 |
"group_size": 4096,
|
| 240 |
"in_features": 4096,
|
|
|
|
| 258 |
"model.layers.1.mlp.up_proj": {
|
| 259 |
"bias": false,
|
| 260 |
"enable_norm": true,
|
| 261 |
+
"enable_perm": false,
|
| 262 |
"group_num": 1,
|
| 263 |
"group_size": 4096,
|
| 264 |
"in_features": 4096,
|
|
|
|
| 282 |
"model.layers.1.self_attn.k_proj": {
|
| 283 |
"bias": false,
|
| 284 |
"enable_norm": true,
|
| 285 |
+
"enable_perm": false,
|
| 286 |
"group_num": 1,
|
| 287 |
"group_size": 4096,
|
| 288 |
"in_features": 4096,
|
|
|
|
| 306 |
"model.layers.1.self_attn.o_proj": {
|
| 307 |
"bias": false,
|
| 308 |
"enable_norm": true,
|
| 309 |
+
"enable_perm": false,
|
| 310 |
"group_num": 1,
|
| 311 |
"group_size": 4096,
|
| 312 |
"in_features": 4096,
|
|
|
|
| 330 |
"model.layers.1.self_attn.q_proj": {
|
| 331 |
"bias": false,
|
| 332 |
"enable_norm": true,
|
| 333 |
+
"enable_perm": false,
|
| 334 |
"group_num": 1,
|
| 335 |
"group_size": 4096,
|
| 336 |
"in_features": 4096,
|
|
|
|
| 354 |
"model.layers.1.self_attn.v_proj": {
|
| 355 |
"bias": false,
|
| 356 |
"enable_norm": true,
|
| 357 |
+
"enable_perm": false,
|
| 358 |
"group_num": 1,
|
| 359 |
"group_size": 4096,
|
| 360 |
"in_features": 4096,
|
|
|
|
| 378 |
"model.layers.10.mlp.down_proj": {
|
| 379 |
"bias": false,
|
| 380 |
"enable_norm": true,
|
| 381 |
+
"enable_perm": false,
|
| 382 |
"group_num": 1,
|
| 383 |
"group_size": 14336,
|
| 384 |
"in_features": 14336,
|
|
|
|
| 402 |
"model.layers.10.mlp.gate_proj": {
|
| 403 |
"bias": false,
|
| 404 |
"enable_norm": true,
|
| 405 |
+
"enable_perm": false,
|
| 406 |
"group_num": 1,
|
| 407 |
"group_size": 4096,
|
| 408 |
"in_features": 4096,
|
|
|
|
| 426 |
"model.layers.10.mlp.up_proj": {
|
| 427 |
"bias": false,
|
| 428 |
"enable_norm": true,
|
| 429 |
+
"enable_perm": false,
|
| 430 |
"group_num": 1,
|
| 431 |
"group_size": 4096,
|
| 432 |
"in_features": 4096,
|
|
|
|
| 450 |
"model.layers.10.self_attn.k_proj": {
|
| 451 |
"bias": false,
|
| 452 |
"enable_norm": true,
|
| 453 |
+
"enable_perm": false,
|
| 454 |
"group_num": 1,
|
| 455 |
"group_size": 4096,
|
| 456 |
"in_features": 4096,
|
|
|
|
| 474 |
"model.layers.10.self_attn.o_proj": {
|
| 475 |
"bias": false,
|
| 476 |
"enable_norm": true,
|
| 477 |
+
"enable_perm": false,
|
| 478 |
"group_num": 1,
|
| 479 |
"group_size": 4096,
|
| 480 |
"in_features": 4096,
|
|
|
|
| 498 |
"model.layers.10.self_attn.q_proj": {
|
| 499 |
"bias": false,
|
| 500 |
"enable_norm": true,
|
| 501 |
+
"enable_perm": false,
|
| 502 |
"group_num": 1,
|
| 503 |
"group_size": 4096,
|
| 504 |
"in_features": 4096,
|
|
|
|
| 522 |
"model.layers.10.self_attn.v_proj": {
|
| 523 |
"bias": false,
|
| 524 |
"enable_norm": true,
|
| 525 |
+
"enable_perm": false,
|
| 526 |
"group_num": 1,
|
| 527 |
"group_size": 4096,
|
| 528 |
"in_features": 4096,
|
|
|
|
| 546 |
"model.layers.11.mlp.down_proj": {
|
| 547 |
"bias": false,
|
| 548 |
"enable_norm": true,
|
| 549 |
+
"enable_perm": false,
|
| 550 |
"group_num": 1,
|
| 551 |
"group_size": 14336,
|
| 552 |
"in_features": 14336,
|
|
|
|
| 570 |
"model.layers.11.mlp.gate_proj": {
|
| 571 |
"bias": false,
|
| 572 |
"enable_norm": true,
|
| 573 |
+
"enable_perm": false,
|
| 574 |
"group_num": 1,
|
| 575 |
"group_size": 4096,
|
| 576 |
"in_features": 4096,
|
|
|
|
| 594 |
"model.layers.11.mlp.up_proj": {
|
| 595 |
"bias": false,
|
| 596 |
"enable_norm": true,
|
| 597 |
+
"enable_perm": false,
|
| 598 |
"group_num": 1,
|
| 599 |
"group_size": 4096,
|
| 600 |
"in_features": 4096,
|
|
|
|
| 618 |
"model.layers.11.self_attn.k_proj": {
|
| 619 |
"bias": false,
|
| 620 |
"enable_norm": true,
|
| 621 |
+
"enable_perm": false,
|
| 622 |
"group_num": 1,
|
| 623 |
"group_size": 4096,
|
| 624 |
"in_features": 4096,
|
|
|
|
| 642 |
"model.layers.11.self_attn.o_proj": {
|
| 643 |
"bias": false,
|
| 644 |
"enable_norm": true,
|
| 645 |
+
"enable_perm": false,
|
| 646 |
"group_num": 1,
|
| 647 |
"group_size": 4096,
|
| 648 |
"in_features": 4096,
|
|
|
|
| 666 |
"model.layers.11.self_attn.q_proj": {
|
| 667 |
"bias": false,
|
| 668 |
"enable_norm": true,
|
| 669 |
+
"enable_perm": false,
|
| 670 |
"group_num": 1,
|
| 671 |
"group_size": 4096,
|
| 672 |
"in_features": 4096,
|
|
|
|
| 690 |
"model.layers.11.self_attn.v_proj": {
|
| 691 |
"bias": false,
|
| 692 |
"enable_norm": true,
|
| 693 |
+
"enable_perm": false,
|
| 694 |
"group_num": 1,
|
| 695 |
"group_size": 4096,
|
| 696 |
"in_features": 4096,
|
|
|
|
| 714 |
"model.layers.12.mlp.down_proj": {
|
| 715 |
"bias": false,
|
| 716 |
"enable_norm": true,
|
| 717 |
+
"enable_perm": false,
|
| 718 |
"group_num": 1,
|
| 719 |
"group_size": 14336,
|
| 720 |
"in_features": 14336,
|
|
|
|
| 738 |
"model.layers.12.mlp.gate_proj": {
|
| 739 |
"bias": false,
|
| 740 |
"enable_norm": true,
|
| 741 |
+
"enable_perm": false,
|
| 742 |
"group_num": 1,
|
| 743 |
"group_size": 4096,
|
| 744 |
"in_features": 4096,
|
|
|
|
| 762 |
"model.layers.12.mlp.up_proj": {
|
| 763 |
"bias": false,
|
| 764 |
"enable_norm": true,
|
| 765 |
+
"enable_perm": false,
|
| 766 |
"group_num": 1,
|
| 767 |
"group_size": 4096,
|
| 768 |
"in_features": 4096,
|
|
|
|
| 786 |
"model.layers.12.self_attn.k_proj": {
|
| 787 |
"bias": false,
|
| 788 |
"enable_norm": true,
|
| 789 |
+
"enable_perm": false,
|
| 790 |
"group_num": 1,
|
| 791 |
"group_size": 4096,
|
| 792 |
"in_features": 4096,
|
|
|
|
| 810 |
"model.layers.12.self_attn.o_proj": {
|
| 811 |
"bias": false,
|
| 812 |
"enable_norm": true,
|
| 813 |
+
"enable_perm": false,
|
| 814 |
"group_num": 1,
|
| 815 |
"group_size": 4096,
|
| 816 |
"in_features": 4096,
|
|
|
|
| 834 |
"model.layers.12.self_attn.q_proj": {
|
| 835 |
"bias": false,
|
| 836 |
"enable_norm": true,
|
| 837 |
+
"enable_perm": false,
|
| 838 |
"group_num": 1,
|
| 839 |
"group_size": 4096,
|
| 840 |
"in_features": 4096,
|
|
|
|
| 858 |
"model.layers.12.self_attn.v_proj": {
|
| 859 |
"bias": false,
|
| 860 |
"enable_norm": true,
|
| 861 |
+
"enable_perm": false,
|
| 862 |
"group_num": 1,
|
| 863 |
"group_size": 4096,
|
| 864 |
"in_features": 4096,
|
|
|
|
| 882 |
"model.layers.13.mlp.down_proj": {
|
| 883 |
"bias": false,
|
| 884 |
"enable_norm": true,
|
| 885 |
+
"enable_perm": false,
|
| 886 |
"group_num": 1,
|
| 887 |
"group_size": 14336,
|
| 888 |
"in_features": 14336,
|
|
|
|
| 906 |
"model.layers.13.mlp.gate_proj": {
|
| 907 |
"bias": false,
|
| 908 |
"enable_norm": true,
|
| 909 |
+
"enable_perm": false,
|
| 910 |
"group_num": 1,
|
| 911 |
"group_size": 4096,
|
| 912 |
"in_features": 4096,
|
|
|
|
| 930 |
"model.layers.13.mlp.up_proj": {
|
| 931 |
"bias": false,
|
| 932 |
"enable_norm": true,
|
| 933 |
+
"enable_perm": false,
|
| 934 |
"group_num": 1,
|
| 935 |
"group_size": 4096,
|
| 936 |
"in_features": 4096,
|
|
|
|
| 954 |
"model.layers.13.self_attn.k_proj": {
|
| 955 |
"bias": false,
|
| 956 |
"enable_norm": true,
|
| 957 |
+
"enable_perm": false,
|
| 958 |
"group_num": 1,
|
| 959 |
"group_size": 4096,
|
| 960 |
"in_features": 4096,
|
|
|
|
| 978 |
"model.layers.13.self_attn.o_proj": {
|
| 979 |
"bias": false,
|
| 980 |
"enable_norm": true,
|
| 981 |
+
"enable_perm": false,
|
| 982 |
"group_num": 1,
|
| 983 |
"group_size": 4096,
|
| 984 |
"in_features": 4096,
|
|
|
|
| 1002 |
"model.layers.13.self_attn.q_proj": {
|
| 1003 |
"bias": false,
|
| 1004 |
"enable_norm": true,
|
| 1005 |
+
"enable_perm": false,
|
| 1006 |
"group_num": 1,
|
| 1007 |
"group_size": 4096,
|
| 1008 |
"in_features": 4096,
|
|
|
|
| 1026 |
"model.layers.13.self_attn.v_proj": {
|
| 1027 |
"bias": false,
|
| 1028 |
"enable_norm": true,
|
| 1029 |
+
"enable_perm": false,
|
| 1030 |
"group_num": 1,
|
| 1031 |
"group_size": 4096,
|
| 1032 |
"in_features": 4096,
|
|
|
|
| 1050 |
"model.layers.14.mlp.down_proj": {
|
| 1051 |
"bias": false,
|
| 1052 |
"enable_norm": true,
|
| 1053 |
+
"enable_perm": false,
|
| 1054 |
"group_num": 1,
|
| 1055 |
"group_size": 14336,
|
| 1056 |
"in_features": 14336,
|
|
|
|
| 1074 |
"model.layers.14.mlp.gate_proj": {
|
| 1075 |
"bias": false,
|
| 1076 |
"enable_norm": true,
|
| 1077 |
+
"enable_perm": false,
|
| 1078 |
"group_num": 1,
|
| 1079 |
"group_size": 4096,
|
| 1080 |
"in_features": 4096,
|
|
|
|
| 1098 |
"model.layers.14.mlp.up_proj": {
|
| 1099 |
"bias": false,
|
| 1100 |
"enable_norm": true,
|
| 1101 |
+
"enable_perm": false,
|
| 1102 |
"group_num": 1,
|
| 1103 |
"group_size": 4096,
|
| 1104 |
"in_features": 4096,
|
|
|
|
| 1122 |
"model.layers.14.self_attn.k_proj": {
|
| 1123 |
"bias": false,
|
| 1124 |
"enable_norm": true,
|
| 1125 |
+
"enable_perm": false,
|
| 1126 |
"group_num": 1,
|
| 1127 |
"group_size": 4096,
|
| 1128 |
"in_features": 4096,
|
|
|
|
| 1146 |
"model.layers.14.self_attn.o_proj": {
|
| 1147 |
"bias": false,
|
| 1148 |
"enable_norm": true,
|
| 1149 |
+
"enable_perm": false,
|
| 1150 |
"group_num": 1,
|
| 1151 |
"group_size": 4096,
|
| 1152 |
"in_features": 4096,
|
|
|
|
| 1170 |
"model.layers.14.self_attn.q_proj": {
|
| 1171 |
"bias": false,
|
| 1172 |
"enable_norm": true,
|
| 1173 |
+
"enable_perm": false,
|
| 1174 |
"group_num": 1,
|
| 1175 |
"group_size": 4096,
|
| 1176 |
"in_features": 4096,
|
|
|
|
| 1194 |
"model.layers.14.self_attn.v_proj": {
|
| 1195 |
"bias": false,
|
| 1196 |
"enable_norm": true,
|
| 1197 |
+
"enable_perm": false,
|
| 1198 |
"group_num": 1,
|
| 1199 |
"group_size": 4096,
|
| 1200 |
"in_features": 4096,
|
|
|
|
| 1218 |
"model.layers.15.mlp.down_proj": {
|
| 1219 |
"bias": false,
|
| 1220 |
"enable_norm": true,
|
| 1221 |
+
"enable_perm": false,
|
| 1222 |
"group_num": 1,
|
| 1223 |
"group_size": 14336,
|
| 1224 |
"in_features": 14336,
|
|
|
|
| 1242 |
"model.layers.15.mlp.gate_proj": {
|
| 1243 |
"bias": false,
|
| 1244 |
"enable_norm": true,
|
| 1245 |
+
"enable_perm": false,
|
| 1246 |
"group_num": 1,
|
| 1247 |
"group_size": 4096,
|
| 1248 |
"in_features": 4096,
|
|
|
|
| 1266 |
"model.layers.15.mlp.up_proj": {
|
| 1267 |
"bias": false,
|
| 1268 |
"enable_norm": true,
|
| 1269 |
+
"enable_perm": false,
|
| 1270 |
"group_num": 1,
|
| 1271 |
"group_size": 4096,
|
| 1272 |
"in_features": 4096,
|
|
|
|
| 1290 |
"model.layers.15.self_attn.k_proj": {
|
| 1291 |
"bias": false,
|
| 1292 |
"enable_norm": true,
|
| 1293 |
+
"enable_perm": false,
|
| 1294 |
"group_num": 1,
|
| 1295 |
"group_size": 4096,
|
| 1296 |
"in_features": 4096,
|
|
|
|
| 1314 |
"model.layers.15.self_attn.o_proj": {
|
| 1315 |
"bias": false,
|
| 1316 |
"enable_norm": true,
|
| 1317 |
+
"enable_perm": false,
|
| 1318 |
"group_num": 1,
|
| 1319 |
"group_size": 4096,
|
| 1320 |
"in_features": 4096,
|
|
|
|
| 1338 |
"model.layers.15.self_attn.q_proj": {
|
| 1339 |
"bias": false,
|
| 1340 |
"enable_norm": true,
|
| 1341 |
+
"enable_perm": false,
|
| 1342 |
"group_num": 1,
|
| 1343 |
"group_size": 4096,
|
| 1344 |
"in_features": 4096,
|
|
|
|
| 1362 |
"model.layers.15.self_attn.v_proj": {
|
| 1363 |
"bias": false,
|
| 1364 |
"enable_norm": true,
|
| 1365 |
+
"enable_perm": false,
|
| 1366 |
"group_num": 1,
|
| 1367 |
"group_size": 4096,
|
| 1368 |
"in_features": 4096,
|
|
|
|
| 1386 |
"model.layers.16.mlp.down_proj": {
|
| 1387 |
"bias": false,
|
| 1388 |
"enable_norm": true,
|
| 1389 |
+
"enable_perm": false,
|
| 1390 |
"group_num": 1,
|
| 1391 |
"group_size": 14336,
|
| 1392 |
"in_features": 14336,
|
|
|
|
| 1410 |
"model.layers.16.mlp.gate_proj": {
|
| 1411 |
"bias": false,
|
| 1412 |
"enable_norm": true,
|
| 1413 |
+
"enable_perm": false,
|
| 1414 |
"group_num": 1,
|
| 1415 |
"group_size": 4096,
|
| 1416 |
"in_features": 4096,
|
|
|
|
| 1434 |
"model.layers.16.mlp.up_proj": {
|
| 1435 |
"bias": false,
|
| 1436 |
"enable_norm": true,
|
| 1437 |
+
"enable_perm": false,
|
| 1438 |
"group_num": 1,
|
| 1439 |
"group_size": 4096,
|
| 1440 |
"in_features": 4096,
|
|
|
|
| 1458 |
"model.layers.16.self_attn.k_proj": {
|
| 1459 |
"bias": false,
|
| 1460 |
"enable_norm": true,
|
| 1461 |
+
"enable_perm": false,
|
| 1462 |
"group_num": 1,
|
| 1463 |
"group_size": 4096,
|
| 1464 |
"in_features": 4096,
|
|
|
|
| 1482 |
"model.layers.16.self_attn.o_proj": {
|
| 1483 |
"bias": false,
|
| 1484 |
"enable_norm": true,
|
| 1485 |
+
"enable_perm": false,
|
| 1486 |
"group_num": 1,
|
| 1487 |
"group_size": 4096,
|
| 1488 |
"in_features": 4096,
|
|
|
|
| 1506 |
"model.layers.16.self_attn.q_proj": {
|
| 1507 |
"bias": false,
|
| 1508 |
"enable_norm": true,
|
| 1509 |
+
"enable_perm": false,
|
| 1510 |
"group_num": 1,
|
| 1511 |
"group_size": 4096,
|
| 1512 |
"in_features": 4096,
|
|
|
|
| 1530 |
"model.layers.16.self_attn.v_proj": {
|
| 1531 |
"bias": false,
|
| 1532 |
"enable_norm": true,
|
| 1533 |
+
"enable_perm": false,
|
| 1534 |
"group_num": 1,
|
| 1535 |
"group_size": 4096,
|
| 1536 |
"in_features": 4096,
|
|
|
|
| 1554 |
"model.layers.17.mlp.down_proj": {
|
| 1555 |
"bias": false,
|
| 1556 |
"enable_norm": true,
|
| 1557 |
+
"enable_perm": false,
|
| 1558 |
"group_num": 1,
|
| 1559 |
"group_size": 14336,
|
| 1560 |
"in_features": 14336,
|
|
|
|
| 1578 |
"model.layers.17.mlp.gate_proj": {
|
| 1579 |
"bias": false,
|
| 1580 |
"enable_norm": true,
|
| 1581 |
+
"enable_perm": false,
|
| 1582 |
"group_num": 1,
|
| 1583 |
"group_size": 4096,
|
| 1584 |
"in_features": 4096,
|
|
|
|
| 1602 |
"model.layers.17.mlp.up_proj": {
|
| 1603 |
"bias": false,
|
| 1604 |
"enable_norm": true,
|
| 1605 |
+
"enable_perm": false,
|
| 1606 |
"group_num": 1,
|
| 1607 |
"group_size": 4096,
|
| 1608 |
"in_features": 4096,
|
|
|
|
| 1626 |
"model.layers.17.self_attn.k_proj": {
|
| 1627 |
"bias": false,
|
| 1628 |
"enable_norm": true,
|
| 1629 |
+
"enable_perm": false,
|
| 1630 |
"group_num": 1,
|
| 1631 |
"group_size": 4096,
|
| 1632 |
"in_features": 4096,
|
|
|
|
| 1650 |
"model.layers.17.self_attn.o_proj": {
|
| 1651 |
"bias": false,
|
| 1652 |
"enable_norm": true,
|
| 1653 |
+
"enable_perm": false,
|
| 1654 |
"group_num": 1,
|
| 1655 |
"group_size": 4096,
|
| 1656 |
"in_features": 4096,
|
|
|
|
| 1674 |
"model.layers.17.self_attn.q_proj": {
|
| 1675 |
"bias": false,
|
| 1676 |
"enable_norm": true,
|
| 1677 |
+
"enable_perm": false,
|
| 1678 |
"group_num": 1,
|
| 1679 |
"group_size": 4096,
|
| 1680 |
"in_features": 4096,
|
|
|
|
| 1698 |
"model.layers.17.self_attn.v_proj": {
|
| 1699 |
"bias": false,
|
| 1700 |
"enable_norm": true,
|
| 1701 |
+
"enable_perm": false,
|
| 1702 |
"group_num": 1,
|
| 1703 |
"group_size": 4096,
|
| 1704 |
"in_features": 4096,
|
|
|
|
| 1722 |
"model.layers.18.mlp.down_proj": {
|
| 1723 |
"bias": false,
|
| 1724 |
"enable_norm": true,
|
| 1725 |
+
"enable_perm": false,
|
| 1726 |
"group_num": 1,
|
| 1727 |
"group_size": 14336,
|
| 1728 |
"in_features": 14336,
|
|
|
|
| 1746 |
"model.layers.18.mlp.gate_proj": {
|
| 1747 |
"bias": false,
|
| 1748 |
"enable_norm": true,
|
| 1749 |
+
"enable_perm": false,
|
| 1750 |
"group_num": 1,
|
| 1751 |
"group_size": 4096,
|
| 1752 |
"in_features": 4096,
|
|
|
|
| 1770 |
"model.layers.18.mlp.up_proj": {
|
| 1771 |
"bias": false,
|
| 1772 |
"enable_norm": true,
|
| 1773 |
+
"enable_perm": false,
|
| 1774 |
"group_num": 1,
|
| 1775 |
"group_size": 4096,
|
| 1776 |
"in_features": 4096,
|
|
|
|
| 1794 |
"model.layers.18.self_attn.k_proj": {
|
| 1795 |
"bias": false,
|
| 1796 |
"enable_norm": true,
|
| 1797 |
+
"enable_perm": false,
|
| 1798 |
"group_num": 1,
|
| 1799 |
"group_size": 4096,
|
| 1800 |
"in_features": 4096,
|
|
|
|
| 1818 |
"model.layers.18.self_attn.o_proj": {
|
| 1819 |
"bias": false,
|
| 1820 |
"enable_norm": true,
|
| 1821 |
+
"enable_perm": false,
|
| 1822 |
"group_num": 1,
|
| 1823 |
"group_size": 4096,
|
| 1824 |
"in_features": 4096,
|
|
|
|
| 1842 |
"model.layers.18.self_attn.q_proj": {
|
| 1843 |
"bias": false,
|
| 1844 |
"enable_norm": true,
|
| 1845 |
+
"enable_perm": false,
|
| 1846 |
"group_num": 1,
|
| 1847 |
"group_size": 4096,
|
| 1848 |
"in_features": 4096,
|
|
|
|
| 1866 |
"model.layers.18.self_attn.v_proj": {
|
| 1867 |
"bias": false,
|
| 1868 |
"enable_norm": true,
|
| 1869 |
+
"enable_perm": false,
|
| 1870 |
"group_num": 1,
|
| 1871 |
"group_size": 4096,
|
| 1872 |
"in_features": 4096,
|
|
|
|
| 1890 |
"model.layers.19.mlp.down_proj": {
|
| 1891 |
"bias": false,
|
| 1892 |
"enable_norm": true,
|
| 1893 |
+
"enable_perm": false,
|
| 1894 |
"group_num": 1,
|
| 1895 |
"group_size": 14336,
|
| 1896 |
"in_features": 14336,
|
|
|
|
| 1914 |
"model.layers.19.mlp.gate_proj": {
|
| 1915 |
"bias": false,
|
| 1916 |
"enable_norm": true,
|
| 1917 |
+
"enable_perm": false,
|
| 1918 |
"group_num": 1,
|
| 1919 |
"group_size": 4096,
|
| 1920 |
"in_features": 4096,
|
|
|
|
| 1938 |
"model.layers.19.mlp.up_proj": {
|
| 1939 |
"bias": false,
|
| 1940 |
"enable_norm": true,
|
| 1941 |
+
"enable_perm": false,
|
| 1942 |
"group_num": 1,
|
| 1943 |
"group_size": 4096,
|
| 1944 |
"in_features": 4096,
|
|
|
|
| 1962 |
"model.layers.19.self_attn.k_proj": {
|
| 1963 |
"bias": false,
|
| 1964 |
"enable_norm": true,
|
| 1965 |
+
"enable_perm": false,
|
| 1966 |
"group_num": 1,
|
| 1967 |
"group_size": 4096,
|
| 1968 |
"in_features": 4096,
|
|
|
|
| 1986 |
"model.layers.19.self_attn.o_proj": {
|
| 1987 |
"bias": false,
|
| 1988 |
"enable_norm": true,
|
| 1989 |
+
"enable_perm": false,
|
| 1990 |
"group_num": 1,
|
| 1991 |
"group_size": 4096,
|
| 1992 |
"in_features": 4096,
|
|
|
|
| 2010 |
"model.layers.19.self_attn.q_proj": {
|
| 2011 |
"bias": false,
|
| 2012 |
"enable_norm": true,
|
| 2013 |
+
"enable_perm": false,
|
| 2014 |
"group_num": 1,
|
| 2015 |
"group_size": 4096,
|
| 2016 |
"in_features": 4096,
|
|
|
|
| 2034 |
"model.layers.19.self_attn.v_proj": {
|
| 2035 |
"bias": false,
|
| 2036 |
"enable_norm": true,
|
| 2037 |
+
"enable_perm": false,
|
| 2038 |
"group_num": 1,
|
| 2039 |
"group_size": 4096,
|
| 2040 |
"in_features": 4096,
|
|
|
|
| 2058 |
"model.layers.2.mlp.down_proj": {
|
| 2059 |
"bias": false,
|
| 2060 |
"enable_norm": true,
|
| 2061 |
+
"enable_perm": false,
|
| 2062 |
"group_num": 1,
|
| 2063 |
"group_size": 14336,
|
| 2064 |
"in_features": 14336,
|
|
|
|
| 2082 |
"model.layers.2.mlp.gate_proj": {
|
| 2083 |
"bias": false,
|
| 2084 |
"enable_norm": true,
|
| 2085 |
+
"enable_perm": false,
|
| 2086 |
"group_num": 1,
|
| 2087 |
"group_size": 4096,
|
| 2088 |
"in_features": 4096,
|
|
|
|
| 2106 |
"model.layers.2.mlp.up_proj": {
|
| 2107 |
"bias": false,
|
| 2108 |
"enable_norm": true,
|
| 2109 |
+
"enable_perm": false,
|
| 2110 |
"group_num": 1,
|
| 2111 |
"group_size": 4096,
|
| 2112 |
"in_features": 4096,
|
|
|
|
| 2130 |
"model.layers.2.self_attn.k_proj": {
|
| 2131 |
"bias": false,
|
| 2132 |
"enable_norm": true,
|
| 2133 |
+
"enable_perm": false,
|
| 2134 |
"group_num": 1,
|
| 2135 |
"group_size": 4096,
|
| 2136 |
"in_features": 4096,
|
|
|
|
| 2154 |
"model.layers.2.self_attn.o_proj": {
|
| 2155 |
"bias": false,
|
| 2156 |
"enable_norm": true,
|
| 2157 |
+
"enable_perm": false,
|
| 2158 |
"group_num": 1,
|
| 2159 |
"group_size": 4096,
|
| 2160 |
"in_features": 4096,
|
|
|
|
| 2178 |
"model.layers.2.self_attn.q_proj": {
|
| 2179 |
"bias": false,
|
| 2180 |
"enable_norm": true,
|
| 2181 |
+
"enable_perm": false,
|
| 2182 |
"group_num": 1,
|
| 2183 |
"group_size": 4096,
|
| 2184 |
"in_features": 4096,
|
|
|
|
| 2202 |
"model.layers.2.self_attn.v_proj": {
|
| 2203 |
"bias": false,
|
| 2204 |
"enable_norm": true,
|
| 2205 |
+
"enable_perm": false,
|
| 2206 |
"group_num": 1,
|
| 2207 |
"group_size": 4096,
|
| 2208 |
"in_features": 4096,
|
|
|
|
| 2226 |
"model.layers.20.mlp.down_proj": {
|
| 2227 |
"bias": false,
|
| 2228 |
"enable_norm": true,
|
| 2229 |
+
"enable_perm": false,
|
| 2230 |
"group_num": 1,
|
| 2231 |
"group_size": 14336,
|
| 2232 |
"in_features": 14336,
|
|
|
|
| 2250 |
"model.layers.20.mlp.gate_proj": {
|
| 2251 |
"bias": false,
|
| 2252 |
"enable_norm": true,
|
| 2253 |
+
"enable_perm": false,
|
| 2254 |
"group_num": 1,
|
| 2255 |
"group_size": 4096,
|
| 2256 |
"in_features": 4096,
|
|
|
|
| 2274 |
"model.layers.20.mlp.up_proj": {
|
| 2275 |
"bias": false,
|
| 2276 |
"enable_norm": true,
|
| 2277 |
+
"enable_perm": false,
|
| 2278 |
"group_num": 1,
|
| 2279 |
"group_size": 4096,
|
| 2280 |
"in_features": 4096,
|
|
|
|
| 2298 |
"model.layers.20.self_attn.k_proj": {
|
| 2299 |
"bias": false,
|
| 2300 |
"enable_norm": true,
|
| 2301 |
+
"enable_perm": false,
|
| 2302 |
"group_num": 1,
|
| 2303 |
"group_size": 4096,
|
| 2304 |
"in_features": 4096,
|
|
|
|
| 2322 |
"model.layers.20.self_attn.o_proj": {
|
| 2323 |
"bias": false,
|
| 2324 |
"enable_norm": true,
|
| 2325 |
+
"enable_perm": false,
|
| 2326 |
"group_num": 1,
|
| 2327 |
"group_size": 4096,
|
| 2328 |
"in_features": 4096,
|
|
|
|
| 2346 |
"model.layers.20.self_attn.q_proj": {
|
| 2347 |
"bias": false,
|
| 2348 |
"enable_norm": true,
|
| 2349 |
+
"enable_perm": false,
|
| 2350 |
"group_num": 1,
|
| 2351 |
"group_size": 4096,
|
| 2352 |
"in_features": 4096,
|
|
|
|
| 2370 |
"model.layers.20.self_attn.v_proj": {
|
| 2371 |
"bias": false,
|
| 2372 |
"enable_norm": true,
|
| 2373 |
+
"enable_perm": false,
|
| 2374 |
"group_num": 1,
|
| 2375 |
"group_size": 4096,
|
| 2376 |
"in_features": 4096,
|
|
|
|
| 2394 |
"model.layers.21.mlp.down_proj": {
|
| 2395 |
"bias": false,
|
| 2396 |
"enable_norm": true,
|
| 2397 |
+
"enable_perm": false,
|
| 2398 |
"group_num": 1,
|
| 2399 |
"group_size": 14336,
|
| 2400 |
"in_features": 14336,
|
|
|
|
| 2418 |
"model.layers.21.mlp.gate_proj": {
|
| 2419 |
"bias": false,
|
| 2420 |
"enable_norm": true,
|
| 2421 |
+
"enable_perm": false,
|
| 2422 |
"group_num": 1,
|
| 2423 |
"group_size": 4096,
|
| 2424 |
"in_features": 4096,
|
|
|
|
| 2442 |
"model.layers.21.mlp.up_proj": {
|
| 2443 |
"bias": false,
|
| 2444 |
"enable_norm": true,
|
| 2445 |
+
"enable_perm": false,
|
| 2446 |
"group_num": 1,
|
| 2447 |
"group_size": 4096,
|
| 2448 |
"in_features": 4096,
|
|
|
|
| 2466 |
"model.layers.21.self_attn.k_proj": {
|
| 2467 |
"bias": false,
|
| 2468 |
"enable_norm": true,
|
| 2469 |
+
"enable_perm": false,
|
| 2470 |
"group_num": 1,
|
| 2471 |
"group_size": 4096,
|
| 2472 |
"in_features": 4096,
|
|
|
|
| 2490 |
"model.layers.21.self_attn.o_proj": {
|
| 2491 |
"bias": false,
|
| 2492 |
"enable_norm": true,
|
| 2493 |
+
"enable_perm": false,
|
| 2494 |
"group_num": 1,
|
| 2495 |
"group_size": 4096,
|
| 2496 |
"in_features": 4096,
|
|
|
|
| 2514 |
"model.layers.21.self_attn.q_proj": {
|
| 2515 |
"bias": false,
|
| 2516 |
"enable_norm": true,
|
| 2517 |
+
"enable_perm": false,
|
| 2518 |
"group_num": 1,
|
| 2519 |
"group_size": 4096,
|
| 2520 |
"in_features": 4096,
|
|
|
|
| 2538 |
"model.layers.21.self_attn.v_proj": {
|
| 2539 |
"bias": false,
|
| 2540 |
"enable_norm": true,
|
| 2541 |
+
"enable_perm": false,
|
| 2542 |
"group_num": 1,
|
| 2543 |
"group_size": 4096,
|
| 2544 |
"in_features": 4096,
|
|
|
|
| 2562 |
"model.layers.22.mlp.down_proj": {
|
| 2563 |
"bias": false,
|
| 2564 |
"enable_norm": true,
|
| 2565 |
+
"enable_perm": false,
|
| 2566 |
"group_num": 1,
|
| 2567 |
"group_size": 14336,
|
| 2568 |
"in_features": 14336,
|
|
|
|
| 2586 |
"model.layers.22.mlp.gate_proj": {
|
| 2587 |
"bias": false,
|
| 2588 |
"enable_norm": true,
|
| 2589 |
+
"enable_perm": false,
|
| 2590 |
"group_num": 1,
|
| 2591 |
"group_size": 4096,
|
| 2592 |
"in_features": 4096,
|
|
|
|
| 2610 |
"model.layers.22.mlp.up_proj": {
|
| 2611 |
"bias": false,
|
| 2612 |
"enable_norm": true,
|
| 2613 |
+
"enable_perm": false,
|
| 2614 |
"group_num": 1,
|
| 2615 |
"group_size": 4096,
|
| 2616 |
"in_features": 4096,
|
|
|
|
| 2634 |
"model.layers.22.self_attn.k_proj": {
|
| 2635 |
"bias": false,
|
| 2636 |
"enable_norm": true,
|
| 2637 |
+
"enable_perm": false,
|
| 2638 |
"group_num": 1,
|
| 2639 |
"group_size": 4096,
|
| 2640 |
"in_features": 4096,
|
|
|
|
| 2658 |
"model.layers.22.self_attn.o_proj": {
|
| 2659 |
"bias": false,
|
| 2660 |
"enable_norm": true,
|
| 2661 |
+
"enable_perm": false,
|
| 2662 |
"group_num": 1,
|
| 2663 |
"group_size": 4096,
|
| 2664 |
"in_features": 4096,
|
|
|
|
| 2682 |
"model.layers.22.self_attn.q_proj": {
|
| 2683 |
"bias": false,
|
| 2684 |
"enable_norm": true,
|
| 2685 |
+
"enable_perm": false,
|
| 2686 |
"group_num": 1,
|
| 2687 |
"group_size": 4096,
|
| 2688 |
"in_features": 4096,
|
|
|
|
| 2706 |
"model.layers.22.self_attn.v_proj": {
|
| 2707 |
"bias": false,
|
| 2708 |
"enable_norm": true,
|
| 2709 |
+
"enable_perm": false,
|
| 2710 |
"group_num": 1,
|
| 2711 |
"group_size": 4096,
|
| 2712 |
"in_features": 4096,
|
|
|
|
| 2730 |
"model.layers.23.mlp.down_proj": {
|
| 2731 |
"bias": false,
|
| 2732 |
"enable_norm": true,
|
| 2733 |
+
"enable_perm": false,
|
| 2734 |
"group_num": 1,
|
| 2735 |
"group_size": 14336,
|
| 2736 |
"in_features": 14336,
|
|
|
|
| 2754 |
"model.layers.23.mlp.gate_proj": {
|
| 2755 |
"bias": false,
|
| 2756 |
"enable_norm": true,
|
| 2757 |
+
"enable_perm": false,
|
| 2758 |
"group_num": 1,
|
| 2759 |
"group_size": 4096,
|
| 2760 |
"in_features": 4096,
|
|
|
|
| 2778 |
"model.layers.23.mlp.up_proj": {
|
| 2779 |
"bias": false,
|
| 2780 |
"enable_norm": true,
|
| 2781 |
+
"enable_perm": false,
|
| 2782 |
"group_num": 1,
|
| 2783 |
"group_size": 4096,
|
| 2784 |
"in_features": 4096,
|
|
|
|
| 2802 |
"model.layers.23.self_attn.k_proj": {
|
| 2803 |
"bias": false,
|
| 2804 |
"enable_norm": true,
|
| 2805 |
+
"enable_perm": false,
|
| 2806 |
"group_num": 1,
|
| 2807 |
"group_size": 4096,
|
| 2808 |
"in_features": 4096,
|
|
|
|
| 2826 |
"model.layers.23.self_attn.o_proj": {
|
| 2827 |
"bias": false,
|
| 2828 |
"enable_norm": true,
|
| 2829 |
+
"enable_perm": false,
|
| 2830 |
"group_num": 1,
|
| 2831 |
"group_size": 4096,
|
| 2832 |
"in_features": 4096,
|
|
|
|
| 2850 |
"model.layers.23.self_attn.q_proj": {
|
| 2851 |
"bias": false,
|
| 2852 |
"enable_norm": true,
|
| 2853 |
+
"enable_perm": false,
|
| 2854 |
"group_num": 1,
|
| 2855 |
"group_size": 4096,
|
| 2856 |
"in_features": 4096,
|
|
|
|
| 2874 |
"model.layers.23.self_attn.v_proj": {
|
| 2875 |
"bias": false,
|
| 2876 |
"enable_norm": true,
|
| 2877 |
+
"enable_perm": false,
|
| 2878 |
"group_num": 1,
|
| 2879 |
"group_size": 4096,
|
| 2880 |
"in_features": 4096,
|
|
|
|
| 2898 |
"model.layers.24.mlp.down_proj": {
|
| 2899 |
"bias": false,
|
| 2900 |
"enable_norm": true,
|
| 2901 |
+
"enable_perm": false,
|
| 2902 |
"group_num": 1,
|
| 2903 |
"group_size": 14336,
|
| 2904 |
"in_features": 14336,
|
|
|
|
| 2922 |
"model.layers.24.mlp.gate_proj": {
|
| 2923 |
"bias": false,
|
| 2924 |
"enable_norm": true,
|
| 2925 |
+
"enable_perm": false,
|
| 2926 |
"group_num": 1,
|
| 2927 |
"group_size": 4096,
|
| 2928 |
"in_features": 4096,
|
|
|
|
| 2946 |
"model.layers.24.mlp.up_proj": {
|
| 2947 |
"bias": false,
|
| 2948 |
"enable_norm": true,
|
| 2949 |
+
"enable_perm": false,
|
| 2950 |
"group_num": 1,
|
| 2951 |
"group_size": 4096,
|
| 2952 |
"in_features": 4096,
|
|
|
|
| 2970 |
"model.layers.24.self_attn.k_proj": {
|
| 2971 |
"bias": false,
|
| 2972 |
"enable_norm": true,
|
| 2973 |
+
"enable_perm": false,
|
| 2974 |
"group_num": 1,
|
| 2975 |
"group_size": 4096,
|
| 2976 |
"in_features": 4096,
|
|
|
|
| 2994 |
"model.layers.24.self_attn.o_proj": {
|
| 2995 |
"bias": false,
|
| 2996 |
"enable_norm": true,
|
| 2997 |
+
"enable_perm": false,
|
| 2998 |
"group_num": 1,
|
| 2999 |
"group_size": 4096,
|
| 3000 |
"in_features": 4096,
|
|
|
|
| 3018 |
"model.layers.24.self_attn.q_proj": {
|
| 3019 |
"bias": false,
|
| 3020 |
"enable_norm": true,
|
| 3021 |
+
"enable_perm": false,
|
| 3022 |
"group_num": 1,
|
| 3023 |
"group_size": 4096,
|
| 3024 |
"in_features": 4096,
|
|
|
|
| 3042 |
"model.layers.24.self_attn.v_proj": {
|
| 3043 |
"bias": false,
|
| 3044 |
"enable_norm": true,
|
| 3045 |
+
"enable_perm": false,
|
| 3046 |
"group_num": 1,
|
| 3047 |
"group_size": 4096,
|
| 3048 |
"in_features": 4096,
|
|
|
|
| 3066 |
"model.layers.25.mlp.down_proj": {
|
| 3067 |
"bias": false,
|
| 3068 |
"enable_norm": true,
|
| 3069 |
+
"enable_perm": false,
|
| 3070 |
"group_num": 1,
|
| 3071 |
"group_size": 14336,
|
| 3072 |
"in_features": 14336,
|
|
|
|
| 3090 |
"model.layers.25.mlp.gate_proj": {
|
| 3091 |
"bias": false,
|
| 3092 |
"enable_norm": true,
|
| 3093 |
+
"enable_perm": false,
|
| 3094 |
"group_num": 1,
|
| 3095 |
"group_size": 4096,
|
| 3096 |
"in_features": 4096,
|
|
|
|
| 3114 |
"model.layers.25.mlp.up_proj": {
|
| 3115 |
"bias": false,
|
| 3116 |
"enable_norm": true,
|
| 3117 |
+
"enable_perm": false,
|
| 3118 |
"group_num": 1,
|
| 3119 |
"group_size": 4096,
|
| 3120 |
"in_features": 4096,
|
|
|
|
| 3138 |
"model.layers.25.self_attn.k_proj": {
|
| 3139 |
"bias": false,
|
| 3140 |
"enable_norm": true,
|
| 3141 |
+
"enable_perm": false,
|
| 3142 |
"group_num": 1,
|
| 3143 |
"group_size": 4096,
|
| 3144 |
"in_features": 4096,
|
|
|
|
| 3162 |
"model.layers.25.self_attn.o_proj": {
|
| 3163 |
"bias": false,
|
| 3164 |
"enable_norm": true,
|
| 3165 |
+
"enable_perm": false,
|
| 3166 |
"group_num": 1,
|
| 3167 |
"group_size": 4096,
|
| 3168 |
"in_features": 4096,
|
|
|
|
| 3186 |
"model.layers.25.self_attn.q_proj": {
|
| 3187 |
"bias": false,
|
| 3188 |
"enable_norm": true,
|
| 3189 |
+
"enable_perm": false,
|
| 3190 |
"group_num": 1,
|
| 3191 |
"group_size": 4096,
|
| 3192 |
"in_features": 4096,
|
|
|
|
| 3210 |
"model.layers.25.self_attn.v_proj": {
|
| 3211 |
"bias": false,
|
| 3212 |
"enable_norm": true,
|
| 3213 |
+
"enable_perm": false,
|
| 3214 |
"group_num": 1,
|
| 3215 |
"group_size": 4096,
|
| 3216 |
"in_features": 4096,
|
|
|
|
| 3234 |
"model.layers.26.mlp.down_proj": {
|
| 3235 |
"bias": false,
|
| 3236 |
"enable_norm": true,
|
| 3237 |
+
"enable_perm": false,
|
| 3238 |
"group_num": 1,
|
| 3239 |
"group_size": 14336,
|
| 3240 |
"in_features": 14336,
|
|
|
|
| 3258 |
"model.layers.26.mlp.gate_proj": {
|
| 3259 |
"bias": false,
|
| 3260 |
"enable_norm": true,
|
| 3261 |
+
"enable_perm": false,
|
| 3262 |
"group_num": 1,
|
| 3263 |
"group_size": 4096,
|
| 3264 |
"in_features": 4096,
|
|
|
|
| 3282 |
"model.layers.26.mlp.up_proj": {
|
| 3283 |
"bias": false,
|
| 3284 |
"enable_norm": true,
|
| 3285 |
+
"enable_perm": false,
|
| 3286 |
"group_num": 1,
|
| 3287 |
"group_size": 4096,
|
| 3288 |
"in_features": 4096,
|
|
|
|
| 3306 |
"model.layers.26.self_attn.k_proj": {
|
| 3307 |
"bias": false,
|
| 3308 |
"enable_norm": true,
|
| 3309 |
+
"enable_perm": false,
|
| 3310 |
"group_num": 1,
|
| 3311 |
"group_size": 4096,
|
| 3312 |
"in_features": 4096,
|
|
|
|
| 3330 |
"model.layers.26.self_attn.o_proj": {
|
| 3331 |
"bias": false,
|
| 3332 |
"enable_norm": true,
|
| 3333 |
+
"enable_perm": false,
|
| 3334 |
"group_num": 1,
|
| 3335 |
"group_size": 4096,
|
| 3336 |
"in_features": 4096,
|
|
|
|
| 3354 |
"model.layers.26.self_attn.q_proj": {
|
| 3355 |
"bias": false,
|
| 3356 |
"enable_norm": true,
|
| 3357 |
+
"enable_perm": false,
|
| 3358 |
"group_num": 1,
|
| 3359 |
"group_size": 4096,
|
| 3360 |
"in_features": 4096,
|
|
|
|
| 3378 |
"model.layers.26.self_attn.v_proj": {
|
| 3379 |
"bias": false,
|
| 3380 |
"enable_norm": true,
|
| 3381 |
+
"enable_perm": false,
|
| 3382 |
"group_num": 1,
|
| 3383 |
"group_size": 4096,
|
| 3384 |
"in_features": 4096,
|
|
|
|
| 3402 |
"model.layers.27.mlp.down_proj": {
|
| 3403 |
"bias": false,
|
| 3404 |
"enable_norm": true,
|
| 3405 |
+
"enable_perm": false,
|
| 3406 |
"group_num": 1,
|
| 3407 |
"group_size": 14336,
|
| 3408 |
"in_features": 14336,
|
|
|
|
| 3426 |
"model.layers.27.mlp.gate_proj": {
|
| 3427 |
"bias": false,
|
| 3428 |
"enable_norm": true,
|
| 3429 |
+
"enable_perm": false,
|
| 3430 |
"group_num": 1,
|
| 3431 |
"group_size": 4096,
|
| 3432 |
"in_features": 4096,
|
|
|
|
| 3450 |
"model.layers.27.mlp.up_proj": {
|
| 3451 |
"bias": false,
|
| 3452 |
"enable_norm": true,
|
| 3453 |
+
"enable_perm": false,
|
| 3454 |
"group_num": 1,
|
| 3455 |
"group_size": 4096,
|
| 3456 |
"in_features": 4096,
|
|
|
|
| 3474 |
"model.layers.27.self_attn.k_proj": {
|
| 3475 |
"bias": false,
|
| 3476 |
"enable_norm": true,
|
| 3477 |
+
"enable_perm": false,
|
| 3478 |
"group_num": 1,
|
| 3479 |
"group_size": 4096,
|
| 3480 |
"in_features": 4096,
|
|
|
|
| 3498 |
"model.layers.27.self_attn.o_proj": {
|
| 3499 |
"bias": false,
|
| 3500 |
"enable_norm": true,
|
| 3501 |
+
"enable_perm": false,
|
| 3502 |
"group_num": 1,
|
| 3503 |
"group_size": 4096,
|
| 3504 |
"in_features": 4096,
|
|
|
|
| 3522 |
"model.layers.27.self_attn.q_proj": {
|
| 3523 |
"bias": false,
|
| 3524 |
"enable_norm": true,
|
| 3525 |
+
"enable_perm": false,
|
| 3526 |
"group_num": 1,
|
| 3527 |
"group_size": 4096,
|
| 3528 |
"in_features": 4096,
|
|
|
|
| 3546 |
"model.layers.27.self_attn.v_proj": {
|
| 3547 |
"bias": false,
|
| 3548 |
"enable_norm": true,
|
| 3549 |
+
"enable_perm": false,
|
| 3550 |
"group_num": 1,
|
| 3551 |
"group_size": 4096,
|
| 3552 |
"in_features": 4096,
|
|
|
|
| 3570 |
"model.layers.28.mlp.down_proj": {
|
| 3571 |
"bias": false,
|
| 3572 |
"enable_norm": true,
|
| 3573 |
+
"enable_perm": false,
|
| 3574 |
"group_num": 1,
|
| 3575 |
"group_size": 14336,
|
| 3576 |
"in_features": 14336,
|
|
|
|
| 3594 |
"model.layers.28.mlp.gate_proj": {
|
| 3595 |
"bias": false,
|
| 3596 |
"enable_norm": true,
|
| 3597 |
+
"enable_perm": false,
|
| 3598 |
"group_num": 1,
|
| 3599 |
"group_size": 4096,
|
| 3600 |
"in_features": 4096,
|
|
|
|
| 3618 |
"model.layers.28.mlp.up_proj": {
|
| 3619 |
"bias": false,
|
| 3620 |
"enable_norm": true,
|
| 3621 |
+
"enable_perm": false,
|
| 3622 |
"group_num": 1,
|
| 3623 |
"group_size": 4096,
|
| 3624 |
"in_features": 4096,
|
|
|
|
| 3642 |
"model.layers.28.self_attn.k_proj": {
|
| 3643 |
"bias": false,
|
| 3644 |
"enable_norm": true,
|
| 3645 |
+
"enable_perm": false,
|
| 3646 |
"group_num": 1,
|
| 3647 |
"group_size": 4096,
|
| 3648 |
"in_features": 4096,
|
|
|
|
| 3666 |
"model.layers.28.self_attn.o_proj": {
|
| 3667 |
"bias": false,
|
| 3668 |
"enable_norm": true,
|
| 3669 |
+
"enable_perm": false,
|
| 3670 |
"group_num": 1,
|
| 3671 |
"group_size": 4096,
|
| 3672 |
"in_features": 4096,
|
|
|
|
| 3690 |
"model.layers.28.self_attn.q_proj": {
|
| 3691 |
"bias": false,
|
| 3692 |
"enable_norm": true,
|
| 3693 |
+
"enable_perm": false,
|
| 3694 |
"group_num": 1,
|
| 3695 |
"group_size": 4096,
|
| 3696 |
"in_features": 4096,
|
|
|
|
| 3714 |
"model.layers.28.self_attn.v_proj": {
|
| 3715 |
"bias": false,
|
| 3716 |
"enable_norm": true,
|
| 3717 |
+
"enable_perm": false,
|
| 3718 |
"group_num": 1,
|
| 3719 |
"group_size": 4096,
|
| 3720 |
"in_features": 4096,
|
|
|
|
| 3738 |
"model.layers.29.mlp.down_proj": {
|
| 3739 |
"bias": false,
|
| 3740 |
"enable_norm": true,
|
| 3741 |
+
"enable_perm": false,
|
| 3742 |
"group_num": 1,
|
| 3743 |
"group_size": 14336,
|
| 3744 |
"in_features": 14336,
|
|
|
|
| 3762 |
"model.layers.29.mlp.gate_proj": {
|
| 3763 |
"bias": false,
|
| 3764 |
"enable_norm": true,
|
| 3765 |
+
"enable_perm": false,
|
| 3766 |
"group_num": 1,
|
| 3767 |
"group_size": 4096,
|
| 3768 |
"in_features": 4096,
|
|
|
|
| 3786 |
"model.layers.29.mlp.up_proj": {
|
| 3787 |
"bias": false,
|
| 3788 |
"enable_norm": true,
|
| 3789 |
+
"enable_perm": false,
|
| 3790 |
"group_num": 1,
|
| 3791 |
"group_size": 4096,
|
| 3792 |
"in_features": 4096,
|
|
|
|
| 3810 |
"model.layers.29.self_attn.k_proj": {
|
| 3811 |
"bias": false,
|
| 3812 |
"enable_norm": true,
|
| 3813 |
+
"enable_perm": false,
|
| 3814 |
"group_num": 1,
|
| 3815 |
"group_size": 4096,
|
| 3816 |
"in_features": 4096,
|
|
|
|
| 3834 |
"model.layers.29.self_attn.o_proj": {
|
| 3835 |
"bias": false,
|
| 3836 |
"enable_norm": true,
|
| 3837 |
+
"enable_perm": false,
|
| 3838 |
"group_num": 1,
|
| 3839 |
"group_size": 4096,
|
| 3840 |
"in_features": 4096,
|
|
|
|
| 3858 |
"model.layers.29.self_attn.q_proj": {
|
| 3859 |
"bias": false,
|
| 3860 |
"enable_norm": true,
|
| 3861 |
+
"enable_perm": false,
|
| 3862 |
"group_num": 1,
|
| 3863 |
"group_size": 4096,
|
| 3864 |
"in_features": 4096,
|
|
|
|
| 3882 |
"model.layers.29.self_attn.v_proj": {
|
| 3883 |
"bias": false,
|
| 3884 |
"enable_norm": true,
|
| 3885 |
+
"enable_perm": false,
|
| 3886 |
"group_num": 1,
|
| 3887 |
"group_size": 4096,
|
| 3888 |
"in_features": 4096,
|
|
|
|
| 3906 |
"model.layers.3.mlp.down_proj": {
|
| 3907 |
"bias": false,
|
| 3908 |
"enable_norm": true,
|
| 3909 |
+
"enable_perm": false,
|
| 3910 |
"group_num": 1,
|
| 3911 |
"group_size": 14336,
|
| 3912 |
"in_features": 14336,
|
|
|
|
| 3930 |
"model.layers.3.mlp.gate_proj": {
|
| 3931 |
"bias": false,
|
| 3932 |
"enable_norm": true,
|
| 3933 |
+
"enable_perm": false,
|
| 3934 |
"group_num": 1,
|
| 3935 |
"group_size": 4096,
|
| 3936 |
"in_features": 4096,
|
|
|
|
| 3954 |
"model.layers.3.mlp.up_proj": {
|
| 3955 |
"bias": false,
|
| 3956 |
"enable_norm": true,
|
| 3957 |
+
"enable_perm": false,
|
| 3958 |
"group_num": 1,
|
| 3959 |
"group_size": 4096,
|
| 3960 |
"in_features": 4096,
|
|
|
|
| 3978 |
"model.layers.3.self_attn.k_proj": {
|
| 3979 |
"bias": false,
|
| 3980 |
"enable_norm": true,
|
| 3981 |
+
"enable_perm": false,
|
| 3982 |
"group_num": 1,
|
| 3983 |
"group_size": 4096,
|
| 3984 |
"in_features": 4096,
|
|
|
|
| 4002 |
"model.layers.3.self_attn.o_proj": {
|
| 4003 |
"bias": false,
|
| 4004 |
"enable_norm": true,
|
| 4005 |
+
"enable_perm": false,
|
| 4006 |
"group_num": 1,
|
| 4007 |
"group_size": 4096,
|
| 4008 |
"in_features": 4096,
|
|
|
|
| 4026 |
"model.layers.3.self_attn.q_proj": {
|
| 4027 |
"bias": false,
|
| 4028 |
"enable_norm": true,
|
| 4029 |
+
"enable_perm": false,
|
| 4030 |
"group_num": 1,
|
| 4031 |
"group_size": 4096,
|
| 4032 |
"in_features": 4096,
|
|
|
|
| 4050 |
"model.layers.3.self_attn.v_proj": {
|
| 4051 |
"bias": false,
|
| 4052 |
"enable_norm": true,
|
| 4053 |
+
"enable_perm": false,
|
| 4054 |
"group_num": 1,
|
| 4055 |
"group_size": 4096,
|
| 4056 |
"in_features": 4096,
|
|
|
|
| 4074 |
"model.layers.30.mlp.down_proj": {
|
| 4075 |
"bias": false,
|
| 4076 |
"enable_norm": true,
|
| 4077 |
+
"enable_perm": false,
|
| 4078 |
"group_num": 1,
|
| 4079 |
"group_size": 14336,
|
| 4080 |
"in_features": 14336,
|
|
|
|
| 4098 |
"model.layers.30.mlp.gate_proj": {
|
| 4099 |
"bias": false,
|
| 4100 |
"enable_norm": true,
|
| 4101 |
+
"enable_perm": false,
|
| 4102 |
"group_num": 1,
|
| 4103 |
"group_size": 4096,
|
| 4104 |
"in_features": 4096,
|
|
|
|
| 4122 |
"model.layers.30.mlp.up_proj": {
|
| 4123 |
"bias": false,
|
| 4124 |
"enable_norm": true,
|
| 4125 |
+
"enable_perm": false,
|
| 4126 |
"group_num": 1,
|
| 4127 |
"group_size": 4096,
|
| 4128 |
"in_features": 4096,
|
|
|
|
| 4146 |
"model.layers.30.self_attn.k_proj": {
|
| 4147 |
"bias": false,
|
| 4148 |
"enable_norm": true,
|
| 4149 |
+
"enable_perm": false,
|
| 4150 |
"group_num": 1,
|
| 4151 |
"group_size": 4096,
|
| 4152 |
"in_features": 4096,
|
|
|
|
| 4170 |
"model.layers.30.self_attn.o_proj": {
|
| 4171 |
"bias": false,
|
| 4172 |
"enable_norm": true,
|
| 4173 |
+
"enable_perm": false,
|
| 4174 |
"group_num": 1,
|
| 4175 |
"group_size": 4096,
|
| 4176 |
"in_features": 4096,
|
|
|
|
| 4194 |
"model.layers.30.self_attn.q_proj": {
|
| 4195 |
"bias": false,
|
| 4196 |
"enable_norm": true,
|
| 4197 |
+
"enable_perm": false,
|
| 4198 |
"group_num": 1,
|
| 4199 |
"group_size": 4096,
|
| 4200 |
"in_features": 4096,
|
|
|
|
| 4218 |
"model.layers.30.self_attn.v_proj": {
|
| 4219 |
"bias": false,
|
| 4220 |
"enable_norm": true,
|
| 4221 |
+
"enable_perm": false,
|
| 4222 |
"group_num": 1,
|
| 4223 |
"group_size": 4096,
|
| 4224 |
"in_features": 4096,
|
|
|
|
| 4242 |
"model.layers.31.mlp.down_proj": {
|
| 4243 |
"bias": false,
|
| 4244 |
"enable_norm": true,
|
| 4245 |
+
"enable_perm": false,
|
| 4246 |
"group_num": 1,
|
| 4247 |
"group_size": 14336,
|
| 4248 |
"in_features": 14336,
|
|
|
|
| 4266 |
"model.layers.31.mlp.gate_proj": {
|
| 4267 |
"bias": false,
|
| 4268 |
"enable_norm": true,
|
| 4269 |
+
"enable_perm": false,
|
| 4270 |
"group_num": 1,
|
| 4271 |
"group_size": 4096,
|
| 4272 |
"in_features": 4096,
|
|
|
|
| 4290 |
"model.layers.31.mlp.up_proj": {
|
| 4291 |
"bias": false,
|
| 4292 |
"enable_norm": true,
|
| 4293 |
+
"enable_perm": false,
|
| 4294 |
"group_num": 1,
|
| 4295 |
"group_size": 4096,
|
| 4296 |
"in_features": 4096,
|
|
|
|
| 4314 |
"model.layers.31.self_attn.k_proj": {
|
| 4315 |
"bias": false,
|
| 4316 |
"enable_norm": true,
|
| 4317 |
+
"enable_perm": false,
|
| 4318 |
"group_num": 1,
|
| 4319 |
"group_size": 4096,
|
| 4320 |
"in_features": 4096,
|
|
|
|
| 4338 |
"model.layers.31.self_attn.o_proj": {
|
| 4339 |
"bias": false,
|
| 4340 |
"enable_norm": true,
|
| 4341 |
+
"enable_perm": false,
|
| 4342 |
"group_num": 1,
|
| 4343 |
"group_size": 4096,
|
| 4344 |
"in_features": 4096,
|
|
|
|
| 4362 |
"model.layers.31.self_attn.q_proj": {
|
| 4363 |
"bias": false,
|
| 4364 |
"enable_norm": true,
|
| 4365 |
+
"enable_perm": false,
|
| 4366 |
"group_num": 1,
|
| 4367 |
"group_size": 4096,
|
| 4368 |
"in_features": 4096,
|
|
|
|
| 4386 |
"model.layers.31.self_attn.v_proj": {
|
| 4387 |
"bias": false,
|
| 4388 |
"enable_norm": true,
|
| 4389 |
+
"enable_perm": false,
|
| 4390 |
"group_num": 1,
|
| 4391 |
"group_size": 4096,
|
| 4392 |
"in_features": 4096,
|
|
|
|
| 4410 |
"model.layers.4.mlp.down_proj": {
|
| 4411 |
"bias": false,
|
| 4412 |
"enable_norm": true,
|
| 4413 |
+
"enable_perm": false,
|
| 4414 |
"group_num": 1,
|
| 4415 |
"group_size": 14336,
|
| 4416 |
"in_features": 14336,
|
|
|
|
| 4434 |
"model.layers.4.mlp.gate_proj": {
|
| 4435 |
"bias": false,
|
| 4436 |
"enable_norm": true,
|
| 4437 |
+
"enable_perm": false,
|
| 4438 |
"group_num": 1,
|
| 4439 |
"group_size": 4096,
|
| 4440 |
"in_features": 4096,
|
|
|
|
| 4458 |
"model.layers.4.mlp.up_proj": {
|
| 4459 |
"bias": false,
|
| 4460 |
"enable_norm": true,
|
| 4461 |
+
"enable_perm": false,
|
| 4462 |
"group_num": 1,
|
| 4463 |
"group_size": 4096,
|
| 4464 |
"in_features": 4096,
|
|
|
|
| 4482 |
"model.layers.4.self_attn.k_proj": {
|
| 4483 |
"bias": false,
|
| 4484 |
"enable_norm": true,
|
| 4485 |
+
"enable_perm": false,
|
| 4486 |
"group_num": 1,
|
| 4487 |
"group_size": 4096,
|
| 4488 |
"in_features": 4096,
|
|
|
|
| 4506 |
"model.layers.4.self_attn.o_proj": {
|
| 4507 |
"bias": false,
|
| 4508 |
"enable_norm": true,
|
| 4509 |
+
"enable_perm": false,
|
| 4510 |
"group_num": 1,
|
| 4511 |
"group_size": 4096,
|
| 4512 |
"in_features": 4096,
|
|
|
|
| 4530 |
"model.layers.4.self_attn.q_proj": {
|
| 4531 |
"bias": false,
|
| 4532 |
"enable_norm": true,
|
| 4533 |
+
"enable_perm": false,
|
| 4534 |
"group_num": 1,
|
| 4535 |
"group_size": 4096,
|
| 4536 |
"in_features": 4096,
|
|
|
|
| 4554 |
"model.layers.4.self_attn.v_proj": {
|
| 4555 |
"bias": false,
|
| 4556 |
"enable_norm": true,
|
| 4557 |
+
"enable_perm": false,
|
| 4558 |
"group_num": 1,
|
| 4559 |
"group_size": 4096,
|
| 4560 |
"in_features": 4096,
|
|
|
|
| 4578 |
"model.layers.5.mlp.down_proj": {
|
| 4579 |
"bias": false,
|
| 4580 |
"enable_norm": true,
|
| 4581 |
+
"enable_perm": false,
|
| 4582 |
"group_num": 1,
|
| 4583 |
"group_size": 14336,
|
| 4584 |
"in_features": 14336,
|
|
|
|
| 4602 |
"model.layers.5.mlp.gate_proj": {
|
| 4603 |
"bias": false,
|
| 4604 |
"enable_norm": true,
|
| 4605 |
+
"enable_perm": false,
|
| 4606 |
"group_num": 1,
|
| 4607 |
"group_size": 4096,
|
| 4608 |
"in_features": 4096,
|
|
|
|
| 4626 |
"model.layers.5.mlp.up_proj": {
|
| 4627 |
"bias": false,
|
| 4628 |
"enable_norm": true,
|
| 4629 |
+
"enable_perm": false,
|
| 4630 |
"group_num": 1,
|
| 4631 |
"group_size": 4096,
|
| 4632 |
"in_features": 4096,
|
|
|
|
| 4650 |
"model.layers.5.self_attn.k_proj": {
|
| 4651 |
"bias": false,
|
| 4652 |
"enable_norm": true,
|
| 4653 |
+
"enable_perm": false,
|
| 4654 |
"group_num": 1,
|
| 4655 |
"group_size": 4096,
|
| 4656 |
"in_features": 4096,
|
|
|
|
| 4674 |
"model.layers.5.self_attn.o_proj": {
|
| 4675 |
"bias": false,
|
| 4676 |
"enable_norm": true,
|
| 4677 |
+
"enable_perm": false,
|
| 4678 |
"group_num": 1,
|
| 4679 |
"group_size": 4096,
|
| 4680 |
"in_features": 4096,
|
|
|
|
| 4698 |
"model.layers.5.self_attn.q_proj": {
|
| 4699 |
"bias": false,
|
| 4700 |
"enable_norm": true,
|
| 4701 |
+
"enable_perm": false,
|
| 4702 |
"group_num": 1,
|
| 4703 |
"group_size": 4096,
|
| 4704 |
"in_features": 4096,
|
|
|
|
| 4722 |
"model.layers.5.self_attn.v_proj": {
|
| 4723 |
"bias": false,
|
| 4724 |
"enable_norm": true,
|
| 4725 |
+
"enable_perm": false,
|
| 4726 |
"group_num": 1,
|
| 4727 |
"group_size": 4096,
|
| 4728 |
"in_features": 4096,
|
|
|
|
| 4746 |
"model.layers.6.mlp.down_proj": {
|
| 4747 |
"bias": false,
|
| 4748 |
"enable_norm": true,
|
| 4749 |
+
"enable_perm": false,
|
| 4750 |
"group_num": 1,
|
| 4751 |
"group_size": 14336,
|
| 4752 |
"in_features": 14336,
|
|
|
|
| 4770 |
"model.layers.6.mlp.gate_proj": {
|
| 4771 |
"bias": false,
|
| 4772 |
"enable_norm": true,
|
| 4773 |
+
"enable_perm": false,
|
| 4774 |
"group_num": 1,
|
| 4775 |
"group_size": 4096,
|
| 4776 |
"in_features": 4096,
|
|
|
|
| 4794 |
"model.layers.6.mlp.up_proj": {
|
| 4795 |
"bias": false,
|
| 4796 |
"enable_norm": true,
|
| 4797 |
+
"enable_perm": false,
|
| 4798 |
"group_num": 1,
|
| 4799 |
"group_size": 4096,
|
| 4800 |
"in_features": 4096,
|
|
|
|
| 4818 |
"model.layers.6.self_attn.k_proj": {
|
| 4819 |
"bias": false,
|
| 4820 |
"enable_norm": true,
|
| 4821 |
+
"enable_perm": false,
|
| 4822 |
"group_num": 1,
|
| 4823 |
"group_size": 4096,
|
| 4824 |
"in_features": 4096,
|
|
|
|
| 4842 |
"model.layers.6.self_attn.o_proj": {
|
| 4843 |
"bias": false,
|
| 4844 |
"enable_norm": true,
|
| 4845 |
+
"enable_perm": false,
|
| 4846 |
"group_num": 1,
|
| 4847 |
"group_size": 4096,
|
| 4848 |
"in_features": 4096,
|
|
|
|
| 4866 |
"model.layers.6.self_attn.q_proj": {
|
| 4867 |
"bias": false,
|
| 4868 |
"enable_norm": true,
|
| 4869 |
+
"enable_perm": false,
|
| 4870 |
"group_num": 1,
|
| 4871 |
"group_size": 4096,
|
| 4872 |
"in_features": 4096,
|
|
|
|
| 4890 |
"model.layers.6.self_attn.v_proj": {
|
| 4891 |
"bias": false,
|
| 4892 |
"enable_norm": true,
|
| 4893 |
+
"enable_perm": false,
|
| 4894 |
"group_num": 1,
|
| 4895 |
"group_size": 4096,
|
| 4896 |
"in_features": 4096,
|
|
|
|
| 4914 |
"model.layers.7.mlp.down_proj": {
|
| 4915 |
"bias": false,
|
| 4916 |
"enable_norm": true,
|
| 4917 |
+
"enable_perm": false,
|
| 4918 |
"group_num": 1,
|
| 4919 |
"group_size": 14336,
|
| 4920 |
"in_features": 14336,
|
|
|
|
| 4938 |
"model.layers.7.mlp.gate_proj": {
|
| 4939 |
"bias": false,
|
| 4940 |
"enable_norm": true,
|
| 4941 |
+
"enable_perm": false,
|
| 4942 |
"group_num": 1,
|
| 4943 |
"group_size": 4096,
|
| 4944 |
"in_features": 4096,
|
|
|
|
| 4962 |
"model.layers.7.mlp.up_proj": {
|
| 4963 |
"bias": false,
|
| 4964 |
"enable_norm": true,
|
| 4965 |
+
"enable_perm": false,
|
| 4966 |
"group_num": 1,
|
| 4967 |
"group_size": 4096,
|
| 4968 |
"in_features": 4096,
|
|
|
|
| 4986 |
"model.layers.7.self_attn.k_proj": {
|
| 4987 |
"bias": false,
|
| 4988 |
"enable_norm": true,
|
| 4989 |
+
"enable_perm": false,
|
| 4990 |
"group_num": 1,
|
| 4991 |
"group_size": 4096,
|
| 4992 |
"in_features": 4096,
|
|
|
|
| 5010 |
"model.layers.7.self_attn.o_proj": {
|
| 5011 |
"bias": false,
|
| 5012 |
"enable_norm": true,
|
| 5013 |
+
"enable_perm": false,
|
| 5014 |
"group_num": 1,
|
| 5015 |
"group_size": 4096,
|
| 5016 |
"in_features": 4096,
|
|
|
|
| 5034 |
"model.layers.7.self_attn.q_proj": {
|
| 5035 |
"bias": false,
|
| 5036 |
"enable_norm": true,
|
| 5037 |
+
"enable_perm": false,
|
| 5038 |
"group_num": 1,
|
| 5039 |
"group_size": 4096,
|
| 5040 |
"in_features": 4096,
|
|
|
|
| 5058 |
"model.layers.7.self_attn.v_proj": {
|
| 5059 |
"bias": false,
|
| 5060 |
"enable_norm": true,
|
| 5061 |
+
"enable_perm": false,
|
| 5062 |
"group_num": 1,
|
| 5063 |
"group_size": 4096,
|
| 5064 |
"in_features": 4096,
|
|
|
|
| 5082 |
"model.layers.8.mlp.down_proj": {
|
| 5083 |
"bias": false,
|
| 5084 |
"enable_norm": true,
|
| 5085 |
+
"enable_perm": false,
|
| 5086 |
"group_num": 1,
|
| 5087 |
"group_size": 14336,
|
| 5088 |
"in_features": 14336,
|
|
|
|
| 5106 |
"model.layers.8.mlp.gate_proj": {
|
| 5107 |
"bias": false,
|
| 5108 |
"enable_norm": true,
|
| 5109 |
+
"enable_perm": false,
|
| 5110 |
"group_num": 1,
|
| 5111 |
"group_size": 4096,
|
| 5112 |
"in_features": 4096,
|
|
|
|
| 5130 |
"model.layers.8.mlp.up_proj": {
|
| 5131 |
"bias": false,
|
| 5132 |
"enable_norm": true,
|
| 5133 |
+
"enable_perm": false,
|
| 5134 |
"group_num": 1,
|
| 5135 |
"group_size": 4096,
|
| 5136 |
"in_features": 4096,
|
|
|
|
| 5154 |
"model.layers.8.self_attn.k_proj": {
|
| 5155 |
"bias": false,
|
| 5156 |
"enable_norm": true,
|
| 5157 |
+
"enable_perm": false,
|
| 5158 |
"group_num": 1,
|
| 5159 |
"group_size": 4096,
|
| 5160 |
"in_features": 4096,
|
|
|
|
| 5178 |
"model.layers.8.self_attn.o_proj": {
|
| 5179 |
"bias": false,
|
| 5180 |
"enable_norm": true,
|
| 5181 |
+
"enable_perm": false,
|
| 5182 |
"group_num": 1,
|
| 5183 |
"group_size": 4096,
|
| 5184 |
"in_features": 4096,
|
|
|
|
| 5202 |
"model.layers.8.self_attn.q_proj": {
|
| 5203 |
"bias": false,
|
| 5204 |
"enable_norm": true,
|
| 5205 |
+
"enable_perm": false,
|
| 5206 |
"group_num": 1,
|
| 5207 |
"group_size": 4096,
|
| 5208 |
"in_features": 4096,
|
|
|
|
| 5226 |
"model.layers.8.self_attn.v_proj": {
|
| 5227 |
"bias": false,
|
| 5228 |
"enable_norm": true,
|
| 5229 |
+
"enable_perm": false,
|
| 5230 |
"group_num": 1,
|
| 5231 |
"group_size": 4096,
|
| 5232 |
"in_features": 4096,
|
|
|
|
| 5250 |
"model.layers.9.mlp.down_proj": {
|
| 5251 |
"bias": false,
|
| 5252 |
"enable_norm": true,
|
| 5253 |
+
"enable_perm": false,
|
| 5254 |
"group_num": 1,
|
| 5255 |
"group_size": 14336,
|
| 5256 |
"in_features": 14336,
|
|
|
|
| 5274 |
"model.layers.9.mlp.gate_proj": {
|
| 5275 |
"bias": false,
|
| 5276 |
"enable_norm": true,
|
| 5277 |
+
"enable_perm": false,
|
| 5278 |
"group_num": 1,
|
| 5279 |
"group_size": 4096,
|
| 5280 |
"in_features": 4096,
|
|
|
|
| 5298 |
"model.layers.9.mlp.up_proj": {
|
| 5299 |
"bias": false,
|
| 5300 |
"enable_norm": true,
|
| 5301 |
+
"enable_perm": false,
|
| 5302 |
"group_num": 1,
|
| 5303 |
"group_size": 4096,
|
| 5304 |
"in_features": 4096,
|
|
|
|
| 5322 |
"model.layers.9.self_attn.k_proj": {
|
| 5323 |
"bias": false,
|
| 5324 |
"enable_norm": true,
|
| 5325 |
+
"enable_perm": false,
|
| 5326 |
"group_num": 1,
|
| 5327 |
"group_size": 4096,
|
| 5328 |
"in_features": 4096,
|
|
|
|
| 5346 |
"model.layers.9.self_attn.o_proj": {
|
| 5347 |
"bias": false,
|
| 5348 |
"enable_norm": true,
|
| 5349 |
+
"enable_perm": false,
|
| 5350 |
"group_num": 1,
|
| 5351 |
"group_size": 4096,
|
| 5352 |
"in_features": 4096,
|
|
|
|
| 5370 |
"model.layers.9.self_attn.q_proj": {
|
| 5371 |
"bias": false,
|
| 5372 |
"enable_norm": true,
|
| 5373 |
+
"enable_perm": false,
|
| 5374 |
"group_num": 1,
|
| 5375 |
"group_size": 4096,
|
| 5376 |
"in_features": 4096,
|
|
|
|
| 5394 |
"model.layers.9.self_attn.v_proj": {
|
| 5395 |
"bias": false,
|
| 5396 |
"enable_norm": true,
|
| 5397 |
+
"enable_perm": false,
|
| 5398 |
"group_num": 1,
|
| 5399 |
"group_size": 4096,
|
| 5400 |
"in_features": 4096,
|