{ "lambdas": [ [ 0.3883242607116699, 0.3623622953891754, 0.37503084540367126 ], [ 0.34870514273643494, 0.37118518352508545, 0.3382723927497864 ], [ 0.31293290853500366, 0.360230952501297, 0.3535623252391815 ], [ 0.3090986907482147, 0.3132426142692566, 0.3656841218471527 ], [ 0.39302244782447815, 0.22840334475040436, 0.2582692801952362 ], [ 0.40333759784698486, 0.3732066750526428, 0.3943216800689697 ], [ 0.35193100571632385, 0.31543511152267456, 0.3743206262588501 ], [ 0.31697705388069153, 0.32546091079711914, 0.3447988033294678 ], [ 0.2927347719669342, 0.34207165241241455, 0.31707763671875 ], [ 0.335234671831131, 0.36454805731773376, 0.2865229547023773 ], [ 0.28157103061676025, 0.31637653708457947, 0.25314876437187195 ], [ 0.29041609168052673, 0.24318921566009521, 0.25509101152420044 ], [ 0.268072247505188, 0.3103688359260559, 0.385242760181427 ], [ 0.27690190076828003, 0.2568579316139221, 0.23891136050224304 ], [ 0.29108649492263794, 0.431260883808136, 0.3767254948616028 ], [ 0.2667229473590851, 0.4261268675327301, 0.2987927794456482 ], [ 0.3375805616378784, 0.2593033015727997, 0.3820607364177704 ], [ 0.29160040616989136, 0.29833856225013733, 0.2513200640678406 ], [ 0.41337326169013977, 0.2919619381427765, 0.18059132993221283 ], [ 0.35069024562835693, 0.3335267901420593, 0.3373335003852844 ], [ 0.34498515725135803, 0.3440103530883789, 0.32793059945106506 ], [ 0.20089201629161835, 0.39681825041770935, 0.18193161487579346 ], [ 0.39463290572166443, 0.37697556614875793, 0.1828676462173462 ], [ 0.38112935423851013, 0.31183040142059326, 0.39470359683036804 ], [ 0.3648071587085724, 0.31466302275657654, 0.38174504041671753 ], [ 0.28994786739349365, 0.2857919931411743, 0.2895475924015045 ], [ 0.23836930096149445, 0.25469934940338135, 0.20404624938964844 ], [ 0.3720416724681854, 0.2436002939939499, 0.29299506545066833 ], [ 0.3178931474685669, 0.25512033700942993, 0.26282304525375366 ], [ 0.3162326216697693, 0.2617306113243103, 0.2617321014404297 ], [ 0.19119736552238464, 0.22181381285190582, 0.19509093463420868 ], [ 0.39264020323753357, 0.21596397459506989, 0.2147778570652008 ], [ 0.2696373760700226, 0.3549906313419342, 0.38246554136276245 ], [ 0.3275165557861328, 0.34763646125793457, 0.3809554874897003 ], [ 0.2816231846809387, 0.2845311760902405, 0.26921606063842773 ], [ 0.24686750769615173, 0.26908186078071594, 0.36267468333244324 ], [ 0.37496766448020935, 0.22052006423473358, 0.2156810164451599 ], [ 0.3127693831920624, 0.3327944874763489, 0.31126803159713745 ], [ 0.24814096093177795, 0.33962151408195496, 0.3332516551017761 ], [ 0.28147992491722107, 0.20830675959587097, 0.2245684117078781 ], [ 0.3766174018383026, 0.20415450632572174, 0.20479559898376465 ], [ 0.325105756521225, 0.29087960720062256, 0.28362035751342773 ], [ 0.32374194264411926, 0.24844275414943695, 0.3287203907966614 ], [ 0.3562851548194885, 0.34408044815063477, 0.33212822675704956 ], [ 0.24317413568496704, 0.3342779576778412, 0.3679232597351074 ], [ 0.28208789229393005, 0.23189561069011688, 0.3551972210407257 ], [ 0.31444409489631653, 0.28177592158317566, 0.2999891936779022 ], [ 0.27754905819892883, 0.3250216841697693, 0.3399043083190918 ], [ 0.22077758610248566, 0.24111232161521912, 0.20977291464805603 ], [ 0.21510633826255798, 0.20415279269218445, 0.34861651062965393 ], [ 0.33469030261039734, 0.22629156708717346, 0.3223295509815216 ], [ 0.4043980538845062, 0.29894107580184937, 0.2800503373146057 ], [ 0.3535716235637665, 0.24572545289993286, 0.24943549931049347 ], [ 0.27787351608276367, 0.19825978577136993, 0.2795927822589874 ], [ 0.3531836271286011, 0.28581106662750244, 0.251440167427063 ], [ 0.40898948907852173, 0.38315364718437195, 0.3664551079273224 ], [ 0.3686511814594269, 0.3719461262226105, 0.3300400376319885 ], [ 0.21359319984912872, 0.2162943333387375, 0.19426186382770538 ], [ 0.2621611952781677, 0.19908210635185242, 0.27713072299957275 ], [ 0.3029685318470001, 0.32555097341537476, 0.28010445833206177 ], [ 0.23895785212516785, 0.21855375170707703, 0.3792441487312317 ], [ 0.38740450143814087, 0.23142819106578827, 0.298105388879776 ], [ 0.2297668755054474, 0.4054788649082184, 0.33879172801971436 ], [ 0.23646755516529083, 0.2156289964914322, 0.33678293228149414 ], [ 0.2598417401313782, 0.3474366068840027, 0.3489058017730713 ], [ 0.3630068302154541, 0.37734073400497437, 0.3596450090408325 ], [ 0.3690379858016968, 0.33453360199928284, 0.23141241073608398 ], [ 0.2112365961074829, 0.20301534235477448, 0.42213261127471924 ], [ 0.37722212076187134, 0.2605799436569214, 0.27029818296432495 ], [ 0.4015510082244873, 0.2545139491558075, 0.2512376010417938 ], [ 0.3890366852283478, 0.2706148624420166, 0.3350306749343872 ], [ 0.3031468093395233, 0.2991064488887787, 0.30880942940711975 ], [ 0.22063690423965454, 0.3345808684825897, 0.2924988865852356 ], [ 0.2651190161705017, 0.2819417119026184, 0.23100979626178741 ], [ 0.33170583844184875, 0.25984010100364685, 0.24688838422298431 ], [ 0.3842375874519348, 0.40348589420318604, 0.39649391174316406 ], [ 0.20523031055927277, 0.24208150804042816, 0.4134710431098938 ], [ 0.25265148282051086, 0.24085453152656555, 0.3437962532043457 ], [ 0.2726953327655792, 0.23532339930534363, 0.2919284999370575 ], [ 0.3999418318271637, 0.20833371579647064, 0.24912938475608826 ], [ 0.3016860783100128, 0.2991064488887787, 0.291384756565094 ], [ 0.2326887547969818, 0.20517869293689728, 0.3575997054576874 ], [ 0.35019832849502563, 0.3487541377544403, 0.3654879331588745 ], [ 0.2886132001876831, 0.3061191737651825, 0.30866026878356934 ], [ 0.2705903947353363, 0.29687198996543884, 0.200924351811409 ], [ 0.3555452525615692, 0.3329847455024719, 0.22329427301883698 ], [ 0.27235499024391174, 0.2524101436138153, 0.23791800439357758 ], [ 0.23980098962783813, 0.40368956327438354, 0.22931145131587982 ], [ 0.2874733507633209, 0.22713828086853027, 0.4087265431880951 ], [ 0.30632638931274414, 0.3704463839530945, 0.3880924582481384 ], [ 0.26323893666267395, 0.21290108561515808, 0.3708689510822296 ], [ 0.3029574453830719, 0.2707969546318054, 0.3391459882259369 ], [ 0.3461504876613617, 0.3970031142234802, 0.3742915093898773 ], [ 0.1840425431728363, 0.2419389933347702, 0.2897161841392517 ], [ 0.353685200214386, 0.3755283057689667, 0.34905770421028137 ], [ 0.24832041561603546, 0.2902229130268097, 0.2151690125465393 ], [ 0.29903411865234375, 0.37380513548851013, 0.2661544382572174 ], [ 0.39497217535972595, 0.33879315853118896, 0.35845938324928284 ], [ 0.24488131701946259, 0.2991064488887787, 0.22929759323596954 ], [ 0.3821658492088318, 0.37903353571891785, 0.181688129901886 ], [ 0.3847050666809082, 0.23709334433078766, 0.3856165409088135 ], [ 0.3608607053756714, 0.3747178912162781, 0.3704005181789398 ], [ 0.37809160351753235, 0.3927466571331024, 0.37344133853912354 ], [ 0.22261585295200348, 0.1994251310825348, 0.21917051076889038 ], [ 0.2696945071220398, 0.33371502161026, 0.4003022313117981 ], [ 0.3206605017185211, 0.3373804986476898, 0.280136376619339 ], [ 0.260978102684021, 0.24670591950416565, 0.39149796962738037 ], [ 0.34051597118377686, 0.37631818652153015, 0.25929149985313416 ], [ 0.3907065987586975, 0.38418787717819214, 0.20959261059761047 ], [ 0.2965652048587799, 0.29861822724342346, 0.31973996758461 ], [ 0.2142055779695511, 0.22334299981594086, 0.23120635747909546 ], [ 0.33672475814819336, 0.23128989338874817, 0.241999089717865 ], [ 0.262056827545166, 0.25463253259658813, 0.3349643051624298 ], [ 0.3651188313961029, 0.3639734387397766, 0.3300137221813202 ], [ 0.26690706610679626, 0.3640022873878479, 0.297036737203598 ], [ 0.40695059299468994, 0.3969866931438446, 0.36739039421081543 ], [ 0.35780614614486694, 0.3466162383556366, 0.22807283699512482 ], [ 0.3790653347969055, 0.36814889311790466, 0.35832932591438293 ], [ 0.37500086426734924, 0.40332910418510437, 0.3131495416164398 ], [ 0.2898157835006714, 0.32542046904563904, 0.2649974226951599 ], [ 0.27568769454956055, 0.35351213812828064, 0.19069719314575195 ], [ 0.3537901043891907, 0.21330882608890533, 0.2154841274023056 ], [ 0.39461657404899597, 0.22658494114875793, 0.40037137269973755 ], [ 0.32012900710105896, 0.22551126778125763, 0.1952424943447113 ], [ 0.4044588506221771, 0.2870043218135834, 0.41318216919898987 ], [ 0.24841244518756866, 0.2991064488887787, 0.359804630279541 ], [ 0.20572419464588165, 0.37685900926589966, 0.20065918564796448 ], [ 0.2231069654226303, 0.36666762828826904, 0.376221239566803 ], [ 0.31154945492744446, 0.33882391452789307, 0.3518316149711609 ], [ 0.275085985660553, 0.3718191385269165, 0.20597892999649048 ], [ 0.21165381371974945, 0.3918326497077942, 0.20218192040920258 ], [ 0.2142898589372635, 0.20601648092269897, 0.2011740654706955 ], [ 0.2238815873861313, 0.34816882014274597, 0.21244102716445923 ], [ 0.3886909484863281, 0.3721139132976532, 0.39156341552734375 ], [ 0.2719188928604126, 0.24027907848358154, 0.2529193162918091 ], [ 0.21617725491523743, 0.31258508563041687, 0.21470600366592407 ], [ 0.37306591868400574, 0.3109174966812134, 0.25874271988868713 ], [ 0.30624455213546753, 0.34182608127593994, 0.35792678594589233 ], [ 0.20845377445220947, 0.38172876834869385, 0.21456488966941833 ], [ 0.19948387145996094, 0.24893465638160706, 0.39198803901672363 ], [ 0.33647772669792175, 0.3894765377044678, 0.23352539539337158 ], [ 0.37234070897102356, 0.3916969895362854, 0.23035132884979248 ], [ 0.4004880487918854, 0.2232302725315094, 0.28357723355293274 ], [ 0.34418413043022156, 0.2991064488887787, 0.3528580069541931 ], [ 0.22602835297584534, 0.36724552512168884, 0.29569509625434875 ], [ 0.267846941947937, 0.30040356516838074, 0.3327115774154663 ], [ 0.3129286766052246, 0.35984790325164795, 0.33761999011039734 ], [ 0.30865105986595154, 0.3076038062572479, 0.27449706196784973 ], [ 0.21445733308792114, 0.3321622908115387, 0.2170647233724594 ], [ 0.34500300884246826, 0.30192336440086365, 0.3100443482398987 ], [ 0.4023292064666748, 0.19847708940505981, 0.22118930518627167 ], [ 0.41021573543548584, 0.3522898256778717, 0.2509464919567108 ], [ 0.2809571623802185, 0.2991064488887787, 0.3436353802680969 ], [ 0.3859373927116394, 0.39896419644355774, 0.2005540281534195 ], [ 0.3510185182094574, 0.24334318935871124, 0.32767847180366516 ], [ 0.2913549244403839, 0.36552032828330994, 0.3718450665473938 ], [ 0.1889614313840866, 0.40470564365386963, 0.3549887537956238 ], [ 0.19629013538360596, 0.225816011428833, 0.2007388472557068 ], [ 0.37456709146499634, 0.3941386938095093, 0.3517758548259735 ], [ 0.2556999921798706, 0.20328563451766968, 0.39413779973983765 ], [ 0.39653894305229187, 0.41154104471206665, 0.30622464418411255 ], [ 0.38080868124961853, 0.2991064488887787, 0.24592125415802002 ], [ 0.3313809335231781, 0.27195805311203003, 0.2208387404680252 ], [ 0.3555898666381836, 0.31074872612953186, 0.36027273535728455 ], [ 0.37996920943260193, 0.3645731508731842, 0.3388481140136719 ], [ 0.28789493441581726, 0.3893921971321106, 0.22932587563991547 ], [ 0.41367262601852417, 0.4106283485889435, 0.37968260049819946 ], [ 0.3753623068332672, 0.37106749415397644, 0.23023924231529236 ], [ 0.18712924420833588, 0.3284003436565399, 0.25042104721069336 ], [ 0.35538166761398315, 0.42108720541000366, 0.285932332277298 ], [ 0.3677556812763214, 0.2991064488887787, 0.36747413873672485 ], [ 0.2664182186126709, 0.32455241680145264, 0.3703777492046356 ], [ 0.16972634196281433, 0.3778935968875885, 0.2932637631893158 ], [ 0.4039979577064514, 0.40259143710136414, 0.4136446416378021 ], [ 0.20013238489627838, 0.30873557925224304, 0.20913326740264893 ], [ 0.19595161080360413, 0.30672842264175415, 0.2452753186225891 ], [ 0.391770601272583, 0.30807214975357056, 0.32796725630760193 ], [ 0.23373982310295105, 0.19680923223495483, 0.19121727347373962 ], [ 0.2541869878768921, 0.3787055015563965, 0.4557085633277893 ], [ 0.39971253275871277, 0.2991064488887787, 0.2321351170539856 ], [ 0.2782626152038574, 0.19146595895290375, 0.2829567790031433 ], [ 0.25045517086982727, 0.3819902539253235, 0.2539924085140228 ], [ 0.3782663345336914, 0.41245192289352417, 0.24768783152103424 ], [ 0.30407649278640747, 0.20620810985565186, 0.22133873403072357 ], [ 0.21233916282653809, 0.19252246618270874, 0.4018924832344055 ], [ 0.432591050863266, 0.23376668989658356, 0.24733340740203857 ], [ 0.4144866466522217, 0.24681518971920013, 0.2397996336221695 ], [ 0.3041132092475891, 0.3268588185310364, 0.28527912497520447 ], [ 0.39036187529563904, 0.2991064488887787, 0.39712661504745483 ], [ 0.3849163353443146, 0.23858928680419922, 0.3843546509742737 ], [ 0.3309418857097626, 0.3194034695625305, 0.30275431275367737 ], [ 0.28447964787483215, 0.27185967564582825, 0.24963083863258362 ], [ 0.35889676213264465, 0.3565056324005127, 0.43252548575401306 ], [ 0.38172295689582825, 0.16952233016490936, 0.3983158767223358 ], [ 0.3377133011817932, 0.33234190940856934, 0.21213269233703613 ], [ 0.3571989834308624, 0.17591749131679535, 0.23240876197814941 ], [ 0.4361167848110199, 0.27213338017463684, 0.32464003562927246 ], [ 0.3173784911632538, 0.2991064488887787, 0.30699804425239563 ], [ 0.3447878360748291, 0.21996575593948364, 0.32738256454467773 ], [ 0.36943623423576355, 0.20948199927806854, 0.25647279620170593 ], [ 0.23839440941810608, 0.2092297524213791, 0.24825681746006012 ], [ 0.17821237444877625, 0.3349408805370331, 0.32569828629493713 ], [ 0.40162524580955505, 0.3743557929992676, 0.3871481716632843 ], [ 0.4137503206729889, 0.32879936695098877, 0.18728595972061157 ], [ 0.37369292974472046, 0.27171987295150757, 0.26730969548225403 ], [ 0.43127456307411194, 0.40427279472351074, 0.44116276502609253 ], [ 0.1914208084344864, 0.35479849576950073, 0.38604509830474854 ], [ 0.39088666439056396, 0.37367960810661316, 0.20381216704845428 ], [ 0.16992826759815216, 0.19981710612773895, 0.21016918122768402 ], [ 0.27916085720062256, 0.21161127090454102, 0.19437551498413086 ], [ 0.25305089354515076, 0.39059704542160034, 0.2465861737728119 ], [ 0.40665534138679504, 0.3627412021160126, 0.23964105546474457 ], [ 0.19555631279945374, 0.3164346218109131, 0.29416850209236145 ], [ 0.3135971426963806, 0.2342560589313507, 0.24193967878818512 ], [ 0.43025943636894226, 0.19899214804172516, 0.4302777647972107 ], [ 0.2256966531276703, 0.2991064488887787, 0.23316973447799683 ], [ 0.23912502825260162, 0.20957373082637787, 0.19495394825935364 ], [ 0.23008137941360474, 0.27278825640678406, 0.41580867767333984 ], [ 0.24287232756614685, 0.30981460213661194, 0.27125540375709534 ], [ 0.1671687513589859, 0.31038036942481995, 0.34663137793540955 ], [ 0.42514514923095703, 0.17283941805362701, 0.4156189560890198 ], [ 0.4168146550655365, 0.4137655794620514, 0.4279871881008148 ], [ 0.42051973938941956, 0.41071876883506775, 0.36794158816337585 ], [ 0.2451215535402298, 0.3930608630180359, 0.41934332251548767 ], [ 0.22715631127357483, 0.32089999318122864, 0.3635891079902649 ], [ 0.3904862403869629, 0.3821772336959839, 0.362802654504776 ], [ 0.3917807638645172, 0.4336768090724945, 0.393928200006485 ], [ 0.4068164527416229, 0.36042362451553345, 0.38350456953048706 ], [ 0.3348886966705322, 0.26808488368988037, 0.29866161942481995 ], [ 0.4200945198535919, 0.2508523166179657, 0.4419146180152893 ], [ 0.1653856486082077, 0.3859034776687622, 0.23257014155387878 ], [ 0.410726934671402, 0.2249910980463028, 0.40766558051109314 ], [ 0.24586184322834015, 0.3176778554916382, 0.4229857921600342 ], [ 0.4058119058609009, 0.24759773910045624, 0.20222556591033936 ], [ 0.29266124963760376, 0.21434326469898224, 0.37526044249534607 ], [ 0.35474422574043274, 0.4108079969882965, 0.33864670991897583 ], [ 0.35048675537109375, 0.17898240685462952, 0.19693785905838013 ], [ 0.4039932191371918, 0.2854495942592621, 0.3935033977031708 ], [ 0.435630202293396, 0.4274696707725525, 0.1990002989768982 ], [ 0.44515126943588257, 0.412635862827301, 0.39780303835868835 ], [ 0.4206939935684204, 0.19284026324748993, 0.4129420816898346 ], [ 0.2042851448059082, 0.28650370240211487, 0.43368643522262573 ], [ 0.3875381350517273, 0.2991064488887787, 0.3154343068599701 ], [ 0.3956281542778015, 0.384502649307251, 0.2036466896533966 ], [ 0.35721901059150696, 0.29928287863731384, 0.2603643238544464 ], [ 0.2551652491092682, 0.2146013081073761, 0.22017168998718262 ], [ 0.1848575323820114, 0.34243515133857727, 0.21192139387130737 ], [ 0.3702299892902374, 0.18249468505382538, 0.19415122270584106 ], [ 0.4407699704170227, 0.2862900197505951, 0.20192617177963257 ], [ 0.2997152507305145, 0.20137713849544525, 0.3658028244972229 ], [ 0.22969289124011993, 0.31545186042785645, 0.4249734580516815 ], [ 0.19194096326828003, 0.2991064488887787, 0.29746097326278687 ], [ 0.38833168148994446, 0.2264913022518158, 0.21622535586357117 ], [ 0.3419898748397827, 0.32490473985671997, 0.28989049792289734 ], [ 0.3733155131340027, 0.2695937156677246, 0.327462375164032 ], [ 0.16603581607341766, 0.2065427005290985, 0.19622394442558289 ], [ 0.39046502113342285, 0.3112553358078003, 0.30394476652145386 ], [ 0.19536720216274261, 0.2507157623767853, 0.17685969173908234 ], [ 0.4354845881462097, 0.1725979447364807, 0.4108954966068268 ], [ 0.1687704473733902, 0.42737966775894165, 0.3615131080150604 ], [ 0.3073764443397522, 0.30011141300201416, 0.2252797782421112 ], [ 0.34714677929878235, 0.38223206996917725, 0.3392619490623474 ], [ 0.29542046785354614, 0.22996902465820312, 0.3547264635562897 ], [ 0.4252474904060364, 0.3659898638725281, 0.21462717652320862 ], [ 0.20209349691867828, 0.1717177927494049, 0.42266833782196045 ], [ 0.4049817621707916, 0.3163258135318756, 0.4065883457660675 ], [ 0.40989041328430176, 0.16486363112926483, 0.3653058409690857 ], [ 0.2331339716911316, 0.20539624989032745, 0.35222890973091125 ], [ 0.15381211042404175, 0.19398052990436554, 0.24659037590026855 ], [ 0.3463601768016815, 0.2991064488887787, 0.2792954742908478 ], [ 0.2949776351451874, 0.37481826543807983, 0.33055099844932556 ], [ 0.36206018924713135, 0.3842589557170868, 0.2699452340602875 ], [ 0.23476417362689972, 0.36749500036239624, 0.2768413722515106 ], [ 0.24786831438541412, 0.41904473304748535, 0.3566398620605469 ], [ 0.40717893838882446, 0.40397050976753235, 0.4254368543624878 ], [ 0.28767332434654236, 0.17419925332069397, 0.17939725518226624 ], [ 0.2289697378873825, 0.2704469561576843, 0.2715151906013489 ], [ 0.18049292266368866, 0.24434413015842438, 0.369288831949234 ], [ 0.26268142461776733, 0.33503296971321106, 0.23925383388996124 ], [ 0.3088991343975067, 0.4438924789428711, 0.3879861533641815 ], [ 0.20758670568466187, 0.29517585039138794, 0.3329733610153198 ], [ 0.2305309772491455, 0.21366733312606812, 0.25563523173332214 ], [ 0.16969293355941772, 0.3407335579395294, 0.4500804841518402 ], [ 0.3295100927352905, 0.4438541829586029, 0.4291228652000427 ], [ 0.4329433739185333, 0.16973084211349487, 0.19713136553764343 ], [ 0.44385749101638794, 0.21979974210262299, 0.1725056916475296 ], [ 0.144502654671669, 0.46215271949768066, 0.4555712938308716 ], [ 0.40285441279411316, 0.2991064488887787, 0.23847955465316772 ], [ 0.13851185142993927, 0.2991064488887787, 0.4169304072856903 ], [ 0.48601633310317993, 0.160946324467659, 0.10018127411603928 ], [ 0.5619405508041382, 0.559393584728241, 0.4568527340888977 ] ], "model_names": [ "instruct", "math", "code" ], "num_models": 3, "num_params": 291, "param_names": [ "model.embed_tokens.weight", "model.layers.0.self_attn.q_proj.weight", "model.layers.0.self_attn.k_proj.weight", "model.layers.0.self_attn.v_proj.weight", "model.layers.0.self_attn.o_proj.weight", "model.layers.0.mlp.gate_proj.weight", "model.layers.0.mlp.up_proj.weight", "model.layers.0.mlp.down_proj.weight", "model.layers.0.input_layernorm.weight", "model.layers.0.post_attention_layernorm.weight", "model.layers.1.self_attn.q_proj.weight", "model.layers.1.self_attn.k_proj.weight", "model.layers.1.self_attn.v_proj.weight", "model.layers.1.self_attn.o_proj.weight", "model.layers.1.mlp.gate_proj.weight", "model.layers.1.mlp.up_proj.weight", "model.layers.1.mlp.down_proj.weight", "model.layers.1.input_layernorm.weight", "model.layers.1.post_attention_layernorm.weight", "model.layers.2.self_attn.q_proj.weight", "model.layers.2.self_attn.k_proj.weight", "model.layers.2.self_attn.v_proj.weight", "model.layers.2.self_attn.o_proj.weight", "model.layers.2.mlp.gate_proj.weight", "model.layers.2.mlp.up_proj.weight", "model.layers.2.mlp.down_proj.weight", "model.layers.2.input_layernorm.weight", "model.layers.2.post_attention_layernorm.weight", "model.layers.3.self_attn.q_proj.weight", "model.layers.3.self_attn.k_proj.weight", "model.layers.3.self_attn.v_proj.weight", "model.layers.3.self_attn.o_proj.weight", "model.layers.3.mlp.gate_proj.weight", "model.layers.3.mlp.up_proj.weight", "model.layers.3.mlp.down_proj.weight", "model.layers.3.input_layernorm.weight", "model.layers.3.post_attention_layernorm.weight", "model.layers.4.self_attn.q_proj.weight", "model.layers.4.self_attn.k_proj.weight", "model.layers.4.self_attn.v_proj.weight", "model.layers.4.self_attn.o_proj.weight", "model.layers.4.mlp.gate_proj.weight", "model.layers.4.mlp.up_proj.weight", "model.layers.4.mlp.down_proj.weight", "model.layers.4.input_layernorm.weight", "model.layers.4.post_attention_layernorm.weight", "model.layers.5.self_attn.q_proj.weight", "model.layers.5.self_attn.k_proj.weight", "model.layers.5.self_attn.v_proj.weight", "model.layers.5.self_attn.o_proj.weight", "model.layers.5.mlp.gate_proj.weight", "model.layers.5.mlp.up_proj.weight", "model.layers.5.mlp.down_proj.weight", "model.layers.5.input_layernorm.weight", "model.layers.5.post_attention_layernorm.weight", "model.layers.6.self_attn.q_proj.weight", "model.layers.6.self_attn.k_proj.weight", "model.layers.6.self_attn.v_proj.weight", "model.layers.6.self_attn.o_proj.weight", "model.layers.6.mlp.gate_proj.weight", "model.layers.6.mlp.up_proj.weight", "model.layers.6.mlp.down_proj.weight", "model.layers.6.input_layernorm.weight", "model.layers.6.post_attention_layernorm.weight", "model.layers.7.self_attn.q_proj.weight", "model.layers.7.self_attn.k_proj.weight", "model.layers.7.self_attn.v_proj.weight", "model.layers.7.self_attn.o_proj.weight", "model.layers.7.mlp.gate_proj.weight", "model.layers.7.mlp.up_proj.weight", "model.layers.7.mlp.down_proj.weight", "model.layers.7.input_layernorm.weight", "model.layers.7.post_attention_layernorm.weight", "model.layers.8.self_attn.q_proj.weight", "model.layers.8.self_attn.k_proj.weight", "model.layers.8.self_attn.v_proj.weight", "model.layers.8.self_attn.o_proj.weight", "model.layers.8.mlp.gate_proj.weight", "model.layers.8.mlp.up_proj.weight", "model.layers.8.mlp.down_proj.weight", "model.layers.8.input_layernorm.weight", "model.layers.8.post_attention_layernorm.weight", "model.layers.9.self_attn.q_proj.weight", "model.layers.9.self_attn.k_proj.weight", "model.layers.9.self_attn.v_proj.weight", "model.layers.9.self_attn.o_proj.weight", "model.layers.9.mlp.gate_proj.weight", "model.layers.9.mlp.up_proj.weight", "model.layers.9.mlp.down_proj.weight", "model.layers.9.input_layernorm.weight", "model.layers.9.post_attention_layernorm.weight", "model.layers.10.self_attn.q_proj.weight", "model.layers.10.self_attn.k_proj.weight", "model.layers.10.self_attn.v_proj.weight", "model.layers.10.self_attn.o_proj.weight", "model.layers.10.mlp.gate_proj.weight", "model.layers.10.mlp.up_proj.weight", "model.layers.10.mlp.down_proj.weight", "model.layers.10.input_layernorm.weight", "model.layers.10.post_attention_layernorm.weight", "model.layers.11.self_attn.q_proj.weight", "model.layers.11.self_attn.k_proj.weight", "model.layers.11.self_attn.v_proj.weight", "model.layers.11.self_attn.o_proj.weight", "model.layers.11.mlp.gate_proj.weight", "model.layers.11.mlp.up_proj.weight", "model.layers.11.mlp.down_proj.weight", "model.layers.11.input_layernorm.weight", "model.layers.11.post_attention_layernorm.weight", "model.layers.12.self_attn.q_proj.weight", "model.layers.12.self_attn.k_proj.weight", "model.layers.12.self_attn.v_proj.weight", "model.layers.12.self_attn.o_proj.weight", "model.layers.12.mlp.gate_proj.weight", "model.layers.12.mlp.up_proj.weight", "model.layers.12.mlp.down_proj.weight", "model.layers.12.input_layernorm.weight", "model.layers.12.post_attention_layernorm.weight", "model.layers.13.self_attn.q_proj.weight", "model.layers.13.self_attn.k_proj.weight", "model.layers.13.self_attn.v_proj.weight", "model.layers.13.self_attn.o_proj.weight", "model.layers.13.mlp.gate_proj.weight", "model.layers.13.mlp.up_proj.weight", "model.layers.13.mlp.down_proj.weight", "model.layers.13.input_layernorm.weight", "model.layers.13.post_attention_layernorm.weight", "model.layers.14.self_attn.q_proj.weight", "model.layers.14.self_attn.k_proj.weight", "model.layers.14.self_attn.v_proj.weight", "model.layers.14.self_attn.o_proj.weight", "model.layers.14.mlp.gate_proj.weight", "model.layers.14.mlp.up_proj.weight", "model.layers.14.mlp.down_proj.weight", "model.layers.14.input_layernorm.weight", "model.layers.14.post_attention_layernorm.weight", "model.layers.15.self_attn.q_proj.weight", "model.layers.15.self_attn.k_proj.weight", "model.layers.15.self_attn.v_proj.weight", "model.layers.15.self_attn.o_proj.weight", "model.layers.15.mlp.gate_proj.weight", "model.layers.15.mlp.up_proj.weight", "model.layers.15.mlp.down_proj.weight", "model.layers.15.input_layernorm.weight", "model.layers.15.post_attention_layernorm.weight", "model.layers.16.self_attn.q_proj.weight", "model.layers.16.self_attn.k_proj.weight", "model.layers.16.self_attn.v_proj.weight", "model.layers.16.self_attn.o_proj.weight", "model.layers.16.mlp.gate_proj.weight", "model.layers.16.mlp.up_proj.weight", "model.layers.16.mlp.down_proj.weight", "model.layers.16.input_layernorm.weight", "model.layers.16.post_attention_layernorm.weight", "model.layers.17.self_attn.q_proj.weight", "model.layers.17.self_attn.k_proj.weight", "model.layers.17.self_attn.v_proj.weight", "model.layers.17.self_attn.o_proj.weight", "model.layers.17.mlp.gate_proj.weight", "model.layers.17.mlp.up_proj.weight", "model.layers.17.mlp.down_proj.weight", "model.layers.17.input_layernorm.weight", "model.layers.17.post_attention_layernorm.weight", "model.layers.18.self_attn.q_proj.weight", "model.layers.18.self_attn.k_proj.weight", "model.layers.18.self_attn.v_proj.weight", "model.layers.18.self_attn.o_proj.weight", "model.layers.18.mlp.gate_proj.weight", "model.layers.18.mlp.up_proj.weight", "model.layers.18.mlp.down_proj.weight", "model.layers.18.input_layernorm.weight", "model.layers.18.post_attention_layernorm.weight", "model.layers.19.self_attn.q_proj.weight", "model.layers.19.self_attn.k_proj.weight", "model.layers.19.self_attn.v_proj.weight", "model.layers.19.self_attn.o_proj.weight", "model.layers.19.mlp.gate_proj.weight", "model.layers.19.mlp.up_proj.weight", "model.layers.19.mlp.down_proj.weight", "model.layers.19.input_layernorm.weight", "model.layers.19.post_attention_layernorm.weight", "model.layers.20.self_attn.q_proj.weight", "model.layers.20.self_attn.k_proj.weight", "model.layers.20.self_attn.v_proj.weight", "model.layers.20.self_attn.o_proj.weight", "model.layers.20.mlp.gate_proj.weight", "model.layers.20.mlp.up_proj.weight", "model.layers.20.mlp.down_proj.weight", "model.layers.20.input_layernorm.weight", "model.layers.20.post_attention_layernorm.weight", "model.layers.21.self_attn.q_proj.weight", "model.layers.21.self_attn.k_proj.weight", "model.layers.21.self_attn.v_proj.weight", "model.layers.21.self_attn.o_proj.weight", "model.layers.21.mlp.gate_proj.weight", "model.layers.21.mlp.up_proj.weight", "model.layers.21.mlp.down_proj.weight", "model.layers.21.input_layernorm.weight", "model.layers.21.post_attention_layernorm.weight", "model.layers.22.self_attn.q_proj.weight", "model.layers.22.self_attn.k_proj.weight", "model.layers.22.self_attn.v_proj.weight", "model.layers.22.self_attn.o_proj.weight", "model.layers.22.mlp.gate_proj.weight", "model.layers.22.mlp.up_proj.weight", "model.layers.22.mlp.down_proj.weight", "model.layers.22.input_layernorm.weight", "model.layers.22.post_attention_layernorm.weight", "model.layers.23.self_attn.q_proj.weight", "model.layers.23.self_attn.k_proj.weight", "model.layers.23.self_attn.v_proj.weight", "model.layers.23.self_attn.o_proj.weight", "model.layers.23.mlp.gate_proj.weight", "model.layers.23.mlp.up_proj.weight", "model.layers.23.mlp.down_proj.weight", "model.layers.23.input_layernorm.weight", "model.layers.23.post_attention_layernorm.weight", "model.layers.24.self_attn.q_proj.weight", "model.layers.24.self_attn.k_proj.weight", "model.layers.24.self_attn.v_proj.weight", "model.layers.24.self_attn.o_proj.weight", "model.layers.24.mlp.gate_proj.weight", "model.layers.24.mlp.up_proj.weight", "model.layers.24.mlp.down_proj.weight", "model.layers.24.input_layernorm.weight", "model.layers.24.post_attention_layernorm.weight", "model.layers.25.self_attn.q_proj.weight", "model.layers.25.self_attn.k_proj.weight", "model.layers.25.self_attn.v_proj.weight", "model.layers.25.self_attn.o_proj.weight", "model.layers.25.mlp.gate_proj.weight", "model.layers.25.mlp.up_proj.weight", "model.layers.25.mlp.down_proj.weight", "model.layers.25.input_layernorm.weight", "model.layers.25.post_attention_layernorm.weight", "model.layers.26.self_attn.q_proj.weight", "model.layers.26.self_attn.k_proj.weight", "model.layers.26.self_attn.v_proj.weight", "model.layers.26.self_attn.o_proj.weight", "model.layers.26.mlp.gate_proj.weight", "model.layers.26.mlp.up_proj.weight", "model.layers.26.mlp.down_proj.weight", "model.layers.26.input_layernorm.weight", "model.layers.26.post_attention_layernorm.weight", "model.layers.27.self_attn.q_proj.weight", "model.layers.27.self_attn.k_proj.weight", "model.layers.27.self_attn.v_proj.weight", "model.layers.27.self_attn.o_proj.weight", "model.layers.27.mlp.gate_proj.weight", "model.layers.27.mlp.up_proj.weight", "model.layers.27.mlp.down_proj.weight", "model.layers.27.input_layernorm.weight", "model.layers.27.post_attention_layernorm.weight", "model.layers.28.self_attn.q_proj.weight", "model.layers.28.self_attn.k_proj.weight", "model.layers.28.self_attn.v_proj.weight", "model.layers.28.self_attn.o_proj.weight", "model.layers.28.mlp.gate_proj.weight", "model.layers.28.mlp.up_proj.weight", "model.layers.28.mlp.down_proj.weight", "model.layers.28.input_layernorm.weight", "model.layers.28.post_attention_layernorm.weight", "model.layers.29.self_attn.q_proj.weight", "model.layers.29.self_attn.k_proj.weight", "model.layers.29.self_attn.v_proj.weight", "model.layers.29.self_attn.o_proj.weight", "model.layers.29.mlp.gate_proj.weight", "model.layers.29.mlp.up_proj.weight", "model.layers.29.mlp.down_proj.weight", "model.layers.29.input_layernorm.weight", "model.layers.29.post_attention_layernorm.weight", "model.layers.30.self_attn.q_proj.weight", "model.layers.30.self_attn.k_proj.weight", "model.layers.30.self_attn.v_proj.weight", "model.layers.30.self_attn.o_proj.weight", "model.layers.30.mlp.gate_proj.weight", "model.layers.30.mlp.up_proj.weight", "model.layers.30.mlp.down_proj.weight", "model.layers.30.input_layernorm.weight", "model.layers.30.post_attention_layernorm.weight", "model.layers.31.self_attn.q_proj.weight", "model.layers.31.self_attn.k_proj.weight", "model.layers.31.self_attn.v_proj.weight", "model.layers.31.self_attn.o_proj.weight", "model.layers.31.mlp.gate_proj.weight", "model.layers.31.mlp.up_proj.weight", "model.layers.31.mlp.down_proj.weight", "model.layers.31.input_layernorm.weight", "model.layers.31.post_attention_layernorm.weight", "model.norm.weight", "lm_head.weight" ] }