root commited on
Commit
f1bc1f0
·
1 Parent(s): e8fd46a

Signed-off-by: root <root@rack-ge107-ya-20250517193240.idczw.hb1.kwaidc.com>

model-00001-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fde24f4cea04717461df1cd21e38273038f8ffbdfa0f5cf2fe90771a47d3694a
3
- size 4932328296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c8245d49c65c1be03de3d97310f58951b25d867353e0f8409d2e25d5308b0a5
3
+ size 4932307584
model-00002-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3de75caf4a466d6fd92522304ab807e37ee99de2b11ecb9e5b11e56faafa526a
3
- size 4875989704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:025ccf9d44b268e397cb49caf5fd8836ddc0861c6920c8223ce84bd4dde5d07a
3
+ size 4875989696
model-00003-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f3991d2cbddc4f1dc33d3677284992363da669636e38629e14947ed0edfbf31
3
- size 4875989728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0a2b1d7d09da8e1f2007877ef78c6d60cfefee1f181a7a519a3a184405fc725
3
+ size 4875989720
model-00004-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:27324de98e4a78a731f64890aa46072476e4ccb5756764f1999a64bdc447b84f
3
- size 4875989760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e56acc07cb1c87d2d8576ecd0435e302951de7fbfca41f4f65ab32b367a167c
3
+ size 4875989752
model-00005-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c1061acf0b748e83ec0dfc17b8b8cbb2244c8f22282d5391d7682006d3f47a72
3
- size 4875989760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33d43d09da1b62b5fdad2f9134c11a7598573b39d4cc6daa62feaf6ce0d5146a
3
+ size 4875989752
model-00006-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:05f64b81a0e453905a76400be1c6132694fb4834373dfd31dad108f099a8dded
3
- size 4875989760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b180d65bfda25d31931a5887bec20e59cb8a3fd24d866bc1bfb6edf4fcd8283
3
+ size 4875989752
model-00007-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:531343b3a974527b51fb8e6af98c684ea515160c34c54fb6cec14d3e14f731fb
3
- size 4875989760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba71af62d1e7f8d3ff2fbcc7537adfe97223501b45bfb8802f3d40221123b9d3
3
+ size 4875989752
model-00008-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3b6394061a920b7fd1eb9adec4a89438118f7d4a545d9c3bbe31b030ebf41b47
3
- size 4875989760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3791cc769a06c3dd965b594d0ebed36026f97f3cf59ae859aa92ba3fd6781e6
3
+ size 4875989752
model-00009-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ab196998abe5e68d8b8bcc00e2eb5a8c582ae52cc4da18664d3a0f64b921943d
3
- size 4875989760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03556cfb7159e2f3999ef1610c5eb44456a29bb6328af26889d785c4d2f7f6de
3
+ size 4875989752
model-00010-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b018e919d2c29bee4a93c3403915fcc98530a2ce1c95f186adcfe52d57768a5c
3
- size 4875989760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5149f2d34a2e3bc4248ae89c9a046491066395254b9c35765b95257f0e1c2bdf
3
+ size 4875989752
model-00011-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c3ab15280b6c9575d65e5bc04501bf3f0287e3ec9062578c89865acc3432d1a6
3
- size 4875989760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb2b02a41b63208152df53fa5b9d55ab17a052b604051047d3058c3d54940dc9
3
+ size 4875989752
model-00012-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b68ad8f2254f6fa81129bfa17a2673444e93008504fc8c2735780819e5f21083
3
- size 4875989760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9104fdfb1cccaf14d2114a8b767578630c7cbe218e0fd1af14638bdbc9737f1
3
+ size 4875989752
model-00013-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:926f1467d0486592835c4cf4b10f68512639203c0e11c4555163ea44c6989084
3
- size 4875989760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01b02c17fd7ea0ff6ea9d836a601ab38a74b1c8593c4800dfdbe7c8392828cd1
3
+ size 4875989752
model-00014-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6302ce334d630c09cafdaab88d6a3197e3243d1616fda6327ff3b92e82328826
3
- size 2080123328
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4e83c27b30b91e83a2c0984e6ddd57baac2b0a2ce30ade1386ed2e534e8ff6c
3
+ size 2080144040
model.safetensors.index.json CHANGED
@@ -60,11 +60,11 @@
60
  "model.layers.12.self_attn.q_norm.weight": "model-00003-of-00014.safetensors",
61
  "model.layers.12.self_attn.q_proj.weight": "model-00003-of-00014.safetensors",
62
  "model.layers.12.self_attn.v_proj.weight": "model-00003-of-00014.safetensors",
63
- "model.layers.13.input_layernorm.weight": "model-00003-of-00014.safetensors",
64
  "model.layers.13.mlp.down_proj.weight": "model-00004-of-00014.safetensors",
65
  "model.layers.13.mlp.gate_proj.weight": "model-00003-of-00014.safetensors",
66
  "model.layers.13.mlp.up_proj.weight": "model-00004-of-00014.safetensors",
67
- "model.layers.13.post_attention_layernorm.weight": "model-00003-of-00014.safetensors",
68
  "model.layers.13.self_attn.k_norm.weight": "model-00003-of-00014.safetensors",
69
  "model.layers.13.self_attn.k_proj.weight": "model-00003-of-00014.safetensors",
70
  "model.layers.13.self_attn.o_proj.weight": "model-00003-of-00014.safetensors",
@@ -115,11 +115,11 @@
115
  "model.layers.17.self_attn.q_norm.weight": "model-00004-of-00014.safetensors",
116
  "model.layers.17.self_attn.q_proj.weight": "model-00004-of-00014.safetensors",
117
  "model.layers.17.self_attn.v_proj.weight": "model-00004-of-00014.safetensors",
118
- "model.layers.18.input_layernorm.weight": "model-00004-of-00014.safetensors",
119
  "model.layers.18.mlp.down_proj.weight": "model-00005-of-00014.safetensors",
120
  "model.layers.18.mlp.gate_proj.weight": "model-00004-of-00014.safetensors",
121
  "model.layers.18.mlp.up_proj.weight": "model-00005-of-00014.safetensors",
122
- "model.layers.18.post_attention_layernorm.weight": "model-00004-of-00014.safetensors",
123
  "model.layers.18.self_attn.k_norm.weight": "model-00004-of-00014.safetensors",
124
  "model.layers.18.self_attn.k_proj.weight": "model-00004-of-00014.safetensors",
125
  "model.layers.18.self_attn.o_proj.weight": "model-00004-of-00014.safetensors",
@@ -181,11 +181,11 @@
181
  "model.layers.22.self_attn.q_norm.weight": "model-00005-of-00014.safetensors",
182
  "model.layers.22.self_attn.q_proj.weight": "model-00005-of-00014.safetensors",
183
  "model.layers.22.self_attn.v_proj.weight": "model-00005-of-00014.safetensors",
184
- "model.layers.23.input_layernorm.weight": "model-00005-of-00014.safetensors",
185
  "model.layers.23.mlp.down_proj.weight": "model-00006-of-00014.safetensors",
186
  "model.layers.23.mlp.gate_proj.weight": "model-00005-of-00014.safetensors",
187
  "model.layers.23.mlp.up_proj.weight": "model-00006-of-00014.safetensors",
188
- "model.layers.23.post_attention_layernorm.weight": "model-00005-of-00014.safetensors",
189
  "model.layers.23.self_attn.k_norm.weight": "model-00005-of-00014.safetensors",
190
  "model.layers.23.self_attn.k_proj.weight": "model-00005-of-00014.safetensors",
191
  "model.layers.23.self_attn.o_proj.weight": "model-00005-of-00014.safetensors",
@@ -236,11 +236,11 @@
236
  "model.layers.27.self_attn.q_norm.weight": "model-00006-of-00014.safetensors",
237
  "model.layers.27.self_attn.q_proj.weight": "model-00006-of-00014.safetensors",
238
  "model.layers.27.self_attn.v_proj.weight": "model-00006-of-00014.safetensors",
239
- "model.layers.28.input_layernorm.weight": "model-00006-of-00014.safetensors",
240
  "model.layers.28.mlp.down_proj.weight": "model-00007-of-00014.safetensors",
241
  "model.layers.28.mlp.gate_proj.weight": "model-00006-of-00014.safetensors",
242
  "model.layers.28.mlp.up_proj.weight": "model-00007-of-00014.safetensors",
243
- "model.layers.28.post_attention_layernorm.weight": "model-00006-of-00014.safetensors",
244
  "model.layers.28.self_attn.k_norm.weight": "model-00006-of-00014.safetensors",
245
  "model.layers.28.self_attn.k_proj.weight": "model-00006-of-00014.safetensors",
246
  "model.layers.28.self_attn.o_proj.weight": "model-00006-of-00014.safetensors",
@@ -258,11 +258,11 @@
258
  "model.layers.29.self_attn.q_norm.weight": "model-00007-of-00014.safetensors",
259
  "model.layers.29.self_attn.q_proj.weight": "model-00007-of-00014.safetensors",
260
  "model.layers.29.self_attn.v_proj.weight": "model-00007-of-00014.safetensors",
261
- "model.layers.3.input_layernorm.weight": "model-00001-of-00014.safetensors",
262
  "model.layers.3.mlp.down_proj.weight": "model-00002-of-00014.safetensors",
263
  "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00014.safetensors",
264
  "model.layers.3.mlp.up_proj.weight": "model-00002-of-00014.safetensors",
265
- "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00014.safetensors",
266
  "model.layers.3.self_attn.k_norm.weight": "model-00001-of-00014.safetensors",
267
  "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00014.safetensors",
268
  "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00014.safetensors",
@@ -302,11 +302,11 @@
302
  "model.layers.32.self_attn.q_norm.weight": "model-00007-of-00014.safetensors",
303
  "model.layers.32.self_attn.q_proj.weight": "model-00007-of-00014.safetensors",
304
  "model.layers.32.self_attn.v_proj.weight": "model-00007-of-00014.safetensors",
305
- "model.layers.33.input_layernorm.weight": "model-00007-of-00014.safetensors",
306
  "model.layers.33.mlp.down_proj.weight": "model-00008-of-00014.safetensors",
307
  "model.layers.33.mlp.gate_proj.weight": "model-00007-of-00014.safetensors",
308
  "model.layers.33.mlp.up_proj.weight": "model-00008-of-00014.safetensors",
309
- "model.layers.33.post_attention_layernorm.weight": "model-00007-of-00014.safetensors",
310
  "model.layers.33.self_attn.k_norm.weight": "model-00007-of-00014.safetensors",
311
  "model.layers.33.self_attn.k_proj.weight": "model-00007-of-00014.safetensors",
312
  "model.layers.33.self_attn.o_proj.weight": "model-00007-of-00014.safetensors",
@@ -357,11 +357,11 @@
357
  "model.layers.37.self_attn.q_norm.weight": "model-00008-of-00014.safetensors",
358
  "model.layers.37.self_attn.q_proj.weight": "model-00008-of-00014.safetensors",
359
  "model.layers.37.self_attn.v_proj.weight": "model-00008-of-00014.safetensors",
360
- "model.layers.38.input_layernorm.weight": "model-00008-of-00014.safetensors",
361
  "model.layers.38.mlp.down_proj.weight": "model-00009-of-00014.safetensors",
362
  "model.layers.38.mlp.gate_proj.weight": "model-00008-of-00014.safetensors",
363
  "model.layers.38.mlp.up_proj.weight": "model-00009-of-00014.safetensors",
364
- "model.layers.38.post_attention_layernorm.weight": "model-00008-of-00014.safetensors",
365
  "model.layers.38.self_attn.k_norm.weight": "model-00008-of-00014.safetensors",
366
  "model.layers.38.self_attn.k_proj.weight": "model-00008-of-00014.safetensors",
367
  "model.layers.38.self_attn.o_proj.weight": "model-00008-of-00014.safetensors",
@@ -423,11 +423,11 @@
423
  "model.layers.42.self_attn.q_norm.weight": "model-00009-of-00014.safetensors",
424
  "model.layers.42.self_attn.q_proj.weight": "model-00009-of-00014.safetensors",
425
  "model.layers.42.self_attn.v_proj.weight": "model-00009-of-00014.safetensors",
426
- "model.layers.43.input_layernorm.weight": "model-00009-of-00014.safetensors",
427
  "model.layers.43.mlp.down_proj.weight": "model-00010-of-00014.safetensors",
428
  "model.layers.43.mlp.gate_proj.weight": "model-00009-of-00014.safetensors",
429
  "model.layers.43.mlp.up_proj.weight": "model-00010-of-00014.safetensors",
430
- "model.layers.43.post_attention_layernorm.weight": "model-00009-of-00014.safetensors",
431
  "model.layers.43.self_attn.k_norm.weight": "model-00009-of-00014.safetensors",
432
  "model.layers.43.self_attn.k_proj.weight": "model-00009-of-00014.safetensors",
433
  "model.layers.43.self_attn.o_proj.weight": "model-00009-of-00014.safetensors",
@@ -478,11 +478,11 @@
478
  "model.layers.47.self_attn.q_norm.weight": "model-00010-of-00014.safetensors",
479
  "model.layers.47.self_attn.q_proj.weight": "model-00010-of-00014.safetensors",
480
  "model.layers.47.self_attn.v_proj.weight": "model-00010-of-00014.safetensors",
481
- "model.layers.48.input_layernorm.weight": "model-00010-of-00014.safetensors",
482
  "model.layers.48.mlp.down_proj.weight": "model-00011-of-00014.safetensors",
483
  "model.layers.48.mlp.gate_proj.weight": "model-00010-of-00014.safetensors",
484
  "model.layers.48.mlp.up_proj.weight": "model-00011-of-00014.safetensors",
485
- "model.layers.48.post_attention_layernorm.weight": "model-00010-of-00014.safetensors",
486
  "model.layers.48.self_attn.k_norm.weight": "model-00010-of-00014.safetensors",
487
  "model.layers.48.self_attn.k_proj.weight": "model-00010-of-00014.safetensors",
488
  "model.layers.48.self_attn.o_proj.weight": "model-00010-of-00014.safetensors",
@@ -544,11 +544,11 @@
544
  "model.layers.52.self_attn.q_norm.weight": "model-00011-of-00014.safetensors",
545
  "model.layers.52.self_attn.q_proj.weight": "model-00011-of-00014.safetensors",
546
  "model.layers.52.self_attn.v_proj.weight": "model-00011-of-00014.safetensors",
547
- "model.layers.53.input_layernorm.weight": "model-00011-of-00014.safetensors",
548
  "model.layers.53.mlp.down_proj.weight": "model-00012-of-00014.safetensors",
549
  "model.layers.53.mlp.gate_proj.weight": "model-00011-of-00014.safetensors",
550
  "model.layers.53.mlp.up_proj.weight": "model-00012-of-00014.safetensors",
551
- "model.layers.53.post_attention_layernorm.weight": "model-00011-of-00014.safetensors",
552
  "model.layers.53.self_attn.k_norm.weight": "model-00011-of-00014.safetensors",
553
  "model.layers.53.self_attn.k_proj.weight": "model-00011-of-00014.safetensors",
554
  "model.layers.53.self_attn.o_proj.weight": "model-00011-of-00014.safetensors",
@@ -599,11 +599,11 @@
599
  "model.layers.57.self_attn.q_norm.weight": "model-00012-of-00014.safetensors",
600
  "model.layers.57.self_attn.q_proj.weight": "model-00012-of-00014.safetensors",
601
  "model.layers.57.self_attn.v_proj.weight": "model-00012-of-00014.safetensors",
602
- "model.layers.58.input_layernorm.weight": "model-00012-of-00014.safetensors",
603
  "model.layers.58.mlp.down_proj.weight": "model-00013-of-00014.safetensors",
604
  "model.layers.58.mlp.gate_proj.weight": "model-00012-of-00014.safetensors",
605
  "model.layers.58.mlp.up_proj.weight": "model-00013-of-00014.safetensors",
606
- "model.layers.58.post_attention_layernorm.weight": "model-00012-of-00014.safetensors",
607
  "model.layers.58.self_attn.k_norm.weight": "model-00012-of-00014.safetensors",
608
  "model.layers.58.self_attn.k_proj.weight": "model-00012-of-00014.safetensors",
609
  "model.layers.58.self_attn.o_proj.weight": "model-00012-of-00014.safetensors",
@@ -665,11 +665,11 @@
665
  "model.layers.62.self_attn.q_norm.weight": "model-00013-of-00014.safetensors",
666
  "model.layers.62.self_attn.q_proj.weight": "model-00013-of-00014.safetensors",
667
  "model.layers.62.self_attn.v_proj.weight": "model-00013-of-00014.safetensors",
668
- "model.layers.63.input_layernorm.weight": "model-00013-of-00014.safetensors",
669
  "model.layers.63.mlp.down_proj.weight": "model-00014-of-00014.safetensors",
670
  "model.layers.63.mlp.gate_proj.weight": "model-00013-of-00014.safetensors",
671
  "model.layers.63.mlp.up_proj.weight": "model-00014-of-00014.safetensors",
672
- "model.layers.63.post_attention_layernorm.weight": "model-00013-of-00014.safetensors",
673
  "model.layers.63.self_attn.k_norm.weight": "model-00013-of-00014.safetensors",
674
  "model.layers.63.self_attn.k_proj.weight": "model-00013-of-00014.safetensors",
675
  "model.layers.63.self_attn.o_proj.weight": "model-00013-of-00014.safetensors",
@@ -687,11 +687,11 @@
687
  "model.layers.7.self_attn.q_norm.weight": "model-00002-of-00014.safetensors",
688
  "model.layers.7.self_attn.q_proj.weight": "model-00002-of-00014.safetensors",
689
  "model.layers.7.self_attn.v_proj.weight": "model-00002-of-00014.safetensors",
690
- "model.layers.8.input_layernorm.weight": "model-00002-of-00014.safetensors",
691
  "model.layers.8.mlp.down_proj.weight": "model-00003-of-00014.safetensors",
692
  "model.layers.8.mlp.gate_proj.weight": "model-00002-of-00014.safetensors",
693
  "model.layers.8.mlp.up_proj.weight": "model-00003-of-00014.safetensors",
694
- "model.layers.8.post_attention_layernorm.weight": "model-00002-of-00014.safetensors",
695
  "model.layers.8.self_attn.k_norm.weight": "model-00002-of-00014.safetensors",
696
  "model.layers.8.self_attn.k_proj.weight": "model-00002-of-00014.safetensors",
697
  "model.layers.8.self_attn.o_proj.weight": "model-00002-of-00014.safetensors",
 
60
  "model.layers.12.self_attn.q_norm.weight": "model-00003-of-00014.safetensors",
61
  "model.layers.12.self_attn.q_proj.weight": "model-00003-of-00014.safetensors",
62
  "model.layers.12.self_attn.v_proj.weight": "model-00003-of-00014.safetensors",
63
+ "model.layers.13.input_layernorm.weight": "model-00004-of-00014.safetensors",
64
  "model.layers.13.mlp.down_proj.weight": "model-00004-of-00014.safetensors",
65
  "model.layers.13.mlp.gate_proj.weight": "model-00003-of-00014.safetensors",
66
  "model.layers.13.mlp.up_proj.weight": "model-00004-of-00014.safetensors",
67
+ "model.layers.13.post_attention_layernorm.weight": "model-00004-of-00014.safetensors",
68
  "model.layers.13.self_attn.k_norm.weight": "model-00003-of-00014.safetensors",
69
  "model.layers.13.self_attn.k_proj.weight": "model-00003-of-00014.safetensors",
70
  "model.layers.13.self_attn.o_proj.weight": "model-00003-of-00014.safetensors",
 
115
  "model.layers.17.self_attn.q_norm.weight": "model-00004-of-00014.safetensors",
116
  "model.layers.17.self_attn.q_proj.weight": "model-00004-of-00014.safetensors",
117
  "model.layers.17.self_attn.v_proj.weight": "model-00004-of-00014.safetensors",
118
+ "model.layers.18.input_layernorm.weight": "model-00005-of-00014.safetensors",
119
  "model.layers.18.mlp.down_proj.weight": "model-00005-of-00014.safetensors",
120
  "model.layers.18.mlp.gate_proj.weight": "model-00004-of-00014.safetensors",
121
  "model.layers.18.mlp.up_proj.weight": "model-00005-of-00014.safetensors",
122
+ "model.layers.18.post_attention_layernorm.weight": "model-00005-of-00014.safetensors",
123
  "model.layers.18.self_attn.k_norm.weight": "model-00004-of-00014.safetensors",
124
  "model.layers.18.self_attn.k_proj.weight": "model-00004-of-00014.safetensors",
125
  "model.layers.18.self_attn.o_proj.weight": "model-00004-of-00014.safetensors",
 
181
  "model.layers.22.self_attn.q_norm.weight": "model-00005-of-00014.safetensors",
182
  "model.layers.22.self_attn.q_proj.weight": "model-00005-of-00014.safetensors",
183
  "model.layers.22.self_attn.v_proj.weight": "model-00005-of-00014.safetensors",
184
+ "model.layers.23.input_layernorm.weight": "model-00006-of-00014.safetensors",
185
  "model.layers.23.mlp.down_proj.weight": "model-00006-of-00014.safetensors",
186
  "model.layers.23.mlp.gate_proj.weight": "model-00005-of-00014.safetensors",
187
  "model.layers.23.mlp.up_proj.weight": "model-00006-of-00014.safetensors",
188
+ "model.layers.23.post_attention_layernorm.weight": "model-00006-of-00014.safetensors",
189
  "model.layers.23.self_attn.k_norm.weight": "model-00005-of-00014.safetensors",
190
  "model.layers.23.self_attn.k_proj.weight": "model-00005-of-00014.safetensors",
191
  "model.layers.23.self_attn.o_proj.weight": "model-00005-of-00014.safetensors",
 
236
  "model.layers.27.self_attn.q_norm.weight": "model-00006-of-00014.safetensors",
237
  "model.layers.27.self_attn.q_proj.weight": "model-00006-of-00014.safetensors",
238
  "model.layers.27.self_attn.v_proj.weight": "model-00006-of-00014.safetensors",
239
+ "model.layers.28.input_layernorm.weight": "model-00007-of-00014.safetensors",
240
  "model.layers.28.mlp.down_proj.weight": "model-00007-of-00014.safetensors",
241
  "model.layers.28.mlp.gate_proj.weight": "model-00006-of-00014.safetensors",
242
  "model.layers.28.mlp.up_proj.weight": "model-00007-of-00014.safetensors",
243
+ "model.layers.28.post_attention_layernorm.weight": "model-00007-of-00014.safetensors",
244
  "model.layers.28.self_attn.k_norm.weight": "model-00006-of-00014.safetensors",
245
  "model.layers.28.self_attn.k_proj.weight": "model-00006-of-00014.safetensors",
246
  "model.layers.28.self_attn.o_proj.weight": "model-00006-of-00014.safetensors",
 
258
  "model.layers.29.self_attn.q_norm.weight": "model-00007-of-00014.safetensors",
259
  "model.layers.29.self_attn.q_proj.weight": "model-00007-of-00014.safetensors",
260
  "model.layers.29.self_attn.v_proj.weight": "model-00007-of-00014.safetensors",
261
+ "model.layers.3.input_layernorm.weight": "model-00002-of-00014.safetensors",
262
  "model.layers.3.mlp.down_proj.weight": "model-00002-of-00014.safetensors",
263
  "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00014.safetensors",
264
  "model.layers.3.mlp.up_proj.weight": "model-00002-of-00014.safetensors",
265
+ "model.layers.3.post_attention_layernorm.weight": "model-00002-of-00014.safetensors",
266
  "model.layers.3.self_attn.k_norm.weight": "model-00001-of-00014.safetensors",
267
  "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00014.safetensors",
268
  "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00014.safetensors",
 
302
  "model.layers.32.self_attn.q_norm.weight": "model-00007-of-00014.safetensors",
303
  "model.layers.32.self_attn.q_proj.weight": "model-00007-of-00014.safetensors",
304
  "model.layers.32.self_attn.v_proj.weight": "model-00007-of-00014.safetensors",
305
+ "model.layers.33.input_layernorm.weight": "model-00008-of-00014.safetensors",
306
  "model.layers.33.mlp.down_proj.weight": "model-00008-of-00014.safetensors",
307
  "model.layers.33.mlp.gate_proj.weight": "model-00007-of-00014.safetensors",
308
  "model.layers.33.mlp.up_proj.weight": "model-00008-of-00014.safetensors",
309
+ "model.layers.33.post_attention_layernorm.weight": "model-00008-of-00014.safetensors",
310
  "model.layers.33.self_attn.k_norm.weight": "model-00007-of-00014.safetensors",
311
  "model.layers.33.self_attn.k_proj.weight": "model-00007-of-00014.safetensors",
312
  "model.layers.33.self_attn.o_proj.weight": "model-00007-of-00014.safetensors",
 
357
  "model.layers.37.self_attn.q_norm.weight": "model-00008-of-00014.safetensors",
358
  "model.layers.37.self_attn.q_proj.weight": "model-00008-of-00014.safetensors",
359
  "model.layers.37.self_attn.v_proj.weight": "model-00008-of-00014.safetensors",
360
+ "model.layers.38.input_layernorm.weight": "model-00009-of-00014.safetensors",
361
  "model.layers.38.mlp.down_proj.weight": "model-00009-of-00014.safetensors",
362
  "model.layers.38.mlp.gate_proj.weight": "model-00008-of-00014.safetensors",
363
  "model.layers.38.mlp.up_proj.weight": "model-00009-of-00014.safetensors",
364
+ "model.layers.38.post_attention_layernorm.weight": "model-00009-of-00014.safetensors",
365
  "model.layers.38.self_attn.k_norm.weight": "model-00008-of-00014.safetensors",
366
  "model.layers.38.self_attn.k_proj.weight": "model-00008-of-00014.safetensors",
367
  "model.layers.38.self_attn.o_proj.weight": "model-00008-of-00014.safetensors",
 
423
  "model.layers.42.self_attn.q_norm.weight": "model-00009-of-00014.safetensors",
424
  "model.layers.42.self_attn.q_proj.weight": "model-00009-of-00014.safetensors",
425
  "model.layers.42.self_attn.v_proj.weight": "model-00009-of-00014.safetensors",
426
+ "model.layers.43.input_layernorm.weight": "model-00010-of-00014.safetensors",
427
  "model.layers.43.mlp.down_proj.weight": "model-00010-of-00014.safetensors",
428
  "model.layers.43.mlp.gate_proj.weight": "model-00009-of-00014.safetensors",
429
  "model.layers.43.mlp.up_proj.weight": "model-00010-of-00014.safetensors",
430
+ "model.layers.43.post_attention_layernorm.weight": "model-00010-of-00014.safetensors",
431
  "model.layers.43.self_attn.k_norm.weight": "model-00009-of-00014.safetensors",
432
  "model.layers.43.self_attn.k_proj.weight": "model-00009-of-00014.safetensors",
433
  "model.layers.43.self_attn.o_proj.weight": "model-00009-of-00014.safetensors",
 
478
  "model.layers.47.self_attn.q_norm.weight": "model-00010-of-00014.safetensors",
479
  "model.layers.47.self_attn.q_proj.weight": "model-00010-of-00014.safetensors",
480
  "model.layers.47.self_attn.v_proj.weight": "model-00010-of-00014.safetensors",
481
+ "model.layers.48.input_layernorm.weight": "model-00011-of-00014.safetensors",
482
  "model.layers.48.mlp.down_proj.weight": "model-00011-of-00014.safetensors",
483
  "model.layers.48.mlp.gate_proj.weight": "model-00010-of-00014.safetensors",
484
  "model.layers.48.mlp.up_proj.weight": "model-00011-of-00014.safetensors",
485
+ "model.layers.48.post_attention_layernorm.weight": "model-00011-of-00014.safetensors",
486
  "model.layers.48.self_attn.k_norm.weight": "model-00010-of-00014.safetensors",
487
  "model.layers.48.self_attn.k_proj.weight": "model-00010-of-00014.safetensors",
488
  "model.layers.48.self_attn.o_proj.weight": "model-00010-of-00014.safetensors",
 
544
  "model.layers.52.self_attn.q_norm.weight": "model-00011-of-00014.safetensors",
545
  "model.layers.52.self_attn.q_proj.weight": "model-00011-of-00014.safetensors",
546
  "model.layers.52.self_attn.v_proj.weight": "model-00011-of-00014.safetensors",
547
+ "model.layers.53.input_layernorm.weight": "model-00012-of-00014.safetensors",
548
  "model.layers.53.mlp.down_proj.weight": "model-00012-of-00014.safetensors",
549
  "model.layers.53.mlp.gate_proj.weight": "model-00011-of-00014.safetensors",
550
  "model.layers.53.mlp.up_proj.weight": "model-00012-of-00014.safetensors",
551
+ "model.layers.53.post_attention_layernorm.weight": "model-00012-of-00014.safetensors",
552
  "model.layers.53.self_attn.k_norm.weight": "model-00011-of-00014.safetensors",
553
  "model.layers.53.self_attn.k_proj.weight": "model-00011-of-00014.safetensors",
554
  "model.layers.53.self_attn.o_proj.weight": "model-00011-of-00014.safetensors",
 
599
  "model.layers.57.self_attn.q_norm.weight": "model-00012-of-00014.safetensors",
600
  "model.layers.57.self_attn.q_proj.weight": "model-00012-of-00014.safetensors",
601
  "model.layers.57.self_attn.v_proj.weight": "model-00012-of-00014.safetensors",
602
+ "model.layers.58.input_layernorm.weight": "model-00013-of-00014.safetensors",
603
  "model.layers.58.mlp.down_proj.weight": "model-00013-of-00014.safetensors",
604
  "model.layers.58.mlp.gate_proj.weight": "model-00012-of-00014.safetensors",
605
  "model.layers.58.mlp.up_proj.weight": "model-00013-of-00014.safetensors",
606
+ "model.layers.58.post_attention_layernorm.weight": "model-00013-of-00014.safetensors",
607
  "model.layers.58.self_attn.k_norm.weight": "model-00012-of-00014.safetensors",
608
  "model.layers.58.self_attn.k_proj.weight": "model-00012-of-00014.safetensors",
609
  "model.layers.58.self_attn.o_proj.weight": "model-00012-of-00014.safetensors",
 
665
  "model.layers.62.self_attn.q_norm.weight": "model-00013-of-00014.safetensors",
666
  "model.layers.62.self_attn.q_proj.weight": "model-00013-of-00014.safetensors",
667
  "model.layers.62.self_attn.v_proj.weight": "model-00013-of-00014.safetensors",
668
+ "model.layers.63.input_layernorm.weight": "model-00014-of-00014.safetensors",
669
  "model.layers.63.mlp.down_proj.weight": "model-00014-of-00014.safetensors",
670
  "model.layers.63.mlp.gate_proj.weight": "model-00013-of-00014.safetensors",
671
  "model.layers.63.mlp.up_proj.weight": "model-00014-of-00014.safetensors",
672
+ "model.layers.63.post_attention_layernorm.weight": "model-00014-of-00014.safetensors",
673
  "model.layers.63.self_attn.k_norm.weight": "model-00013-of-00014.safetensors",
674
  "model.layers.63.self_attn.k_proj.weight": "model-00013-of-00014.safetensors",
675
  "model.layers.63.self_attn.o_proj.weight": "model-00013-of-00014.safetensors",
 
687
  "model.layers.7.self_attn.q_norm.weight": "model-00002-of-00014.safetensors",
688
  "model.layers.7.self_attn.q_proj.weight": "model-00002-of-00014.safetensors",
689
  "model.layers.7.self_attn.v_proj.weight": "model-00002-of-00014.safetensors",
690
+ "model.layers.8.input_layernorm.weight": "model-00003-of-00014.safetensors",
691
  "model.layers.8.mlp.down_proj.weight": "model-00003-of-00014.safetensors",
692
  "model.layers.8.mlp.gate_proj.weight": "model-00002-of-00014.safetensors",
693
  "model.layers.8.mlp.up_proj.weight": "model-00003-of-00014.safetensors",
694
+ "model.layers.8.post_attention_layernorm.weight": "model-00003-of-00014.safetensors",
695
  "model.layers.8.self_attn.k_norm.weight": "model-00002-of-00014.safetensors",
696
  "model.layers.8.self_attn.k_proj.weight": "model-00002-of-00014.safetensors",
697
  "model.layers.8.self_attn.o_proj.weight": "model-00002-of-00014.safetensors",