root
commited on
Commit
·
f1bc1f0
1
Parent(s):
e8fd46a
update
Browse filesSigned-off-by: root <root@rack-ge107-ya-20250517193240.idczw.hb1.kwaidc.com>
- model-00001-of-00014.safetensors +2 -2
- model-00002-of-00014.safetensors +2 -2
- model-00003-of-00014.safetensors +2 -2
- model-00004-of-00014.safetensors +2 -2
- model-00005-of-00014.safetensors +2 -2
- model-00006-of-00014.safetensors +2 -2
- model-00007-of-00014.safetensors +2 -2
- model-00008-of-00014.safetensors +2 -2
- model-00009-of-00014.safetensors +2 -2
- model-00010-of-00014.safetensors +2 -2
- model-00011-of-00014.safetensors +2 -2
- model-00012-of-00014.safetensors +2 -2
- model-00013-of-00014.safetensors +2 -2
- model-00014-of-00014.safetensors +2 -2
- model.safetensors.index.json +26 -26
model-00001-of-00014.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5c8245d49c65c1be03de3d97310f58951b25d867353e0f8409d2e25d5308b0a5
|
3 |
+
size 4932307584
|
model-00002-of-00014.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:025ccf9d44b268e397cb49caf5fd8836ddc0861c6920c8223ce84bd4dde5d07a
|
3 |
+
size 4875989696
|
model-00003-of-00014.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c0a2b1d7d09da8e1f2007877ef78c6d60cfefee1f181a7a519a3a184405fc725
|
3 |
+
size 4875989720
|
model-00004-of-00014.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0e56acc07cb1c87d2d8576ecd0435e302951de7fbfca41f4f65ab32b367a167c
|
3 |
+
size 4875989752
|
model-00005-of-00014.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:33d43d09da1b62b5fdad2f9134c11a7598573b39d4cc6daa62feaf6ce0d5146a
|
3 |
+
size 4875989752
|
model-00006-of-00014.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5b180d65bfda25d31931a5887bec20e59cb8a3fd24d866bc1bfb6edf4fcd8283
|
3 |
+
size 4875989752
|
model-00007-of-00014.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba71af62d1e7f8d3ff2fbcc7537adfe97223501b45bfb8802f3d40221123b9d3
|
3 |
+
size 4875989752
|
model-00008-of-00014.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e3791cc769a06c3dd965b594d0ebed36026f97f3cf59ae859aa92ba3fd6781e6
|
3 |
+
size 4875989752
|
model-00009-of-00014.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:03556cfb7159e2f3999ef1610c5eb44456a29bb6328af26889d785c4d2f7f6de
|
3 |
+
size 4875989752
|
model-00010-of-00014.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5149f2d34a2e3bc4248ae89c9a046491066395254b9c35765b95257f0e1c2bdf
|
3 |
+
size 4875989752
|
model-00011-of-00014.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bb2b02a41b63208152df53fa5b9d55ab17a052b604051047d3058c3d54940dc9
|
3 |
+
size 4875989752
|
model-00012-of-00014.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a9104fdfb1cccaf14d2114a8b767578630c7cbe218e0fd1af14638bdbc9737f1
|
3 |
+
size 4875989752
|
model-00013-of-00014.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:01b02c17fd7ea0ff6ea9d836a601ab38a74b1c8593c4800dfdbe7c8392828cd1
|
3 |
+
size 4875989752
|
model-00014-of-00014.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e4e83c27b30b91e83a2c0984e6ddd57baac2b0a2ce30ade1386ed2e534e8ff6c
|
3 |
+
size 2080144040
|
model.safetensors.index.json
CHANGED
@@ -60,11 +60,11 @@
|
|
60 |
"model.layers.12.self_attn.q_norm.weight": "model-00003-of-00014.safetensors",
|
61 |
"model.layers.12.self_attn.q_proj.weight": "model-00003-of-00014.safetensors",
|
62 |
"model.layers.12.self_attn.v_proj.weight": "model-00003-of-00014.safetensors",
|
63 |
-
"model.layers.13.input_layernorm.weight": "model-
|
64 |
"model.layers.13.mlp.down_proj.weight": "model-00004-of-00014.safetensors",
|
65 |
"model.layers.13.mlp.gate_proj.weight": "model-00003-of-00014.safetensors",
|
66 |
"model.layers.13.mlp.up_proj.weight": "model-00004-of-00014.safetensors",
|
67 |
-
"model.layers.13.post_attention_layernorm.weight": "model-
|
68 |
"model.layers.13.self_attn.k_norm.weight": "model-00003-of-00014.safetensors",
|
69 |
"model.layers.13.self_attn.k_proj.weight": "model-00003-of-00014.safetensors",
|
70 |
"model.layers.13.self_attn.o_proj.weight": "model-00003-of-00014.safetensors",
|
@@ -115,11 +115,11 @@
|
|
115 |
"model.layers.17.self_attn.q_norm.weight": "model-00004-of-00014.safetensors",
|
116 |
"model.layers.17.self_attn.q_proj.weight": "model-00004-of-00014.safetensors",
|
117 |
"model.layers.17.self_attn.v_proj.weight": "model-00004-of-00014.safetensors",
|
118 |
-
"model.layers.18.input_layernorm.weight": "model-
|
119 |
"model.layers.18.mlp.down_proj.weight": "model-00005-of-00014.safetensors",
|
120 |
"model.layers.18.mlp.gate_proj.weight": "model-00004-of-00014.safetensors",
|
121 |
"model.layers.18.mlp.up_proj.weight": "model-00005-of-00014.safetensors",
|
122 |
-
"model.layers.18.post_attention_layernorm.weight": "model-
|
123 |
"model.layers.18.self_attn.k_norm.weight": "model-00004-of-00014.safetensors",
|
124 |
"model.layers.18.self_attn.k_proj.weight": "model-00004-of-00014.safetensors",
|
125 |
"model.layers.18.self_attn.o_proj.weight": "model-00004-of-00014.safetensors",
|
@@ -181,11 +181,11 @@
|
|
181 |
"model.layers.22.self_attn.q_norm.weight": "model-00005-of-00014.safetensors",
|
182 |
"model.layers.22.self_attn.q_proj.weight": "model-00005-of-00014.safetensors",
|
183 |
"model.layers.22.self_attn.v_proj.weight": "model-00005-of-00014.safetensors",
|
184 |
-
"model.layers.23.input_layernorm.weight": "model-
|
185 |
"model.layers.23.mlp.down_proj.weight": "model-00006-of-00014.safetensors",
|
186 |
"model.layers.23.mlp.gate_proj.weight": "model-00005-of-00014.safetensors",
|
187 |
"model.layers.23.mlp.up_proj.weight": "model-00006-of-00014.safetensors",
|
188 |
-
"model.layers.23.post_attention_layernorm.weight": "model-
|
189 |
"model.layers.23.self_attn.k_norm.weight": "model-00005-of-00014.safetensors",
|
190 |
"model.layers.23.self_attn.k_proj.weight": "model-00005-of-00014.safetensors",
|
191 |
"model.layers.23.self_attn.o_proj.weight": "model-00005-of-00014.safetensors",
|
@@ -236,11 +236,11 @@
|
|
236 |
"model.layers.27.self_attn.q_norm.weight": "model-00006-of-00014.safetensors",
|
237 |
"model.layers.27.self_attn.q_proj.weight": "model-00006-of-00014.safetensors",
|
238 |
"model.layers.27.self_attn.v_proj.weight": "model-00006-of-00014.safetensors",
|
239 |
-
"model.layers.28.input_layernorm.weight": "model-
|
240 |
"model.layers.28.mlp.down_proj.weight": "model-00007-of-00014.safetensors",
|
241 |
"model.layers.28.mlp.gate_proj.weight": "model-00006-of-00014.safetensors",
|
242 |
"model.layers.28.mlp.up_proj.weight": "model-00007-of-00014.safetensors",
|
243 |
-
"model.layers.28.post_attention_layernorm.weight": "model-
|
244 |
"model.layers.28.self_attn.k_norm.weight": "model-00006-of-00014.safetensors",
|
245 |
"model.layers.28.self_attn.k_proj.weight": "model-00006-of-00014.safetensors",
|
246 |
"model.layers.28.self_attn.o_proj.weight": "model-00006-of-00014.safetensors",
|
@@ -258,11 +258,11 @@
|
|
258 |
"model.layers.29.self_attn.q_norm.weight": "model-00007-of-00014.safetensors",
|
259 |
"model.layers.29.self_attn.q_proj.weight": "model-00007-of-00014.safetensors",
|
260 |
"model.layers.29.self_attn.v_proj.weight": "model-00007-of-00014.safetensors",
|
261 |
-
"model.layers.3.input_layernorm.weight": "model-
|
262 |
"model.layers.3.mlp.down_proj.weight": "model-00002-of-00014.safetensors",
|
263 |
"model.layers.3.mlp.gate_proj.weight": "model-00001-of-00014.safetensors",
|
264 |
"model.layers.3.mlp.up_proj.weight": "model-00002-of-00014.safetensors",
|
265 |
-
"model.layers.3.post_attention_layernorm.weight": "model-
|
266 |
"model.layers.3.self_attn.k_norm.weight": "model-00001-of-00014.safetensors",
|
267 |
"model.layers.3.self_attn.k_proj.weight": "model-00001-of-00014.safetensors",
|
268 |
"model.layers.3.self_attn.o_proj.weight": "model-00001-of-00014.safetensors",
|
@@ -302,11 +302,11 @@
|
|
302 |
"model.layers.32.self_attn.q_norm.weight": "model-00007-of-00014.safetensors",
|
303 |
"model.layers.32.self_attn.q_proj.weight": "model-00007-of-00014.safetensors",
|
304 |
"model.layers.32.self_attn.v_proj.weight": "model-00007-of-00014.safetensors",
|
305 |
-
"model.layers.33.input_layernorm.weight": "model-
|
306 |
"model.layers.33.mlp.down_proj.weight": "model-00008-of-00014.safetensors",
|
307 |
"model.layers.33.mlp.gate_proj.weight": "model-00007-of-00014.safetensors",
|
308 |
"model.layers.33.mlp.up_proj.weight": "model-00008-of-00014.safetensors",
|
309 |
-
"model.layers.33.post_attention_layernorm.weight": "model-
|
310 |
"model.layers.33.self_attn.k_norm.weight": "model-00007-of-00014.safetensors",
|
311 |
"model.layers.33.self_attn.k_proj.weight": "model-00007-of-00014.safetensors",
|
312 |
"model.layers.33.self_attn.o_proj.weight": "model-00007-of-00014.safetensors",
|
@@ -357,11 +357,11 @@
|
|
357 |
"model.layers.37.self_attn.q_norm.weight": "model-00008-of-00014.safetensors",
|
358 |
"model.layers.37.self_attn.q_proj.weight": "model-00008-of-00014.safetensors",
|
359 |
"model.layers.37.self_attn.v_proj.weight": "model-00008-of-00014.safetensors",
|
360 |
-
"model.layers.38.input_layernorm.weight": "model-
|
361 |
"model.layers.38.mlp.down_proj.weight": "model-00009-of-00014.safetensors",
|
362 |
"model.layers.38.mlp.gate_proj.weight": "model-00008-of-00014.safetensors",
|
363 |
"model.layers.38.mlp.up_proj.weight": "model-00009-of-00014.safetensors",
|
364 |
-
"model.layers.38.post_attention_layernorm.weight": "model-
|
365 |
"model.layers.38.self_attn.k_norm.weight": "model-00008-of-00014.safetensors",
|
366 |
"model.layers.38.self_attn.k_proj.weight": "model-00008-of-00014.safetensors",
|
367 |
"model.layers.38.self_attn.o_proj.weight": "model-00008-of-00014.safetensors",
|
@@ -423,11 +423,11 @@
|
|
423 |
"model.layers.42.self_attn.q_norm.weight": "model-00009-of-00014.safetensors",
|
424 |
"model.layers.42.self_attn.q_proj.weight": "model-00009-of-00014.safetensors",
|
425 |
"model.layers.42.self_attn.v_proj.weight": "model-00009-of-00014.safetensors",
|
426 |
-
"model.layers.43.input_layernorm.weight": "model-
|
427 |
"model.layers.43.mlp.down_proj.weight": "model-00010-of-00014.safetensors",
|
428 |
"model.layers.43.mlp.gate_proj.weight": "model-00009-of-00014.safetensors",
|
429 |
"model.layers.43.mlp.up_proj.weight": "model-00010-of-00014.safetensors",
|
430 |
-
"model.layers.43.post_attention_layernorm.weight": "model-
|
431 |
"model.layers.43.self_attn.k_norm.weight": "model-00009-of-00014.safetensors",
|
432 |
"model.layers.43.self_attn.k_proj.weight": "model-00009-of-00014.safetensors",
|
433 |
"model.layers.43.self_attn.o_proj.weight": "model-00009-of-00014.safetensors",
|
@@ -478,11 +478,11 @@
|
|
478 |
"model.layers.47.self_attn.q_norm.weight": "model-00010-of-00014.safetensors",
|
479 |
"model.layers.47.self_attn.q_proj.weight": "model-00010-of-00014.safetensors",
|
480 |
"model.layers.47.self_attn.v_proj.weight": "model-00010-of-00014.safetensors",
|
481 |
-
"model.layers.48.input_layernorm.weight": "model-
|
482 |
"model.layers.48.mlp.down_proj.weight": "model-00011-of-00014.safetensors",
|
483 |
"model.layers.48.mlp.gate_proj.weight": "model-00010-of-00014.safetensors",
|
484 |
"model.layers.48.mlp.up_proj.weight": "model-00011-of-00014.safetensors",
|
485 |
-
"model.layers.48.post_attention_layernorm.weight": "model-
|
486 |
"model.layers.48.self_attn.k_norm.weight": "model-00010-of-00014.safetensors",
|
487 |
"model.layers.48.self_attn.k_proj.weight": "model-00010-of-00014.safetensors",
|
488 |
"model.layers.48.self_attn.o_proj.weight": "model-00010-of-00014.safetensors",
|
@@ -544,11 +544,11 @@
|
|
544 |
"model.layers.52.self_attn.q_norm.weight": "model-00011-of-00014.safetensors",
|
545 |
"model.layers.52.self_attn.q_proj.weight": "model-00011-of-00014.safetensors",
|
546 |
"model.layers.52.self_attn.v_proj.weight": "model-00011-of-00014.safetensors",
|
547 |
-
"model.layers.53.input_layernorm.weight": "model-
|
548 |
"model.layers.53.mlp.down_proj.weight": "model-00012-of-00014.safetensors",
|
549 |
"model.layers.53.mlp.gate_proj.weight": "model-00011-of-00014.safetensors",
|
550 |
"model.layers.53.mlp.up_proj.weight": "model-00012-of-00014.safetensors",
|
551 |
-
"model.layers.53.post_attention_layernorm.weight": "model-
|
552 |
"model.layers.53.self_attn.k_norm.weight": "model-00011-of-00014.safetensors",
|
553 |
"model.layers.53.self_attn.k_proj.weight": "model-00011-of-00014.safetensors",
|
554 |
"model.layers.53.self_attn.o_proj.weight": "model-00011-of-00014.safetensors",
|
@@ -599,11 +599,11 @@
|
|
599 |
"model.layers.57.self_attn.q_norm.weight": "model-00012-of-00014.safetensors",
|
600 |
"model.layers.57.self_attn.q_proj.weight": "model-00012-of-00014.safetensors",
|
601 |
"model.layers.57.self_attn.v_proj.weight": "model-00012-of-00014.safetensors",
|
602 |
-
"model.layers.58.input_layernorm.weight": "model-
|
603 |
"model.layers.58.mlp.down_proj.weight": "model-00013-of-00014.safetensors",
|
604 |
"model.layers.58.mlp.gate_proj.weight": "model-00012-of-00014.safetensors",
|
605 |
"model.layers.58.mlp.up_proj.weight": "model-00013-of-00014.safetensors",
|
606 |
-
"model.layers.58.post_attention_layernorm.weight": "model-
|
607 |
"model.layers.58.self_attn.k_norm.weight": "model-00012-of-00014.safetensors",
|
608 |
"model.layers.58.self_attn.k_proj.weight": "model-00012-of-00014.safetensors",
|
609 |
"model.layers.58.self_attn.o_proj.weight": "model-00012-of-00014.safetensors",
|
@@ -665,11 +665,11 @@
|
|
665 |
"model.layers.62.self_attn.q_norm.weight": "model-00013-of-00014.safetensors",
|
666 |
"model.layers.62.self_attn.q_proj.weight": "model-00013-of-00014.safetensors",
|
667 |
"model.layers.62.self_attn.v_proj.weight": "model-00013-of-00014.safetensors",
|
668 |
-
"model.layers.63.input_layernorm.weight": "model-
|
669 |
"model.layers.63.mlp.down_proj.weight": "model-00014-of-00014.safetensors",
|
670 |
"model.layers.63.mlp.gate_proj.weight": "model-00013-of-00014.safetensors",
|
671 |
"model.layers.63.mlp.up_proj.weight": "model-00014-of-00014.safetensors",
|
672 |
-
"model.layers.63.post_attention_layernorm.weight": "model-
|
673 |
"model.layers.63.self_attn.k_norm.weight": "model-00013-of-00014.safetensors",
|
674 |
"model.layers.63.self_attn.k_proj.weight": "model-00013-of-00014.safetensors",
|
675 |
"model.layers.63.self_attn.o_proj.weight": "model-00013-of-00014.safetensors",
|
@@ -687,11 +687,11 @@
|
|
687 |
"model.layers.7.self_attn.q_norm.weight": "model-00002-of-00014.safetensors",
|
688 |
"model.layers.7.self_attn.q_proj.weight": "model-00002-of-00014.safetensors",
|
689 |
"model.layers.7.self_attn.v_proj.weight": "model-00002-of-00014.safetensors",
|
690 |
-
"model.layers.8.input_layernorm.weight": "model-
|
691 |
"model.layers.8.mlp.down_proj.weight": "model-00003-of-00014.safetensors",
|
692 |
"model.layers.8.mlp.gate_proj.weight": "model-00002-of-00014.safetensors",
|
693 |
"model.layers.8.mlp.up_proj.weight": "model-00003-of-00014.safetensors",
|
694 |
-
"model.layers.8.post_attention_layernorm.weight": "model-
|
695 |
"model.layers.8.self_attn.k_norm.weight": "model-00002-of-00014.safetensors",
|
696 |
"model.layers.8.self_attn.k_proj.weight": "model-00002-of-00014.safetensors",
|
697 |
"model.layers.8.self_attn.o_proj.weight": "model-00002-of-00014.safetensors",
|
|
|
60 |
"model.layers.12.self_attn.q_norm.weight": "model-00003-of-00014.safetensors",
|
61 |
"model.layers.12.self_attn.q_proj.weight": "model-00003-of-00014.safetensors",
|
62 |
"model.layers.12.self_attn.v_proj.weight": "model-00003-of-00014.safetensors",
|
63 |
+
"model.layers.13.input_layernorm.weight": "model-00004-of-00014.safetensors",
|
64 |
"model.layers.13.mlp.down_proj.weight": "model-00004-of-00014.safetensors",
|
65 |
"model.layers.13.mlp.gate_proj.weight": "model-00003-of-00014.safetensors",
|
66 |
"model.layers.13.mlp.up_proj.weight": "model-00004-of-00014.safetensors",
|
67 |
+
"model.layers.13.post_attention_layernorm.weight": "model-00004-of-00014.safetensors",
|
68 |
"model.layers.13.self_attn.k_norm.weight": "model-00003-of-00014.safetensors",
|
69 |
"model.layers.13.self_attn.k_proj.weight": "model-00003-of-00014.safetensors",
|
70 |
"model.layers.13.self_attn.o_proj.weight": "model-00003-of-00014.safetensors",
|
|
|
115 |
"model.layers.17.self_attn.q_norm.weight": "model-00004-of-00014.safetensors",
|
116 |
"model.layers.17.self_attn.q_proj.weight": "model-00004-of-00014.safetensors",
|
117 |
"model.layers.17.self_attn.v_proj.weight": "model-00004-of-00014.safetensors",
|
118 |
+
"model.layers.18.input_layernorm.weight": "model-00005-of-00014.safetensors",
|
119 |
"model.layers.18.mlp.down_proj.weight": "model-00005-of-00014.safetensors",
|
120 |
"model.layers.18.mlp.gate_proj.weight": "model-00004-of-00014.safetensors",
|
121 |
"model.layers.18.mlp.up_proj.weight": "model-00005-of-00014.safetensors",
|
122 |
+
"model.layers.18.post_attention_layernorm.weight": "model-00005-of-00014.safetensors",
|
123 |
"model.layers.18.self_attn.k_norm.weight": "model-00004-of-00014.safetensors",
|
124 |
"model.layers.18.self_attn.k_proj.weight": "model-00004-of-00014.safetensors",
|
125 |
"model.layers.18.self_attn.o_proj.weight": "model-00004-of-00014.safetensors",
|
|
|
181 |
"model.layers.22.self_attn.q_norm.weight": "model-00005-of-00014.safetensors",
|
182 |
"model.layers.22.self_attn.q_proj.weight": "model-00005-of-00014.safetensors",
|
183 |
"model.layers.22.self_attn.v_proj.weight": "model-00005-of-00014.safetensors",
|
184 |
+
"model.layers.23.input_layernorm.weight": "model-00006-of-00014.safetensors",
|
185 |
"model.layers.23.mlp.down_proj.weight": "model-00006-of-00014.safetensors",
|
186 |
"model.layers.23.mlp.gate_proj.weight": "model-00005-of-00014.safetensors",
|
187 |
"model.layers.23.mlp.up_proj.weight": "model-00006-of-00014.safetensors",
|
188 |
+
"model.layers.23.post_attention_layernorm.weight": "model-00006-of-00014.safetensors",
|
189 |
"model.layers.23.self_attn.k_norm.weight": "model-00005-of-00014.safetensors",
|
190 |
"model.layers.23.self_attn.k_proj.weight": "model-00005-of-00014.safetensors",
|
191 |
"model.layers.23.self_attn.o_proj.weight": "model-00005-of-00014.safetensors",
|
|
|
236 |
"model.layers.27.self_attn.q_norm.weight": "model-00006-of-00014.safetensors",
|
237 |
"model.layers.27.self_attn.q_proj.weight": "model-00006-of-00014.safetensors",
|
238 |
"model.layers.27.self_attn.v_proj.weight": "model-00006-of-00014.safetensors",
|
239 |
+
"model.layers.28.input_layernorm.weight": "model-00007-of-00014.safetensors",
|
240 |
"model.layers.28.mlp.down_proj.weight": "model-00007-of-00014.safetensors",
|
241 |
"model.layers.28.mlp.gate_proj.weight": "model-00006-of-00014.safetensors",
|
242 |
"model.layers.28.mlp.up_proj.weight": "model-00007-of-00014.safetensors",
|
243 |
+
"model.layers.28.post_attention_layernorm.weight": "model-00007-of-00014.safetensors",
|
244 |
"model.layers.28.self_attn.k_norm.weight": "model-00006-of-00014.safetensors",
|
245 |
"model.layers.28.self_attn.k_proj.weight": "model-00006-of-00014.safetensors",
|
246 |
"model.layers.28.self_attn.o_proj.weight": "model-00006-of-00014.safetensors",
|
|
|
258 |
"model.layers.29.self_attn.q_norm.weight": "model-00007-of-00014.safetensors",
|
259 |
"model.layers.29.self_attn.q_proj.weight": "model-00007-of-00014.safetensors",
|
260 |
"model.layers.29.self_attn.v_proj.weight": "model-00007-of-00014.safetensors",
|
261 |
+
"model.layers.3.input_layernorm.weight": "model-00002-of-00014.safetensors",
|
262 |
"model.layers.3.mlp.down_proj.weight": "model-00002-of-00014.safetensors",
|
263 |
"model.layers.3.mlp.gate_proj.weight": "model-00001-of-00014.safetensors",
|
264 |
"model.layers.3.mlp.up_proj.weight": "model-00002-of-00014.safetensors",
|
265 |
+
"model.layers.3.post_attention_layernorm.weight": "model-00002-of-00014.safetensors",
|
266 |
"model.layers.3.self_attn.k_norm.weight": "model-00001-of-00014.safetensors",
|
267 |
"model.layers.3.self_attn.k_proj.weight": "model-00001-of-00014.safetensors",
|
268 |
"model.layers.3.self_attn.o_proj.weight": "model-00001-of-00014.safetensors",
|
|
|
302 |
"model.layers.32.self_attn.q_norm.weight": "model-00007-of-00014.safetensors",
|
303 |
"model.layers.32.self_attn.q_proj.weight": "model-00007-of-00014.safetensors",
|
304 |
"model.layers.32.self_attn.v_proj.weight": "model-00007-of-00014.safetensors",
|
305 |
+
"model.layers.33.input_layernorm.weight": "model-00008-of-00014.safetensors",
|
306 |
"model.layers.33.mlp.down_proj.weight": "model-00008-of-00014.safetensors",
|
307 |
"model.layers.33.mlp.gate_proj.weight": "model-00007-of-00014.safetensors",
|
308 |
"model.layers.33.mlp.up_proj.weight": "model-00008-of-00014.safetensors",
|
309 |
+
"model.layers.33.post_attention_layernorm.weight": "model-00008-of-00014.safetensors",
|
310 |
"model.layers.33.self_attn.k_norm.weight": "model-00007-of-00014.safetensors",
|
311 |
"model.layers.33.self_attn.k_proj.weight": "model-00007-of-00014.safetensors",
|
312 |
"model.layers.33.self_attn.o_proj.weight": "model-00007-of-00014.safetensors",
|
|
|
357 |
"model.layers.37.self_attn.q_norm.weight": "model-00008-of-00014.safetensors",
|
358 |
"model.layers.37.self_attn.q_proj.weight": "model-00008-of-00014.safetensors",
|
359 |
"model.layers.37.self_attn.v_proj.weight": "model-00008-of-00014.safetensors",
|
360 |
+
"model.layers.38.input_layernorm.weight": "model-00009-of-00014.safetensors",
|
361 |
"model.layers.38.mlp.down_proj.weight": "model-00009-of-00014.safetensors",
|
362 |
"model.layers.38.mlp.gate_proj.weight": "model-00008-of-00014.safetensors",
|
363 |
"model.layers.38.mlp.up_proj.weight": "model-00009-of-00014.safetensors",
|
364 |
+
"model.layers.38.post_attention_layernorm.weight": "model-00009-of-00014.safetensors",
|
365 |
"model.layers.38.self_attn.k_norm.weight": "model-00008-of-00014.safetensors",
|
366 |
"model.layers.38.self_attn.k_proj.weight": "model-00008-of-00014.safetensors",
|
367 |
"model.layers.38.self_attn.o_proj.weight": "model-00008-of-00014.safetensors",
|
|
|
423 |
"model.layers.42.self_attn.q_norm.weight": "model-00009-of-00014.safetensors",
|
424 |
"model.layers.42.self_attn.q_proj.weight": "model-00009-of-00014.safetensors",
|
425 |
"model.layers.42.self_attn.v_proj.weight": "model-00009-of-00014.safetensors",
|
426 |
+
"model.layers.43.input_layernorm.weight": "model-00010-of-00014.safetensors",
|
427 |
"model.layers.43.mlp.down_proj.weight": "model-00010-of-00014.safetensors",
|
428 |
"model.layers.43.mlp.gate_proj.weight": "model-00009-of-00014.safetensors",
|
429 |
"model.layers.43.mlp.up_proj.weight": "model-00010-of-00014.safetensors",
|
430 |
+
"model.layers.43.post_attention_layernorm.weight": "model-00010-of-00014.safetensors",
|
431 |
"model.layers.43.self_attn.k_norm.weight": "model-00009-of-00014.safetensors",
|
432 |
"model.layers.43.self_attn.k_proj.weight": "model-00009-of-00014.safetensors",
|
433 |
"model.layers.43.self_attn.o_proj.weight": "model-00009-of-00014.safetensors",
|
|
|
478 |
"model.layers.47.self_attn.q_norm.weight": "model-00010-of-00014.safetensors",
|
479 |
"model.layers.47.self_attn.q_proj.weight": "model-00010-of-00014.safetensors",
|
480 |
"model.layers.47.self_attn.v_proj.weight": "model-00010-of-00014.safetensors",
|
481 |
+
"model.layers.48.input_layernorm.weight": "model-00011-of-00014.safetensors",
|
482 |
"model.layers.48.mlp.down_proj.weight": "model-00011-of-00014.safetensors",
|
483 |
"model.layers.48.mlp.gate_proj.weight": "model-00010-of-00014.safetensors",
|
484 |
"model.layers.48.mlp.up_proj.weight": "model-00011-of-00014.safetensors",
|
485 |
+
"model.layers.48.post_attention_layernorm.weight": "model-00011-of-00014.safetensors",
|
486 |
"model.layers.48.self_attn.k_norm.weight": "model-00010-of-00014.safetensors",
|
487 |
"model.layers.48.self_attn.k_proj.weight": "model-00010-of-00014.safetensors",
|
488 |
"model.layers.48.self_attn.o_proj.weight": "model-00010-of-00014.safetensors",
|
|
|
544 |
"model.layers.52.self_attn.q_norm.weight": "model-00011-of-00014.safetensors",
|
545 |
"model.layers.52.self_attn.q_proj.weight": "model-00011-of-00014.safetensors",
|
546 |
"model.layers.52.self_attn.v_proj.weight": "model-00011-of-00014.safetensors",
|
547 |
+
"model.layers.53.input_layernorm.weight": "model-00012-of-00014.safetensors",
|
548 |
"model.layers.53.mlp.down_proj.weight": "model-00012-of-00014.safetensors",
|
549 |
"model.layers.53.mlp.gate_proj.weight": "model-00011-of-00014.safetensors",
|
550 |
"model.layers.53.mlp.up_proj.weight": "model-00012-of-00014.safetensors",
|
551 |
+
"model.layers.53.post_attention_layernorm.weight": "model-00012-of-00014.safetensors",
|
552 |
"model.layers.53.self_attn.k_norm.weight": "model-00011-of-00014.safetensors",
|
553 |
"model.layers.53.self_attn.k_proj.weight": "model-00011-of-00014.safetensors",
|
554 |
"model.layers.53.self_attn.o_proj.weight": "model-00011-of-00014.safetensors",
|
|
|
599 |
"model.layers.57.self_attn.q_norm.weight": "model-00012-of-00014.safetensors",
|
600 |
"model.layers.57.self_attn.q_proj.weight": "model-00012-of-00014.safetensors",
|
601 |
"model.layers.57.self_attn.v_proj.weight": "model-00012-of-00014.safetensors",
|
602 |
+
"model.layers.58.input_layernorm.weight": "model-00013-of-00014.safetensors",
|
603 |
"model.layers.58.mlp.down_proj.weight": "model-00013-of-00014.safetensors",
|
604 |
"model.layers.58.mlp.gate_proj.weight": "model-00012-of-00014.safetensors",
|
605 |
"model.layers.58.mlp.up_proj.weight": "model-00013-of-00014.safetensors",
|
606 |
+
"model.layers.58.post_attention_layernorm.weight": "model-00013-of-00014.safetensors",
|
607 |
"model.layers.58.self_attn.k_norm.weight": "model-00012-of-00014.safetensors",
|
608 |
"model.layers.58.self_attn.k_proj.weight": "model-00012-of-00014.safetensors",
|
609 |
"model.layers.58.self_attn.o_proj.weight": "model-00012-of-00014.safetensors",
|
|
|
665 |
"model.layers.62.self_attn.q_norm.weight": "model-00013-of-00014.safetensors",
|
666 |
"model.layers.62.self_attn.q_proj.weight": "model-00013-of-00014.safetensors",
|
667 |
"model.layers.62.self_attn.v_proj.weight": "model-00013-of-00014.safetensors",
|
668 |
+
"model.layers.63.input_layernorm.weight": "model-00014-of-00014.safetensors",
|
669 |
"model.layers.63.mlp.down_proj.weight": "model-00014-of-00014.safetensors",
|
670 |
"model.layers.63.mlp.gate_proj.weight": "model-00013-of-00014.safetensors",
|
671 |
"model.layers.63.mlp.up_proj.weight": "model-00014-of-00014.safetensors",
|
672 |
+
"model.layers.63.post_attention_layernorm.weight": "model-00014-of-00014.safetensors",
|
673 |
"model.layers.63.self_attn.k_norm.weight": "model-00013-of-00014.safetensors",
|
674 |
"model.layers.63.self_attn.k_proj.weight": "model-00013-of-00014.safetensors",
|
675 |
"model.layers.63.self_attn.o_proj.weight": "model-00013-of-00014.safetensors",
|
|
|
687 |
"model.layers.7.self_attn.q_norm.weight": "model-00002-of-00014.safetensors",
|
688 |
"model.layers.7.self_attn.q_proj.weight": "model-00002-of-00014.safetensors",
|
689 |
"model.layers.7.self_attn.v_proj.weight": "model-00002-of-00014.safetensors",
|
690 |
+
"model.layers.8.input_layernorm.weight": "model-00003-of-00014.safetensors",
|
691 |
"model.layers.8.mlp.down_proj.weight": "model-00003-of-00014.safetensors",
|
692 |
"model.layers.8.mlp.gate_proj.weight": "model-00002-of-00014.safetensors",
|
693 |
"model.layers.8.mlp.up_proj.weight": "model-00003-of-00014.safetensors",
|
694 |
+
"model.layers.8.post_attention_layernorm.weight": "model-00003-of-00014.safetensors",
|
695 |
"model.layers.8.self_attn.k_norm.weight": "model-00002-of-00014.safetensors",
|
696 |
"model.layers.8.self_attn.k_proj.weight": "model-00002-of-00014.safetensors",
|
697 |
"model.layers.8.self_attn.o_proj.weight": "model-00002-of-00014.safetensors",
|