Update rope (#1)
Browse files- Update rope (95f4c72067bb91338282c18e006dd68449ac1d54)
Co-authored-by: Pavel Iakubovskii <qubvel-hf@users.noreply.huggingface.co>
- config.json +3 -3
config.json
CHANGED
@@ -22,11 +22,11 @@
|
|
22 |
"num_attention_heads": 8,
|
23 |
"num_attention_layers": 4,
|
24 |
"num_key_value_heads": 8,
|
|
|
25 |
"q_aggregation_kernel_size": 4,
|
26 |
"q_aggregation_stride": 4,
|
27 |
"rope_scaling": {
|
28 |
-
"
|
29 |
-
"rope_type": "2d"
|
30 |
},
|
31 |
"rope_theta": 10000.0,
|
32 |
"stage_block_dims": [
|
@@ -54,5 +54,5 @@
|
|
54 |
2
|
55 |
],
|
56 |
"torch_dtype": "float32",
|
57 |
-
"transformers_version": "4.
|
58 |
}
|
|
|
22 |
"num_attention_heads": 8,
|
23 |
"num_attention_layers": 4,
|
24 |
"num_key_value_heads": 8,
|
25 |
+
"partial_rotary_factor": 4.0,
|
26 |
"q_aggregation_kernel_size": 4,
|
27 |
"q_aggregation_stride": 4,
|
28 |
"rope_scaling": {
|
29 |
+
"rope_type": "default"
|
|
|
30 |
},
|
31 |
"rope_theta": 10000.0,
|
32 |
"stage_block_dims": [
|
|
|
54 |
2
|
55 |
],
|
56 |
"torch_dtype": "float32",
|
57 |
+
"transformers_version": "4.54.0.dev0"
|
58 |
}
|