stevenbucaille qubvel-hf HF Staff commited on
Commit
85afdc7
·
verified ·
1 Parent(s): 9ae1310

Update rope (#1)

Browse files

- Update rope (95f4c72067bb91338282c18e006dd68449ac1d54)


Co-authored-by: Pavel Iakubovskii <qubvel-hf@users.noreply.huggingface.co>

Files changed (1) hide show
  1. config.json +3 -3
config.json CHANGED
@@ -22,11 +22,11 @@
22
  "num_attention_heads": 8,
23
  "num_attention_layers": 4,
24
  "num_key_value_heads": 8,
 
25
  "q_aggregation_kernel_size": 4,
26
  "q_aggregation_stride": 4,
27
  "rope_scaling": {
28
- "dim": 64,
29
- "rope_type": "2d"
30
  },
31
  "rope_theta": 10000.0,
32
  "stage_block_dims": [
@@ -54,5 +54,5 @@
54
  2
55
  ],
56
  "torch_dtype": "float32",
57
- "transformers_version": "4.50.0.dev0"
58
  }
 
22
  "num_attention_heads": 8,
23
  "num_attention_layers": 4,
24
  "num_key_value_heads": 8,
25
+ "partial_rotary_factor": 4.0,
26
  "q_aggregation_kernel_size": 4,
27
  "q_aggregation_stride": 4,
28
  "rope_scaling": {
29
+ "rope_type": "default"
 
30
  },
31
  "rope_theta": 10000.0,
32
  "stage_block_dims": [
 
54
  2
55
  ],
56
  "torch_dtype": "float32",
57
+ "transformers_version": "4.54.0.dev0"
58
  }