n1ck-guo commited on
Commit
74c140c
·
verified ·
1 Parent(s): b026a26

Add files using upload-large-folder tool

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ model.safetensors.index.json filter=lfs diff=lfs merge=lfs -text
chat_template.jinja ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% if not thinking is defined %}{% set thinking = false %}{% endif %}{% set ns = namespace(is_first=false, is_tool=false, system_prompt='', is_first_sp=true, is_last_user=false) %}{%- for message in messages %}{%- if message['role'] == 'system' %}{%- if ns.is_first_sp %}{% set ns.system_prompt = ns.system_prompt + message['content'] %}{% set ns.is_first_sp = false %}{%- else %}{% set ns.system_prompt = ns.system_prompt + '
2
+
3
+ ' + message['content'] %}{%- endif %}{%- endif %}{%- endfor %}{{ bos_token }}{{ ns.system_prompt }}{%- for message in messages %}{%- if message['role'] == 'user' %}{%- set ns.is_tool = false -%}{%- set ns.is_first = false -%}{%- set ns.is_last_user = true -%}{{'<|User|>' + message['content']}}{%- endif %}{%- if message['role'] == 'assistant' and message['tool_calls'] is defined and message['tool_calls'] is not none %}{%- if ns.is_last_user %}{{'<|Assistant|></think>'}}{%- endif %}{%- set ns.is_last_user = false -%}{%- set ns.is_first = false %}{%- set ns.is_tool = false -%}{%- for tool in message['tool_calls'] %}{%- if not ns.is_first %}{%- if message['content'] is none %}{{'<|tool▁calls▁begin|><|tool▁call▁begin|>'+ tool['function']['name'] + '<|tool▁sep|>' + tool['function']['arguments'] + '<|tool▁call▁end|>'}}{%- else %}{{message['content'] + '<|tool▁calls▁begin|><|tool▁call▁begin|>' + tool['function']['name'] + '<|tool▁sep|>' + tool['function']['arguments'] + '<|tool▁call▁end|>'}}{%- endif %}{%- set ns.is_first = true -%}{%- else %}{{'<|tool▁call▁begin|>'+ tool['function']['name'] + '<|tool▁sep|>' + tool['function']['arguments'] + '<|tool▁call▁end|>'}}{%- endif %}{%- endfor %}{{'<|tool▁calls▁end|><|end▁of▁sentence|>'}}{%- endif %}{%- if message['role'] == 'assistant' and (message['tool_calls'] is not defined or message['tool_calls'] is none) %}{%- if ns.is_last_user %}{{'<|Assistant|>'}}{%- if message['prefix'] is defined and message['prefix'] and thinking %}{{'<think>'}} {%- else %}{{'</think>'}}{%- endif %}{%- endif %}{%- set ns.is_last_user = false -%}{%- if ns.is_tool %}{{message['content'] + '<|end▁of▁sentence|>'}}{%- set ns.is_tool = false -%}{%- else %}{%- set content = message['content'] -%}{%- if '</think>' in content %}{%- set content = content.split('</think>', 1)[1] -%}{%- endif %}{{content + '<|end▁of▁sentence|>'}}{%- endif %}{%- endif %}{%- if message['role'] == 'tool' %}{%- set ns.is_last_user = false -%}{%- set ns.is_tool = true -%}{{'<|tool▁output▁begin|>' + message['content'] + '<|tool▁output▁end|>'}}{%- endif %}{%- endfor -%}{%- if add_generation_prompt and ns.is_last_user and not ns.is_tool %}{{'<|Assistant|>'}}{%- if not thinking %}{{'</think>'}}{%- else %}{{'<think>'}}{%- endif %}{% endif %}
config.json ADDED
@@ -0,0 +1,1539 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "DeepseekV3ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "auto_map": {
8
+ "AutoConfig": "configuration_deepseek.DeepseekV3Config",
9
+ "AutoModel": "modeling_deepseek.DeepseekV3Model",
10
+ "AutoModelForCausalLM": "modeling_deepseek.DeepseekV3ForCausalLM"
11
+ },
12
+ "bos_token_id": 0,
13
+ "eos_token_id": 1,
14
+ "ep_size": 1,
15
+ "first_k_dense_replace": 3,
16
+ "head_dim": 64,
17
+ "hidden_act": "silu",
18
+ "hidden_size": 7168,
19
+ "initializer_range": 0.02,
20
+ "intermediate_size": 18432,
21
+ "kv_lora_rank": 512,
22
+ "max_position_embeddings": 163840,
23
+ "model_type": "deepseek_v3",
24
+ "moe_intermediate_size": 2048,
25
+ "moe_layer_freq": 1,
26
+ "n_group": 8,
27
+ "n_routed_experts": 256,
28
+ "n_shared_experts": 1,
29
+ "norm_topk_prob": true,
30
+ "num_attention_heads": 128,
31
+ "num_experts_per_tok": 8,
32
+ "num_hidden_layers": 61,
33
+ "num_key_value_heads": 128,
34
+ "num_nextn_predict_layers": 1,
35
+ "pretraining_tp": 1,
36
+ "q_lora_rank": 1536,
37
+ "qk_head_dim": 192,
38
+ "qk_nope_head_dim": 128,
39
+ "qk_rope_head_dim": 64,
40
+ "quantization_config": {
41
+ "autoround_version": "0.6.1.dev",
42
+ "bits": 4,
43
+ "data_type": "int",
44
+ "extra_config": {
45
+ "model.layers.0.mlp.down_proj": {
46
+ "bits": 8
47
+ },
48
+ "model.layers.0.mlp.gate_proj": {
49
+ "bits": 8
50
+ },
51
+ "model.layers.0.mlp.up_proj": {
52
+ "bits": 8
53
+ },
54
+ "model.layers.0.self_attn.kv_a_proj_with_mqa": {
55
+ "bits": 8
56
+ },
57
+ "model.layers.0.self_attn.kv_b_proj": {
58
+ "bits": 8
59
+ },
60
+ "model.layers.0.self_attn.o_proj": {
61
+ "bits": 8
62
+ },
63
+ "model.layers.0.self_attn.q_a_proj": {
64
+ "bits": 8
65
+ },
66
+ "model.layers.0.self_attn.q_b_proj": {
67
+ "bits": 8
68
+ },
69
+ "model.layers.1.mlp.down_proj": {
70
+ "bits": 8
71
+ },
72
+ "model.layers.1.mlp.gate_proj": {
73
+ "bits": 8
74
+ },
75
+ "model.layers.1.mlp.up_proj": {
76
+ "bits": 8
77
+ },
78
+ "model.layers.1.self_attn.kv_a_proj_with_mqa": {
79
+ "bits": 8
80
+ },
81
+ "model.layers.1.self_attn.kv_b_proj": {
82
+ "bits": 8
83
+ },
84
+ "model.layers.1.self_attn.o_proj": {
85
+ "bits": 8
86
+ },
87
+ "model.layers.1.self_attn.q_a_proj": {
88
+ "bits": 8
89
+ },
90
+ "model.layers.1.self_attn.q_b_proj": {
91
+ "bits": 8
92
+ },
93
+ "model.layers.10.mlp.shared_experts.down_proj": {
94
+ "bits": 8
95
+ },
96
+ "model.layers.10.mlp.shared_experts.gate_proj": {
97
+ "bits": 8
98
+ },
99
+ "model.layers.10.mlp.shared_experts.up_proj": {
100
+ "bits": 8
101
+ },
102
+ "model.layers.10.self_attn.kv_a_proj_with_mqa": {
103
+ "bits": 8
104
+ },
105
+ "model.layers.10.self_attn.kv_b_proj": {
106
+ "bits": 8
107
+ },
108
+ "model.layers.10.self_attn.o_proj": {
109
+ "bits": 8
110
+ },
111
+ "model.layers.10.self_attn.q_a_proj": {
112
+ "bits": 8
113
+ },
114
+ "model.layers.10.self_attn.q_b_proj": {
115
+ "bits": 8
116
+ },
117
+ "model.layers.11.mlp.shared_experts.down_proj": {
118
+ "bits": 8
119
+ },
120
+ "model.layers.11.mlp.shared_experts.gate_proj": {
121
+ "bits": 8
122
+ },
123
+ "model.layers.11.mlp.shared_experts.up_proj": {
124
+ "bits": 8
125
+ },
126
+ "model.layers.11.self_attn.kv_a_proj_with_mqa": {
127
+ "bits": 8
128
+ },
129
+ "model.layers.11.self_attn.kv_b_proj": {
130
+ "bits": 8
131
+ },
132
+ "model.layers.11.self_attn.o_proj": {
133
+ "bits": 8
134
+ },
135
+ "model.layers.11.self_attn.q_a_proj": {
136
+ "bits": 8
137
+ },
138
+ "model.layers.11.self_attn.q_b_proj": {
139
+ "bits": 8
140
+ },
141
+ "model.layers.12.mlp.shared_experts.down_proj": {
142
+ "bits": 8
143
+ },
144
+ "model.layers.12.mlp.shared_experts.gate_proj": {
145
+ "bits": 8
146
+ },
147
+ "model.layers.12.mlp.shared_experts.up_proj": {
148
+ "bits": 8
149
+ },
150
+ "model.layers.12.self_attn.kv_a_proj_with_mqa": {
151
+ "bits": 8
152
+ },
153
+ "model.layers.12.self_attn.kv_b_proj": {
154
+ "bits": 8
155
+ },
156
+ "model.layers.12.self_attn.o_proj": {
157
+ "bits": 8
158
+ },
159
+ "model.layers.12.self_attn.q_a_proj": {
160
+ "bits": 8
161
+ },
162
+ "model.layers.12.self_attn.q_b_proj": {
163
+ "bits": 8
164
+ },
165
+ "model.layers.13.mlp.shared_experts.down_proj": {
166
+ "bits": 8
167
+ },
168
+ "model.layers.13.mlp.shared_experts.gate_proj": {
169
+ "bits": 8
170
+ },
171
+ "model.layers.13.mlp.shared_experts.up_proj": {
172
+ "bits": 8
173
+ },
174
+ "model.layers.13.self_attn.kv_a_proj_with_mqa": {
175
+ "bits": 8
176
+ },
177
+ "model.layers.13.self_attn.kv_b_proj": {
178
+ "bits": 8
179
+ },
180
+ "model.layers.13.self_attn.o_proj": {
181
+ "bits": 8
182
+ },
183
+ "model.layers.13.self_attn.q_a_proj": {
184
+ "bits": 8
185
+ },
186
+ "model.layers.13.self_attn.q_b_proj": {
187
+ "bits": 8
188
+ },
189
+ "model.layers.14.mlp.shared_experts.down_proj": {
190
+ "bits": 8
191
+ },
192
+ "model.layers.14.mlp.shared_experts.gate_proj": {
193
+ "bits": 8
194
+ },
195
+ "model.layers.14.mlp.shared_experts.up_proj": {
196
+ "bits": 8
197
+ },
198
+ "model.layers.14.self_attn.kv_a_proj_with_mqa": {
199
+ "bits": 8
200
+ },
201
+ "model.layers.14.self_attn.kv_b_proj": {
202
+ "bits": 8
203
+ },
204
+ "model.layers.14.self_attn.o_proj": {
205
+ "bits": 8
206
+ },
207
+ "model.layers.14.self_attn.q_a_proj": {
208
+ "bits": 8
209
+ },
210
+ "model.layers.14.self_attn.q_b_proj": {
211
+ "bits": 8
212
+ },
213
+ "model.layers.15.mlp.shared_experts.down_proj": {
214
+ "bits": 8
215
+ },
216
+ "model.layers.15.mlp.shared_experts.gate_proj": {
217
+ "bits": 8
218
+ },
219
+ "model.layers.15.mlp.shared_experts.up_proj": {
220
+ "bits": 8
221
+ },
222
+ "model.layers.15.self_attn.kv_a_proj_with_mqa": {
223
+ "bits": 8
224
+ },
225
+ "model.layers.15.self_attn.kv_b_proj": {
226
+ "bits": 8
227
+ },
228
+ "model.layers.15.self_attn.o_proj": {
229
+ "bits": 8
230
+ },
231
+ "model.layers.15.self_attn.q_a_proj": {
232
+ "bits": 8
233
+ },
234
+ "model.layers.15.self_attn.q_b_proj": {
235
+ "bits": 8
236
+ },
237
+ "model.layers.16.mlp.shared_experts.down_proj": {
238
+ "bits": 8
239
+ },
240
+ "model.layers.16.mlp.shared_experts.gate_proj": {
241
+ "bits": 8
242
+ },
243
+ "model.layers.16.mlp.shared_experts.up_proj": {
244
+ "bits": 8
245
+ },
246
+ "model.layers.16.self_attn.kv_a_proj_with_mqa": {
247
+ "bits": 8
248
+ },
249
+ "model.layers.16.self_attn.kv_b_proj": {
250
+ "bits": 8
251
+ },
252
+ "model.layers.16.self_attn.o_proj": {
253
+ "bits": 8
254
+ },
255
+ "model.layers.16.self_attn.q_a_proj": {
256
+ "bits": 8
257
+ },
258
+ "model.layers.16.self_attn.q_b_proj": {
259
+ "bits": 8
260
+ },
261
+ "model.layers.17.mlp.shared_experts.down_proj": {
262
+ "bits": 8
263
+ },
264
+ "model.layers.17.mlp.shared_experts.gate_proj": {
265
+ "bits": 8
266
+ },
267
+ "model.layers.17.mlp.shared_experts.up_proj": {
268
+ "bits": 8
269
+ },
270
+ "model.layers.17.self_attn.kv_a_proj_with_mqa": {
271
+ "bits": 8
272
+ },
273
+ "model.layers.17.self_attn.kv_b_proj": {
274
+ "bits": 8
275
+ },
276
+ "model.layers.17.self_attn.o_proj": {
277
+ "bits": 8
278
+ },
279
+ "model.layers.17.self_attn.q_a_proj": {
280
+ "bits": 8
281
+ },
282
+ "model.layers.17.self_attn.q_b_proj": {
283
+ "bits": 8
284
+ },
285
+ "model.layers.18.mlp.shared_experts.down_proj": {
286
+ "bits": 8
287
+ },
288
+ "model.layers.18.mlp.shared_experts.gate_proj": {
289
+ "bits": 8
290
+ },
291
+ "model.layers.18.mlp.shared_experts.up_proj": {
292
+ "bits": 8
293
+ },
294
+ "model.layers.18.self_attn.kv_a_proj_with_mqa": {
295
+ "bits": 8
296
+ },
297
+ "model.layers.18.self_attn.kv_b_proj": {
298
+ "bits": 8
299
+ },
300
+ "model.layers.18.self_attn.o_proj": {
301
+ "bits": 8
302
+ },
303
+ "model.layers.18.self_attn.q_a_proj": {
304
+ "bits": 8
305
+ },
306
+ "model.layers.18.self_attn.q_b_proj": {
307
+ "bits": 8
308
+ },
309
+ "model.layers.19.mlp.shared_experts.down_proj": {
310
+ "bits": 8
311
+ },
312
+ "model.layers.19.mlp.shared_experts.gate_proj": {
313
+ "bits": 8
314
+ },
315
+ "model.layers.19.mlp.shared_experts.up_proj": {
316
+ "bits": 8
317
+ },
318
+ "model.layers.19.self_attn.kv_a_proj_with_mqa": {
319
+ "bits": 8
320
+ },
321
+ "model.layers.19.self_attn.kv_b_proj": {
322
+ "bits": 8
323
+ },
324
+ "model.layers.19.self_attn.o_proj": {
325
+ "bits": 8
326
+ },
327
+ "model.layers.19.self_attn.q_a_proj": {
328
+ "bits": 8
329
+ },
330
+ "model.layers.19.self_attn.q_b_proj": {
331
+ "bits": 8
332
+ },
333
+ "model.layers.2.mlp.down_proj": {
334
+ "bits": 8
335
+ },
336
+ "model.layers.2.mlp.gate_proj": {
337
+ "bits": 8
338
+ },
339
+ "model.layers.2.mlp.up_proj": {
340
+ "bits": 8
341
+ },
342
+ "model.layers.2.self_attn.kv_a_proj_with_mqa": {
343
+ "bits": 8
344
+ },
345
+ "model.layers.2.self_attn.kv_b_proj": {
346
+ "bits": 8
347
+ },
348
+ "model.layers.2.self_attn.o_proj": {
349
+ "bits": 8
350
+ },
351
+ "model.layers.2.self_attn.q_a_proj": {
352
+ "bits": 8
353
+ },
354
+ "model.layers.2.self_attn.q_b_proj": {
355
+ "bits": 8
356
+ },
357
+ "model.layers.20.mlp.shared_experts.down_proj": {
358
+ "bits": 8
359
+ },
360
+ "model.layers.20.mlp.shared_experts.gate_proj": {
361
+ "bits": 8
362
+ },
363
+ "model.layers.20.mlp.shared_experts.up_proj": {
364
+ "bits": 8
365
+ },
366
+ "model.layers.20.self_attn.kv_a_proj_with_mqa": {
367
+ "bits": 8
368
+ },
369
+ "model.layers.20.self_attn.kv_b_proj": {
370
+ "bits": 8
371
+ },
372
+ "model.layers.20.self_attn.o_proj": {
373
+ "bits": 8
374
+ },
375
+ "model.layers.20.self_attn.q_a_proj": {
376
+ "bits": 8
377
+ },
378
+ "model.layers.20.self_attn.q_b_proj": {
379
+ "bits": 8
380
+ },
381
+ "model.layers.21.mlp.shared_experts.down_proj": {
382
+ "bits": 8
383
+ },
384
+ "model.layers.21.mlp.shared_experts.gate_proj": {
385
+ "bits": 8
386
+ },
387
+ "model.layers.21.mlp.shared_experts.up_proj": {
388
+ "bits": 8
389
+ },
390
+ "model.layers.21.self_attn.kv_a_proj_with_mqa": {
391
+ "bits": 8
392
+ },
393
+ "model.layers.21.self_attn.kv_b_proj": {
394
+ "bits": 8
395
+ },
396
+ "model.layers.21.self_attn.o_proj": {
397
+ "bits": 8
398
+ },
399
+ "model.layers.21.self_attn.q_a_proj": {
400
+ "bits": 8
401
+ },
402
+ "model.layers.21.self_attn.q_b_proj": {
403
+ "bits": 8
404
+ },
405
+ "model.layers.22.mlp.shared_experts.down_proj": {
406
+ "bits": 8
407
+ },
408
+ "model.layers.22.mlp.shared_experts.gate_proj": {
409
+ "bits": 8
410
+ },
411
+ "model.layers.22.mlp.shared_experts.up_proj": {
412
+ "bits": 8
413
+ },
414
+ "model.layers.22.self_attn.kv_a_proj_with_mqa": {
415
+ "bits": 8
416
+ },
417
+ "model.layers.22.self_attn.kv_b_proj": {
418
+ "bits": 8
419
+ },
420
+ "model.layers.22.self_attn.o_proj": {
421
+ "bits": 8
422
+ },
423
+ "model.layers.22.self_attn.q_a_proj": {
424
+ "bits": 8
425
+ },
426
+ "model.layers.22.self_attn.q_b_proj": {
427
+ "bits": 8
428
+ },
429
+ "model.layers.23.mlp.shared_experts.down_proj": {
430
+ "bits": 8
431
+ },
432
+ "model.layers.23.mlp.shared_experts.gate_proj": {
433
+ "bits": 8
434
+ },
435
+ "model.layers.23.mlp.shared_experts.up_proj": {
436
+ "bits": 8
437
+ },
438
+ "model.layers.23.self_attn.kv_a_proj_with_mqa": {
439
+ "bits": 8
440
+ },
441
+ "model.layers.23.self_attn.kv_b_proj": {
442
+ "bits": 8
443
+ },
444
+ "model.layers.23.self_attn.o_proj": {
445
+ "bits": 8
446
+ },
447
+ "model.layers.23.self_attn.q_a_proj": {
448
+ "bits": 8
449
+ },
450
+ "model.layers.23.self_attn.q_b_proj": {
451
+ "bits": 8
452
+ },
453
+ "model.layers.24.mlp.shared_experts.down_proj": {
454
+ "bits": 8
455
+ },
456
+ "model.layers.24.mlp.shared_experts.gate_proj": {
457
+ "bits": 8
458
+ },
459
+ "model.layers.24.mlp.shared_experts.up_proj": {
460
+ "bits": 8
461
+ },
462
+ "model.layers.24.self_attn.kv_a_proj_with_mqa": {
463
+ "bits": 8
464
+ },
465
+ "model.layers.24.self_attn.kv_b_proj": {
466
+ "bits": 8
467
+ },
468
+ "model.layers.24.self_attn.o_proj": {
469
+ "bits": 8
470
+ },
471
+ "model.layers.24.self_attn.q_a_proj": {
472
+ "bits": 8
473
+ },
474
+ "model.layers.24.self_attn.q_b_proj": {
475
+ "bits": 8
476
+ },
477
+ "model.layers.25.mlp.shared_experts.down_proj": {
478
+ "bits": 8
479
+ },
480
+ "model.layers.25.mlp.shared_experts.gate_proj": {
481
+ "bits": 8
482
+ },
483
+ "model.layers.25.mlp.shared_experts.up_proj": {
484
+ "bits": 8
485
+ },
486
+ "model.layers.25.self_attn.kv_a_proj_with_mqa": {
487
+ "bits": 8
488
+ },
489
+ "model.layers.25.self_attn.kv_b_proj": {
490
+ "bits": 8
491
+ },
492
+ "model.layers.25.self_attn.o_proj": {
493
+ "bits": 8
494
+ },
495
+ "model.layers.25.self_attn.q_a_proj": {
496
+ "bits": 8
497
+ },
498
+ "model.layers.25.self_attn.q_b_proj": {
499
+ "bits": 8
500
+ },
501
+ "model.layers.26.mlp.shared_experts.down_proj": {
502
+ "bits": 8
503
+ },
504
+ "model.layers.26.mlp.shared_experts.gate_proj": {
505
+ "bits": 8
506
+ },
507
+ "model.layers.26.mlp.shared_experts.up_proj": {
508
+ "bits": 8
509
+ },
510
+ "model.layers.26.self_attn.kv_a_proj_with_mqa": {
511
+ "bits": 8
512
+ },
513
+ "model.layers.26.self_attn.kv_b_proj": {
514
+ "bits": 8
515
+ },
516
+ "model.layers.26.self_attn.o_proj": {
517
+ "bits": 8
518
+ },
519
+ "model.layers.26.self_attn.q_a_proj": {
520
+ "bits": 8
521
+ },
522
+ "model.layers.26.self_attn.q_b_proj": {
523
+ "bits": 8
524
+ },
525
+ "model.layers.27.mlp.shared_experts.down_proj": {
526
+ "bits": 8
527
+ },
528
+ "model.layers.27.mlp.shared_experts.gate_proj": {
529
+ "bits": 8
530
+ },
531
+ "model.layers.27.mlp.shared_experts.up_proj": {
532
+ "bits": 8
533
+ },
534
+ "model.layers.27.self_attn.kv_a_proj_with_mqa": {
535
+ "bits": 8
536
+ },
537
+ "model.layers.27.self_attn.kv_b_proj": {
538
+ "bits": 8
539
+ },
540
+ "model.layers.27.self_attn.o_proj": {
541
+ "bits": 8
542
+ },
543
+ "model.layers.27.self_attn.q_a_proj": {
544
+ "bits": 8
545
+ },
546
+ "model.layers.27.self_attn.q_b_proj": {
547
+ "bits": 8
548
+ },
549
+ "model.layers.28.mlp.shared_experts.down_proj": {
550
+ "bits": 8
551
+ },
552
+ "model.layers.28.mlp.shared_experts.gate_proj": {
553
+ "bits": 8
554
+ },
555
+ "model.layers.28.mlp.shared_experts.up_proj": {
556
+ "bits": 8
557
+ },
558
+ "model.layers.28.self_attn.kv_a_proj_with_mqa": {
559
+ "bits": 8
560
+ },
561
+ "model.layers.28.self_attn.kv_b_proj": {
562
+ "bits": 8
563
+ },
564
+ "model.layers.28.self_attn.o_proj": {
565
+ "bits": 8
566
+ },
567
+ "model.layers.28.self_attn.q_a_proj": {
568
+ "bits": 8
569
+ },
570
+ "model.layers.28.self_attn.q_b_proj": {
571
+ "bits": 8
572
+ },
573
+ "model.layers.29.mlp.shared_experts.down_proj": {
574
+ "bits": 8
575
+ },
576
+ "model.layers.29.mlp.shared_experts.gate_proj": {
577
+ "bits": 8
578
+ },
579
+ "model.layers.29.mlp.shared_experts.up_proj": {
580
+ "bits": 8
581
+ },
582
+ "model.layers.29.self_attn.kv_a_proj_with_mqa": {
583
+ "bits": 8
584
+ },
585
+ "model.layers.29.self_attn.kv_b_proj": {
586
+ "bits": 8
587
+ },
588
+ "model.layers.29.self_attn.o_proj": {
589
+ "bits": 8
590
+ },
591
+ "model.layers.29.self_attn.q_a_proj": {
592
+ "bits": 8
593
+ },
594
+ "model.layers.29.self_attn.q_b_proj": {
595
+ "bits": 8
596
+ },
597
+ "model.layers.3.mlp.shared_experts.down_proj": {
598
+ "bits": 8
599
+ },
600
+ "model.layers.3.mlp.shared_experts.gate_proj": {
601
+ "bits": 8
602
+ },
603
+ "model.layers.3.mlp.shared_experts.up_proj": {
604
+ "bits": 8
605
+ },
606
+ "model.layers.3.self_attn.kv_a_proj_with_mqa": {
607
+ "bits": 8
608
+ },
609
+ "model.layers.3.self_attn.kv_b_proj": {
610
+ "bits": 8
611
+ },
612
+ "model.layers.3.self_attn.o_proj": {
613
+ "bits": 8
614
+ },
615
+ "model.layers.3.self_attn.q_a_proj": {
616
+ "bits": 8
617
+ },
618
+ "model.layers.3.self_attn.q_b_proj": {
619
+ "bits": 8
620
+ },
621
+ "model.layers.30.mlp.shared_experts.down_proj": {
622
+ "bits": 8
623
+ },
624
+ "model.layers.30.mlp.shared_experts.gate_proj": {
625
+ "bits": 8
626
+ },
627
+ "model.layers.30.mlp.shared_experts.up_proj": {
628
+ "bits": 8
629
+ },
630
+ "model.layers.30.self_attn.kv_a_proj_with_mqa": {
631
+ "bits": 8
632
+ },
633
+ "model.layers.30.self_attn.kv_b_proj": {
634
+ "bits": 8
635
+ },
636
+ "model.layers.30.self_attn.o_proj": {
637
+ "bits": 8
638
+ },
639
+ "model.layers.30.self_attn.q_a_proj": {
640
+ "bits": 8
641
+ },
642
+ "model.layers.30.self_attn.q_b_proj": {
643
+ "bits": 8
644
+ },
645
+ "model.layers.31.mlp.shared_experts.down_proj": {
646
+ "bits": 8
647
+ },
648
+ "model.layers.31.mlp.shared_experts.gate_proj": {
649
+ "bits": 8
650
+ },
651
+ "model.layers.31.mlp.shared_experts.up_proj": {
652
+ "bits": 8
653
+ },
654
+ "model.layers.31.self_attn.kv_a_proj_with_mqa": {
655
+ "bits": 8
656
+ },
657
+ "model.layers.31.self_attn.kv_b_proj": {
658
+ "bits": 8
659
+ },
660
+ "model.layers.31.self_attn.o_proj": {
661
+ "bits": 8
662
+ },
663
+ "model.layers.31.self_attn.q_a_proj": {
664
+ "bits": 8
665
+ },
666
+ "model.layers.31.self_attn.q_b_proj": {
667
+ "bits": 8
668
+ },
669
+ "model.layers.32.mlp.shared_experts.down_proj": {
670
+ "bits": 8
671
+ },
672
+ "model.layers.32.mlp.shared_experts.gate_proj": {
673
+ "bits": 8
674
+ },
675
+ "model.layers.32.mlp.shared_experts.up_proj": {
676
+ "bits": 8
677
+ },
678
+ "model.layers.32.self_attn.kv_a_proj_with_mqa": {
679
+ "bits": 8
680
+ },
681
+ "model.layers.32.self_attn.kv_b_proj": {
682
+ "bits": 8
683
+ },
684
+ "model.layers.32.self_attn.o_proj": {
685
+ "bits": 8
686
+ },
687
+ "model.layers.32.self_attn.q_a_proj": {
688
+ "bits": 8
689
+ },
690
+ "model.layers.32.self_attn.q_b_proj": {
691
+ "bits": 8
692
+ },
693
+ "model.layers.33.mlp.shared_experts.down_proj": {
694
+ "bits": 8
695
+ },
696
+ "model.layers.33.mlp.shared_experts.gate_proj": {
697
+ "bits": 8
698
+ },
699
+ "model.layers.33.mlp.shared_experts.up_proj": {
700
+ "bits": 8
701
+ },
702
+ "model.layers.33.self_attn.kv_a_proj_with_mqa": {
703
+ "bits": 8
704
+ },
705
+ "model.layers.33.self_attn.kv_b_proj": {
706
+ "bits": 8
707
+ },
708
+ "model.layers.33.self_attn.o_proj": {
709
+ "bits": 8
710
+ },
711
+ "model.layers.33.self_attn.q_a_proj": {
712
+ "bits": 8
713
+ },
714
+ "model.layers.33.self_attn.q_b_proj": {
715
+ "bits": 8
716
+ },
717
+ "model.layers.34.mlp.shared_experts.down_proj": {
718
+ "bits": 8
719
+ },
720
+ "model.layers.34.mlp.shared_experts.gate_proj": {
721
+ "bits": 8
722
+ },
723
+ "model.layers.34.mlp.shared_experts.up_proj": {
724
+ "bits": 8
725
+ },
726
+ "model.layers.34.self_attn.kv_a_proj_with_mqa": {
727
+ "bits": 8
728
+ },
729
+ "model.layers.34.self_attn.kv_b_proj": {
730
+ "bits": 8
731
+ },
732
+ "model.layers.34.self_attn.o_proj": {
733
+ "bits": 8
734
+ },
735
+ "model.layers.34.self_attn.q_a_proj": {
736
+ "bits": 8
737
+ },
738
+ "model.layers.34.self_attn.q_b_proj": {
739
+ "bits": 8
740
+ },
741
+ "model.layers.35.mlp.shared_experts.down_proj": {
742
+ "bits": 8
743
+ },
744
+ "model.layers.35.mlp.shared_experts.gate_proj": {
745
+ "bits": 8
746
+ },
747
+ "model.layers.35.mlp.shared_experts.up_proj": {
748
+ "bits": 8
749
+ },
750
+ "model.layers.35.self_attn.kv_a_proj_with_mqa": {
751
+ "bits": 8
752
+ },
753
+ "model.layers.35.self_attn.kv_b_proj": {
754
+ "bits": 8
755
+ },
756
+ "model.layers.35.self_attn.o_proj": {
757
+ "bits": 8
758
+ },
759
+ "model.layers.35.self_attn.q_a_proj": {
760
+ "bits": 8
761
+ },
762
+ "model.layers.35.self_attn.q_b_proj": {
763
+ "bits": 8
764
+ },
765
+ "model.layers.36.mlp.shared_experts.down_proj": {
766
+ "bits": 8
767
+ },
768
+ "model.layers.36.mlp.shared_experts.gate_proj": {
769
+ "bits": 8
770
+ },
771
+ "model.layers.36.mlp.shared_experts.up_proj": {
772
+ "bits": 8
773
+ },
774
+ "model.layers.36.self_attn.kv_a_proj_with_mqa": {
775
+ "bits": 8
776
+ },
777
+ "model.layers.36.self_attn.kv_b_proj": {
778
+ "bits": 8
779
+ },
780
+ "model.layers.36.self_attn.o_proj": {
781
+ "bits": 8
782
+ },
783
+ "model.layers.36.self_attn.q_a_proj": {
784
+ "bits": 8
785
+ },
786
+ "model.layers.36.self_attn.q_b_proj": {
787
+ "bits": 8
788
+ },
789
+ "model.layers.37.mlp.shared_experts.down_proj": {
790
+ "bits": 8
791
+ },
792
+ "model.layers.37.mlp.shared_experts.gate_proj": {
793
+ "bits": 8
794
+ },
795
+ "model.layers.37.mlp.shared_experts.up_proj": {
796
+ "bits": 8
797
+ },
798
+ "model.layers.37.self_attn.kv_a_proj_with_mqa": {
799
+ "bits": 8
800
+ },
801
+ "model.layers.37.self_attn.kv_b_proj": {
802
+ "bits": 8
803
+ },
804
+ "model.layers.37.self_attn.o_proj": {
805
+ "bits": 8
806
+ },
807
+ "model.layers.37.self_attn.q_a_proj": {
808
+ "bits": 8
809
+ },
810
+ "model.layers.37.self_attn.q_b_proj": {
811
+ "bits": 8
812
+ },
813
+ "model.layers.38.mlp.shared_experts.down_proj": {
814
+ "bits": 8
815
+ },
816
+ "model.layers.38.mlp.shared_experts.gate_proj": {
817
+ "bits": 8
818
+ },
819
+ "model.layers.38.mlp.shared_experts.up_proj": {
820
+ "bits": 8
821
+ },
822
+ "model.layers.38.self_attn.kv_a_proj_with_mqa": {
823
+ "bits": 8
824
+ },
825
+ "model.layers.38.self_attn.kv_b_proj": {
826
+ "bits": 8
827
+ },
828
+ "model.layers.38.self_attn.o_proj": {
829
+ "bits": 8
830
+ },
831
+ "model.layers.38.self_attn.q_a_proj": {
832
+ "bits": 8
833
+ },
834
+ "model.layers.38.self_attn.q_b_proj": {
835
+ "bits": 8
836
+ },
837
+ "model.layers.39.mlp.shared_experts.down_proj": {
838
+ "bits": 8
839
+ },
840
+ "model.layers.39.mlp.shared_experts.gate_proj": {
841
+ "bits": 8
842
+ },
843
+ "model.layers.39.mlp.shared_experts.up_proj": {
844
+ "bits": 8
845
+ },
846
+ "model.layers.39.self_attn.kv_a_proj_with_mqa": {
847
+ "bits": 8
848
+ },
849
+ "model.layers.39.self_attn.kv_b_proj": {
850
+ "bits": 8
851
+ },
852
+ "model.layers.39.self_attn.o_proj": {
853
+ "bits": 8
854
+ },
855
+ "model.layers.39.self_attn.q_a_proj": {
856
+ "bits": 8
857
+ },
858
+ "model.layers.39.self_attn.q_b_proj": {
859
+ "bits": 8
860
+ },
861
+ "model.layers.4.mlp.shared_experts.down_proj": {
862
+ "bits": 8
863
+ },
864
+ "model.layers.4.mlp.shared_experts.gate_proj": {
865
+ "bits": 8
866
+ },
867
+ "model.layers.4.mlp.shared_experts.up_proj": {
868
+ "bits": 8
869
+ },
870
+ "model.layers.4.self_attn.kv_a_proj_with_mqa": {
871
+ "bits": 8
872
+ },
873
+ "model.layers.4.self_attn.kv_b_proj": {
874
+ "bits": 8
875
+ },
876
+ "model.layers.4.self_attn.o_proj": {
877
+ "bits": 8
878
+ },
879
+ "model.layers.4.self_attn.q_a_proj": {
880
+ "bits": 8
881
+ },
882
+ "model.layers.4.self_attn.q_b_proj": {
883
+ "bits": 8
884
+ },
885
+ "model.layers.40.mlp.shared_experts.down_proj": {
886
+ "bits": 8
887
+ },
888
+ "model.layers.40.mlp.shared_experts.gate_proj": {
889
+ "bits": 8
890
+ },
891
+ "model.layers.40.mlp.shared_experts.up_proj": {
892
+ "bits": 8
893
+ },
894
+ "model.layers.40.self_attn.kv_a_proj_with_mqa": {
895
+ "bits": 8
896
+ },
897
+ "model.layers.40.self_attn.kv_b_proj": {
898
+ "bits": 8
899
+ },
900
+ "model.layers.40.self_attn.o_proj": {
901
+ "bits": 8
902
+ },
903
+ "model.layers.40.self_attn.q_a_proj": {
904
+ "bits": 8
905
+ },
906
+ "model.layers.40.self_attn.q_b_proj": {
907
+ "bits": 8
908
+ },
909
+ "model.layers.41.mlp.shared_experts.down_proj": {
910
+ "bits": 8
911
+ },
912
+ "model.layers.41.mlp.shared_experts.gate_proj": {
913
+ "bits": 8
914
+ },
915
+ "model.layers.41.mlp.shared_experts.up_proj": {
916
+ "bits": 8
917
+ },
918
+ "model.layers.41.self_attn.kv_a_proj_with_mqa": {
919
+ "bits": 8
920
+ },
921
+ "model.layers.41.self_attn.kv_b_proj": {
922
+ "bits": 8
923
+ },
924
+ "model.layers.41.self_attn.o_proj": {
925
+ "bits": 8
926
+ },
927
+ "model.layers.41.self_attn.q_a_proj": {
928
+ "bits": 8
929
+ },
930
+ "model.layers.41.self_attn.q_b_proj": {
931
+ "bits": 8
932
+ },
933
+ "model.layers.42.mlp.shared_experts.down_proj": {
934
+ "bits": 8
935
+ },
936
+ "model.layers.42.mlp.shared_experts.gate_proj": {
937
+ "bits": 8
938
+ },
939
+ "model.layers.42.mlp.shared_experts.up_proj": {
940
+ "bits": 8
941
+ },
942
+ "model.layers.42.self_attn.kv_a_proj_with_mqa": {
943
+ "bits": 8
944
+ },
945
+ "model.layers.42.self_attn.kv_b_proj": {
946
+ "bits": 8
947
+ },
948
+ "model.layers.42.self_attn.o_proj": {
949
+ "bits": 8
950
+ },
951
+ "model.layers.42.self_attn.q_a_proj": {
952
+ "bits": 8
953
+ },
954
+ "model.layers.42.self_attn.q_b_proj": {
955
+ "bits": 8
956
+ },
957
+ "model.layers.43.mlp.shared_experts.down_proj": {
958
+ "bits": 8
959
+ },
960
+ "model.layers.43.mlp.shared_experts.gate_proj": {
961
+ "bits": 8
962
+ },
963
+ "model.layers.43.mlp.shared_experts.up_proj": {
964
+ "bits": 8
965
+ },
966
+ "model.layers.43.self_attn.kv_a_proj_with_mqa": {
967
+ "bits": 8
968
+ },
969
+ "model.layers.43.self_attn.kv_b_proj": {
970
+ "bits": 8
971
+ },
972
+ "model.layers.43.self_attn.o_proj": {
973
+ "bits": 8
974
+ },
975
+ "model.layers.43.self_attn.q_a_proj": {
976
+ "bits": 8
977
+ },
978
+ "model.layers.43.self_attn.q_b_proj": {
979
+ "bits": 8
980
+ },
981
+ "model.layers.44.mlp.shared_experts.down_proj": {
982
+ "bits": 8
983
+ },
984
+ "model.layers.44.mlp.shared_experts.gate_proj": {
985
+ "bits": 8
986
+ },
987
+ "model.layers.44.mlp.shared_experts.up_proj": {
988
+ "bits": 8
989
+ },
990
+ "model.layers.44.self_attn.kv_a_proj_with_mqa": {
991
+ "bits": 8
992
+ },
993
+ "model.layers.44.self_attn.kv_b_proj": {
994
+ "bits": 8
995
+ },
996
+ "model.layers.44.self_attn.o_proj": {
997
+ "bits": 8
998
+ },
999
+ "model.layers.44.self_attn.q_a_proj": {
1000
+ "bits": 8
1001
+ },
1002
+ "model.layers.44.self_attn.q_b_proj": {
1003
+ "bits": 8
1004
+ },
1005
+ "model.layers.45.mlp.shared_experts.down_proj": {
1006
+ "bits": 8
1007
+ },
1008
+ "model.layers.45.mlp.shared_experts.gate_proj": {
1009
+ "bits": 8
1010
+ },
1011
+ "model.layers.45.mlp.shared_experts.up_proj": {
1012
+ "bits": 8
1013
+ },
1014
+ "model.layers.45.self_attn.kv_a_proj_with_mqa": {
1015
+ "bits": 8
1016
+ },
1017
+ "model.layers.45.self_attn.kv_b_proj": {
1018
+ "bits": 8
1019
+ },
1020
+ "model.layers.45.self_attn.o_proj": {
1021
+ "bits": 8
1022
+ },
1023
+ "model.layers.45.self_attn.q_a_proj": {
1024
+ "bits": 8
1025
+ },
1026
+ "model.layers.45.self_attn.q_b_proj": {
1027
+ "bits": 8
1028
+ },
1029
+ "model.layers.46.mlp.shared_experts.down_proj": {
1030
+ "bits": 8
1031
+ },
1032
+ "model.layers.46.mlp.shared_experts.gate_proj": {
1033
+ "bits": 8
1034
+ },
1035
+ "model.layers.46.mlp.shared_experts.up_proj": {
1036
+ "bits": 8
1037
+ },
1038
+ "model.layers.46.self_attn.kv_a_proj_with_mqa": {
1039
+ "bits": 8
1040
+ },
1041
+ "model.layers.46.self_attn.kv_b_proj": {
1042
+ "bits": 8
1043
+ },
1044
+ "model.layers.46.self_attn.o_proj": {
1045
+ "bits": 8
1046
+ },
1047
+ "model.layers.46.self_attn.q_a_proj": {
1048
+ "bits": 8
1049
+ },
1050
+ "model.layers.46.self_attn.q_b_proj": {
1051
+ "bits": 8
1052
+ },
1053
+ "model.layers.47.mlp.shared_experts.down_proj": {
1054
+ "bits": 8
1055
+ },
1056
+ "model.layers.47.mlp.shared_experts.gate_proj": {
1057
+ "bits": 8
1058
+ },
1059
+ "model.layers.47.mlp.shared_experts.up_proj": {
1060
+ "bits": 8
1061
+ },
1062
+ "model.layers.47.self_attn.kv_a_proj_with_mqa": {
1063
+ "bits": 8
1064
+ },
1065
+ "model.layers.47.self_attn.kv_b_proj": {
1066
+ "bits": 8
1067
+ },
1068
+ "model.layers.47.self_attn.o_proj": {
1069
+ "bits": 8
1070
+ },
1071
+ "model.layers.47.self_attn.q_a_proj": {
1072
+ "bits": 8
1073
+ },
1074
+ "model.layers.47.self_attn.q_b_proj": {
1075
+ "bits": 8
1076
+ },
1077
+ "model.layers.48.mlp.shared_experts.down_proj": {
1078
+ "bits": 8
1079
+ },
1080
+ "model.layers.48.mlp.shared_experts.gate_proj": {
1081
+ "bits": 8
1082
+ },
1083
+ "model.layers.48.mlp.shared_experts.up_proj": {
1084
+ "bits": 8
1085
+ },
1086
+ "model.layers.48.self_attn.kv_a_proj_with_mqa": {
1087
+ "bits": 8
1088
+ },
1089
+ "model.layers.48.self_attn.kv_b_proj": {
1090
+ "bits": 8
1091
+ },
1092
+ "model.layers.48.self_attn.o_proj": {
1093
+ "bits": 8
1094
+ },
1095
+ "model.layers.48.self_attn.q_a_proj": {
1096
+ "bits": 8
1097
+ },
1098
+ "model.layers.48.self_attn.q_b_proj": {
1099
+ "bits": 8
1100
+ },
1101
+ "model.layers.49.mlp.shared_experts.down_proj": {
1102
+ "bits": 8
1103
+ },
1104
+ "model.layers.49.mlp.shared_experts.gate_proj": {
1105
+ "bits": 8
1106
+ },
1107
+ "model.layers.49.mlp.shared_experts.up_proj": {
1108
+ "bits": 8
1109
+ },
1110
+ "model.layers.49.self_attn.kv_a_proj_with_mqa": {
1111
+ "bits": 8
1112
+ },
1113
+ "model.layers.49.self_attn.kv_b_proj": {
1114
+ "bits": 8
1115
+ },
1116
+ "model.layers.49.self_attn.o_proj": {
1117
+ "bits": 8
1118
+ },
1119
+ "model.layers.49.self_attn.q_a_proj": {
1120
+ "bits": 8
1121
+ },
1122
+ "model.layers.49.self_attn.q_b_proj": {
1123
+ "bits": 8
1124
+ },
1125
+ "model.layers.5.mlp.shared_experts.down_proj": {
1126
+ "bits": 8
1127
+ },
1128
+ "model.layers.5.mlp.shared_experts.gate_proj": {
1129
+ "bits": 8
1130
+ },
1131
+ "model.layers.5.mlp.shared_experts.up_proj": {
1132
+ "bits": 8
1133
+ },
1134
+ "model.layers.5.self_attn.kv_a_proj_with_mqa": {
1135
+ "bits": 8
1136
+ },
1137
+ "model.layers.5.self_attn.kv_b_proj": {
1138
+ "bits": 8
1139
+ },
1140
+ "model.layers.5.self_attn.o_proj": {
1141
+ "bits": 8
1142
+ },
1143
+ "model.layers.5.self_attn.q_a_proj": {
1144
+ "bits": 8
1145
+ },
1146
+ "model.layers.5.self_attn.q_b_proj": {
1147
+ "bits": 8
1148
+ },
1149
+ "model.layers.50.mlp.shared_experts.down_proj": {
1150
+ "bits": 8
1151
+ },
1152
+ "model.layers.50.mlp.shared_experts.gate_proj": {
1153
+ "bits": 8
1154
+ },
1155
+ "model.layers.50.mlp.shared_experts.up_proj": {
1156
+ "bits": 8
1157
+ },
1158
+ "model.layers.50.self_attn.kv_a_proj_with_mqa": {
1159
+ "bits": 8
1160
+ },
1161
+ "model.layers.50.self_attn.kv_b_proj": {
1162
+ "bits": 8
1163
+ },
1164
+ "model.layers.50.self_attn.o_proj": {
1165
+ "bits": 8
1166
+ },
1167
+ "model.layers.50.self_attn.q_a_proj": {
1168
+ "bits": 8
1169
+ },
1170
+ "model.layers.50.self_attn.q_b_proj": {
1171
+ "bits": 8
1172
+ },
1173
+ "model.layers.51.mlp.shared_experts.down_proj": {
1174
+ "bits": 8
1175
+ },
1176
+ "model.layers.51.mlp.shared_experts.gate_proj": {
1177
+ "bits": 8
1178
+ },
1179
+ "model.layers.51.mlp.shared_experts.up_proj": {
1180
+ "bits": 8
1181
+ },
1182
+ "model.layers.51.self_attn.kv_a_proj_with_mqa": {
1183
+ "bits": 8
1184
+ },
1185
+ "model.layers.51.self_attn.kv_b_proj": {
1186
+ "bits": 8
1187
+ },
1188
+ "model.layers.51.self_attn.o_proj": {
1189
+ "bits": 8
1190
+ },
1191
+ "model.layers.51.self_attn.q_a_proj": {
1192
+ "bits": 8
1193
+ },
1194
+ "model.layers.51.self_attn.q_b_proj": {
1195
+ "bits": 8
1196
+ },
1197
+ "model.layers.52.mlp.shared_experts.down_proj": {
1198
+ "bits": 8
1199
+ },
1200
+ "model.layers.52.mlp.shared_experts.gate_proj": {
1201
+ "bits": 8
1202
+ },
1203
+ "model.layers.52.mlp.shared_experts.up_proj": {
1204
+ "bits": 8
1205
+ },
1206
+ "model.layers.52.self_attn.kv_a_proj_with_mqa": {
1207
+ "bits": 8
1208
+ },
1209
+ "model.layers.52.self_attn.kv_b_proj": {
1210
+ "bits": 8
1211
+ },
1212
+ "model.layers.52.self_attn.o_proj": {
1213
+ "bits": 8
1214
+ },
1215
+ "model.layers.52.self_attn.q_a_proj": {
1216
+ "bits": 8
1217
+ },
1218
+ "model.layers.52.self_attn.q_b_proj": {
1219
+ "bits": 8
1220
+ },
1221
+ "model.layers.53.mlp.shared_experts.down_proj": {
1222
+ "bits": 8
1223
+ },
1224
+ "model.layers.53.mlp.shared_experts.gate_proj": {
1225
+ "bits": 8
1226
+ },
1227
+ "model.layers.53.mlp.shared_experts.up_proj": {
1228
+ "bits": 8
1229
+ },
1230
+ "model.layers.53.self_attn.kv_a_proj_with_mqa": {
1231
+ "bits": 8
1232
+ },
1233
+ "model.layers.53.self_attn.kv_b_proj": {
1234
+ "bits": 8
1235
+ },
1236
+ "model.layers.53.self_attn.o_proj": {
1237
+ "bits": 8
1238
+ },
1239
+ "model.layers.53.self_attn.q_a_proj": {
1240
+ "bits": 8
1241
+ },
1242
+ "model.layers.53.self_attn.q_b_proj": {
1243
+ "bits": 8
1244
+ },
1245
+ "model.layers.54.mlp.shared_experts.down_proj": {
1246
+ "bits": 8
1247
+ },
1248
+ "model.layers.54.mlp.shared_experts.gate_proj": {
1249
+ "bits": 8
1250
+ },
1251
+ "model.layers.54.mlp.shared_experts.up_proj": {
1252
+ "bits": 8
1253
+ },
1254
+ "model.layers.54.self_attn.kv_a_proj_with_mqa": {
1255
+ "bits": 8
1256
+ },
1257
+ "model.layers.54.self_attn.kv_b_proj": {
1258
+ "bits": 8
1259
+ },
1260
+ "model.layers.54.self_attn.o_proj": {
1261
+ "bits": 8
1262
+ },
1263
+ "model.layers.54.self_attn.q_a_proj": {
1264
+ "bits": 8
1265
+ },
1266
+ "model.layers.54.self_attn.q_b_proj": {
1267
+ "bits": 8
1268
+ },
1269
+ "model.layers.55.mlp.shared_experts.down_proj": {
1270
+ "bits": 8
1271
+ },
1272
+ "model.layers.55.mlp.shared_experts.gate_proj": {
1273
+ "bits": 8
1274
+ },
1275
+ "model.layers.55.mlp.shared_experts.up_proj": {
1276
+ "bits": 8
1277
+ },
1278
+ "model.layers.55.self_attn.kv_a_proj_with_mqa": {
1279
+ "bits": 8
1280
+ },
1281
+ "model.layers.55.self_attn.kv_b_proj": {
1282
+ "bits": 8
1283
+ },
1284
+ "model.layers.55.self_attn.o_proj": {
1285
+ "bits": 8
1286
+ },
1287
+ "model.layers.55.self_attn.q_a_proj": {
1288
+ "bits": 8
1289
+ },
1290
+ "model.layers.55.self_attn.q_b_proj": {
1291
+ "bits": 8
1292
+ },
1293
+ "model.layers.56.mlp.shared_experts.down_proj": {
1294
+ "bits": 8
1295
+ },
1296
+ "model.layers.56.mlp.shared_experts.gate_proj": {
1297
+ "bits": 8
1298
+ },
1299
+ "model.layers.56.mlp.shared_experts.up_proj": {
1300
+ "bits": 8
1301
+ },
1302
+ "model.layers.56.self_attn.kv_a_proj_with_mqa": {
1303
+ "bits": 8
1304
+ },
1305
+ "model.layers.56.self_attn.kv_b_proj": {
1306
+ "bits": 8
1307
+ },
1308
+ "model.layers.56.self_attn.o_proj": {
1309
+ "bits": 8
1310
+ },
1311
+ "model.layers.56.self_attn.q_a_proj": {
1312
+ "bits": 8
1313
+ },
1314
+ "model.layers.56.self_attn.q_b_proj": {
1315
+ "bits": 8
1316
+ },
1317
+ "model.layers.57.mlp.shared_experts.down_proj": {
1318
+ "bits": 8
1319
+ },
1320
+ "model.layers.57.mlp.shared_experts.gate_proj": {
1321
+ "bits": 8
1322
+ },
1323
+ "model.layers.57.mlp.shared_experts.up_proj": {
1324
+ "bits": 8
1325
+ },
1326
+ "model.layers.57.self_attn.kv_a_proj_with_mqa": {
1327
+ "bits": 8
1328
+ },
1329
+ "model.layers.57.self_attn.kv_b_proj": {
1330
+ "bits": 8
1331
+ },
1332
+ "model.layers.57.self_attn.o_proj": {
1333
+ "bits": 8
1334
+ },
1335
+ "model.layers.57.self_attn.q_a_proj": {
1336
+ "bits": 8
1337
+ },
1338
+ "model.layers.57.self_attn.q_b_proj": {
1339
+ "bits": 8
1340
+ },
1341
+ "model.layers.58.mlp.shared_experts.down_proj": {
1342
+ "bits": 8
1343
+ },
1344
+ "model.layers.58.mlp.shared_experts.gate_proj": {
1345
+ "bits": 8
1346
+ },
1347
+ "model.layers.58.mlp.shared_experts.up_proj": {
1348
+ "bits": 8
1349
+ },
1350
+ "model.layers.58.self_attn.kv_a_proj_with_mqa": {
1351
+ "bits": 8
1352
+ },
1353
+ "model.layers.58.self_attn.kv_b_proj": {
1354
+ "bits": 8
1355
+ },
1356
+ "model.layers.58.self_attn.o_proj": {
1357
+ "bits": 8
1358
+ },
1359
+ "model.layers.58.self_attn.q_a_proj": {
1360
+ "bits": 8
1361
+ },
1362
+ "model.layers.58.self_attn.q_b_proj": {
1363
+ "bits": 8
1364
+ },
1365
+ "model.layers.59.mlp.shared_experts.down_proj": {
1366
+ "bits": 8
1367
+ },
1368
+ "model.layers.59.mlp.shared_experts.gate_proj": {
1369
+ "bits": 8
1370
+ },
1371
+ "model.layers.59.mlp.shared_experts.up_proj": {
1372
+ "bits": 8
1373
+ },
1374
+ "model.layers.59.self_attn.kv_a_proj_with_mqa": {
1375
+ "bits": 8
1376
+ },
1377
+ "model.layers.59.self_attn.kv_b_proj": {
1378
+ "bits": 8
1379
+ },
1380
+ "model.layers.59.self_attn.o_proj": {
1381
+ "bits": 8
1382
+ },
1383
+ "model.layers.59.self_attn.q_a_proj": {
1384
+ "bits": 8
1385
+ },
1386
+ "model.layers.59.self_attn.q_b_proj": {
1387
+ "bits": 8
1388
+ },
1389
+ "model.layers.6.mlp.shared_experts.down_proj": {
1390
+ "bits": 8
1391
+ },
1392
+ "model.layers.6.mlp.shared_experts.gate_proj": {
1393
+ "bits": 8
1394
+ },
1395
+ "model.layers.6.mlp.shared_experts.up_proj": {
1396
+ "bits": 8
1397
+ },
1398
+ "model.layers.6.self_attn.kv_a_proj_with_mqa": {
1399
+ "bits": 8
1400
+ },
1401
+ "model.layers.6.self_attn.kv_b_proj": {
1402
+ "bits": 8
1403
+ },
1404
+ "model.layers.6.self_attn.o_proj": {
1405
+ "bits": 8
1406
+ },
1407
+ "model.layers.6.self_attn.q_a_proj": {
1408
+ "bits": 8
1409
+ },
1410
+ "model.layers.6.self_attn.q_b_proj": {
1411
+ "bits": 8
1412
+ },
1413
+ "model.layers.60.mlp.shared_experts.down_proj": {
1414
+ "bits": 8
1415
+ },
1416
+ "model.layers.60.mlp.shared_experts.gate_proj": {
1417
+ "bits": 8
1418
+ },
1419
+ "model.layers.60.mlp.shared_experts.up_proj": {
1420
+ "bits": 8
1421
+ },
1422
+ "model.layers.60.self_attn.kv_a_proj_with_mqa": {
1423
+ "bits": 8
1424
+ },
1425
+ "model.layers.60.self_attn.kv_b_proj": {
1426
+ "bits": 8
1427
+ },
1428
+ "model.layers.60.self_attn.o_proj": {
1429
+ "bits": 8
1430
+ },
1431
+ "model.layers.60.self_attn.q_a_proj": {
1432
+ "bits": 8
1433
+ },
1434
+ "model.layers.60.self_attn.q_b_proj": {
1435
+ "bits": 8
1436
+ },
1437
+ "model.layers.7.mlp.shared_experts.down_proj": {
1438
+ "bits": 8
1439
+ },
1440
+ "model.layers.7.mlp.shared_experts.gate_proj": {
1441
+ "bits": 8
1442
+ },
1443
+ "model.layers.7.mlp.shared_experts.up_proj": {
1444
+ "bits": 8
1445
+ },
1446
+ "model.layers.7.self_attn.kv_a_proj_with_mqa": {
1447
+ "bits": 8
1448
+ },
1449
+ "model.layers.7.self_attn.kv_b_proj": {
1450
+ "bits": 8
1451
+ },
1452
+ "model.layers.7.self_attn.o_proj": {
1453
+ "bits": 8
1454
+ },
1455
+ "model.layers.7.self_attn.q_a_proj": {
1456
+ "bits": 8
1457
+ },
1458
+ "model.layers.7.self_attn.q_b_proj": {
1459
+ "bits": 8
1460
+ },
1461
+ "model.layers.8.mlp.shared_experts.down_proj": {
1462
+ "bits": 8
1463
+ },
1464
+ "model.layers.8.mlp.shared_experts.gate_proj": {
1465
+ "bits": 8
1466
+ },
1467
+ "model.layers.8.mlp.shared_experts.up_proj": {
1468
+ "bits": 8
1469
+ },
1470
+ "model.layers.8.self_attn.kv_a_proj_with_mqa": {
1471
+ "bits": 8
1472
+ },
1473
+ "model.layers.8.self_attn.kv_b_proj": {
1474
+ "bits": 8
1475
+ },
1476
+ "model.layers.8.self_attn.o_proj": {
1477
+ "bits": 8
1478
+ },
1479
+ "model.layers.8.self_attn.q_a_proj": {
1480
+ "bits": 8
1481
+ },
1482
+ "model.layers.8.self_attn.q_b_proj": {
1483
+ "bits": 8
1484
+ },
1485
+ "model.layers.9.mlp.shared_experts.down_proj": {
1486
+ "bits": 8
1487
+ },
1488
+ "model.layers.9.mlp.shared_experts.gate_proj": {
1489
+ "bits": 8
1490
+ },
1491
+ "model.layers.9.mlp.shared_experts.up_proj": {
1492
+ "bits": 8
1493
+ },
1494
+ "model.layers.9.self_attn.kv_a_proj_with_mqa": {
1495
+ "bits": 8
1496
+ },
1497
+ "model.layers.9.self_attn.kv_b_proj": {
1498
+ "bits": 8
1499
+ },
1500
+ "model.layers.9.self_attn.o_proj": {
1501
+ "bits": 8
1502
+ },
1503
+ "model.layers.9.self_attn.q_a_proj": {
1504
+ "bits": 8
1505
+ },
1506
+ "model.layers.9.self_attn.q_b_proj": {
1507
+ "bits": 8
1508
+ }
1509
+ },
1510
+ "group_size": 128,
1511
+ "iters": 0,
1512
+ "packing_format": "auto_round:auto_gptq",
1513
+ "quant_method": "auto-round",
1514
+ "sym": true
1515
+ },
1516
+ "rms_norm_eps": 1e-06,
1517
+ "rope_interleave": true,
1518
+ "rope_scaling": {
1519
+ "beta_fast": 32.0,
1520
+ "beta_slow": 1.0,
1521
+ "factor": 40.0,
1522
+ "mscale": 1.0,
1523
+ "mscale_all_dim": 1.0,
1524
+ "original_max_position_embeddings": 4096,
1525
+ "rope_type": "yarn",
1526
+ "type": "yarn"
1527
+ },
1528
+ "rope_theta": 10000,
1529
+ "routed_scaling_factor": 2.5,
1530
+ "scoring_func": "sigmoid",
1531
+ "tie_word_embeddings": false,
1532
+ "topk_group": 4,
1533
+ "topk_method": "noaux_tc",
1534
+ "torch_dtype": "bfloat16",
1535
+ "transformers_version": "4.56.0.dev0",
1536
+ "use_cache": true,
1537
+ "v_head_dim": 128,
1538
+ "vocab_size": 129280
1539
+ }
generation_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 0,
4
+ "do_sample": true,
5
+ "eos_token_id": 1,
6
+ "temperature": 0.6,
7
+ "top_p": 0.95,
8
+ "transformers_version": "4.56.0.dev0"
9
+ }
model-00001-of-00072.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:118c1bd57564a8cc357b54a83319a1268f6c0e022f4ce2cc8d3549aae56d9c99
3
+ size 4995710664
model-00007-of-00072.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c6fdec602961dd94e355479c076aac7a72a4a535b730466b41e0cb591a7e347
3
+ size 4999591608
model-00016-of-00072.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41efc38b5b63773014829fc209a5be6cb2aeff3ea1b4d01d23d9ae1db1a2d7e1
3
+ size 4999593080
model-00021-of-00072.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:174ccbd9b11842257b80f4da1fd94a8d9c631f18dd49f787447512b19f74f0dd
3
+ size 4999593080
model-00040-of-00072.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42fb3ff28d4bb225af967e6cb1d005ebaad3148d8e0085f5b0d56bb15db5dd77
3
+ size 4999593512
model-00041-of-00072.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba562b08c01e3a4de0d392148a0712e442f05f15e3789334bf46412eb1f327a1
3
+ size 4993807032
model-00043-of-00072.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8fccd19da91d67507c0f9f2c896759506f4899cf6703b6eb692c36227be35fa
3
+ size 4999593096
model-00046-of-00072.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50ebc67ed08abfc78d5d71c172848690594e9179a4ea08b4b28d24631950d185
3
+ size 4999593512
model-00050-of-00072.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:952a7b122d544acecb6c01df90549f3e1c976ae5844f77ce6624b277381c360f
3
+ size 4999593120
model-00062-of-00072.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23ec251eb5545d4fda94786cbd9c0165b825caa87b8c5c1aa23fd40669a02214
3
+ size 4999593512
model-00071-of-00072.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4daf567baa4ef965b996278fac08cf4228a164bb8805db03f5f19bbee5a4a485
3
+ size 4999593080
model-00072-of-00072.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b3af9f0bffb442001896d73f95c2376af6e7e6784780fb986209654360268be
3
+ size 4480112800
model.safetensors.index.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e106ea7643cc5eea90a8e90c2f3e0e22637c40284469021bacf173beb32f11b
3
+ size 12338633
quantization_config.json ADDED
@@ -0,0 +1,1476 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bits": 4,
3
+ "group_size": 128,
4
+ "sym": true,
5
+ "data_type": "int",
6
+ "iters": 0,
7
+ "autoround_version": "0.6.1.dev",
8
+ "quant_method": "auto-round",
9
+ "packing_format": "auto_round:auto_gptq",
10
+ "extra_config": {
11
+ "model.layers.0.self_attn.q_a_proj": {
12
+ "bits": 8
13
+ },
14
+ "model.layers.0.self_attn.q_b_proj": {
15
+ "bits": 8
16
+ },
17
+ "model.layers.0.self_attn.kv_a_proj_with_mqa": {
18
+ "bits": 8
19
+ },
20
+ "model.layers.0.self_attn.kv_b_proj": {
21
+ "bits": 8
22
+ },
23
+ "model.layers.0.self_attn.o_proj": {
24
+ "bits": 8
25
+ },
26
+ "model.layers.0.mlp.gate_proj": {
27
+ "bits": 8
28
+ },
29
+ "model.layers.0.mlp.up_proj": {
30
+ "bits": 8
31
+ },
32
+ "model.layers.0.mlp.down_proj": {
33
+ "bits": 8
34
+ },
35
+ "model.layers.1.self_attn.q_a_proj": {
36
+ "bits": 8
37
+ },
38
+ "model.layers.1.self_attn.q_b_proj": {
39
+ "bits": 8
40
+ },
41
+ "model.layers.1.self_attn.kv_a_proj_with_mqa": {
42
+ "bits": 8
43
+ },
44
+ "model.layers.1.self_attn.kv_b_proj": {
45
+ "bits": 8
46
+ },
47
+ "model.layers.1.self_attn.o_proj": {
48
+ "bits": 8
49
+ },
50
+ "model.layers.1.mlp.gate_proj": {
51
+ "bits": 8
52
+ },
53
+ "model.layers.1.mlp.up_proj": {
54
+ "bits": 8
55
+ },
56
+ "model.layers.1.mlp.down_proj": {
57
+ "bits": 8
58
+ },
59
+ "model.layers.2.self_attn.q_a_proj": {
60
+ "bits": 8
61
+ },
62
+ "model.layers.2.self_attn.q_b_proj": {
63
+ "bits": 8
64
+ },
65
+ "model.layers.2.self_attn.kv_a_proj_with_mqa": {
66
+ "bits": 8
67
+ },
68
+ "model.layers.2.self_attn.kv_b_proj": {
69
+ "bits": 8
70
+ },
71
+ "model.layers.2.self_attn.o_proj": {
72
+ "bits": 8
73
+ },
74
+ "model.layers.2.mlp.gate_proj": {
75
+ "bits": 8
76
+ },
77
+ "model.layers.2.mlp.up_proj": {
78
+ "bits": 8
79
+ },
80
+ "model.layers.2.mlp.down_proj": {
81
+ "bits": 8
82
+ },
83
+ "model.layers.3.self_attn.q_a_proj": {
84
+ "bits": 8
85
+ },
86
+ "model.layers.3.self_attn.q_b_proj": {
87
+ "bits": 8
88
+ },
89
+ "model.layers.3.self_attn.kv_a_proj_with_mqa": {
90
+ "bits": 8
91
+ },
92
+ "model.layers.3.self_attn.kv_b_proj": {
93
+ "bits": 8
94
+ },
95
+ "model.layers.3.self_attn.o_proj": {
96
+ "bits": 8
97
+ },
98
+ "model.layers.3.mlp.shared_experts.gate_proj": {
99
+ "bits": 8
100
+ },
101
+ "model.layers.3.mlp.shared_experts.up_proj": {
102
+ "bits": 8
103
+ },
104
+ "model.layers.3.mlp.shared_experts.down_proj": {
105
+ "bits": 8
106
+ },
107
+ "model.layers.4.self_attn.q_a_proj": {
108
+ "bits": 8
109
+ },
110
+ "model.layers.4.self_attn.q_b_proj": {
111
+ "bits": 8
112
+ },
113
+ "model.layers.4.self_attn.kv_a_proj_with_mqa": {
114
+ "bits": 8
115
+ },
116
+ "model.layers.4.self_attn.kv_b_proj": {
117
+ "bits": 8
118
+ },
119
+ "model.layers.4.self_attn.o_proj": {
120
+ "bits": 8
121
+ },
122
+ "model.layers.4.mlp.shared_experts.gate_proj": {
123
+ "bits": 8
124
+ },
125
+ "model.layers.4.mlp.shared_experts.up_proj": {
126
+ "bits": 8
127
+ },
128
+ "model.layers.4.mlp.shared_experts.down_proj": {
129
+ "bits": 8
130
+ },
131
+ "model.layers.5.self_attn.q_a_proj": {
132
+ "bits": 8
133
+ },
134
+ "model.layers.5.self_attn.q_b_proj": {
135
+ "bits": 8
136
+ },
137
+ "model.layers.5.self_attn.kv_a_proj_with_mqa": {
138
+ "bits": 8
139
+ },
140
+ "model.layers.5.self_attn.kv_b_proj": {
141
+ "bits": 8
142
+ },
143
+ "model.layers.5.self_attn.o_proj": {
144
+ "bits": 8
145
+ },
146
+ "model.layers.5.mlp.shared_experts.gate_proj": {
147
+ "bits": 8
148
+ },
149
+ "model.layers.5.mlp.shared_experts.up_proj": {
150
+ "bits": 8
151
+ },
152
+ "model.layers.5.mlp.shared_experts.down_proj": {
153
+ "bits": 8
154
+ },
155
+ "model.layers.6.self_attn.q_a_proj": {
156
+ "bits": 8
157
+ },
158
+ "model.layers.6.self_attn.q_b_proj": {
159
+ "bits": 8
160
+ },
161
+ "model.layers.6.self_attn.kv_a_proj_with_mqa": {
162
+ "bits": 8
163
+ },
164
+ "model.layers.6.self_attn.kv_b_proj": {
165
+ "bits": 8
166
+ },
167
+ "model.layers.6.self_attn.o_proj": {
168
+ "bits": 8
169
+ },
170
+ "model.layers.6.mlp.shared_experts.gate_proj": {
171
+ "bits": 8
172
+ },
173
+ "model.layers.6.mlp.shared_experts.up_proj": {
174
+ "bits": 8
175
+ },
176
+ "model.layers.6.mlp.shared_experts.down_proj": {
177
+ "bits": 8
178
+ },
179
+ "model.layers.7.self_attn.q_a_proj": {
180
+ "bits": 8
181
+ },
182
+ "model.layers.7.self_attn.q_b_proj": {
183
+ "bits": 8
184
+ },
185
+ "model.layers.7.self_attn.kv_a_proj_with_mqa": {
186
+ "bits": 8
187
+ },
188
+ "model.layers.7.self_attn.kv_b_proj": {
189
+ "bits": 8
190
+ },
191
+ "model.layers.7.self_attn.o_proj": {
192
+ "bits": 8
193
+ },
194
+ "model.layers.7.mlp.shared_experts.gate_proj": {
195
+ "bits": 8
196
+ },
197
+ "model.layers.7.mlp.shared_experts.up_proj": {
198
+ "bits": 8
199
+ },
200
+ "model.layers.7.mlp.shared_experts.down_proj": {
201
+ "bits": 8
202
+ },
203
+ "model.layers.8.self_attn.q_a_proj": {
204
+ "bits": 8
205
+ },
206
+ "model.layers.8.self_attn.q_b_proj": {
207
+ "bits": 8
208
+ },
209
+ "model.layers.8.self_attn.kv_a_proj_with_mqa": {
210
+ "bits": 8
211
+ },
212
+ "model.layers.8.self_attn.kv_b_proj": {
213
+ "bits": 8
214
+ },
215
+ "model.layers.8.self_attn.o_proj": {
216
+ "bits": 8
217
+ },
218
+ "model.layers.8.mlp.shared_experts.gate_proj": {
219
+ "bits": 8
220
+ },
221
+ "model.layers.8.mlp.shared_experts.up_proj": {
222
+ "bits": 8
223
+ },
224
+ "model.layers.8.mlp.shared_experts.down_proj": {
225
+ "bits": 8
226
+ },
227
+ "model.layers.9.self_attn.q_a_proj": {
228
+ "bits": 8
229
+ },
230
+ "model.layers.9.self_attn.q_b_proj": {
231
+ "bits": 8
232
+ },
233
+ "model.layers.9.self_attn.kv_a_proj_with_mqa": {
234
+ "bits": 8
235
+ },
236
+ "model.layers.9.self_attn.kv_b_proj": {
237
+ "bits": 8
238
+ },
239
+ "model.layers.9.self_attn.o_proj": {
240
+ "bits": 8
241
+ },
242
+ "model.layers.9.mlp.shared_experts.gate_proj": {
243
+ "bits": 8
244
+ },
245
+ "model.layers.9.mlp.shared_experts.up_proj": {
246
+ "bits": 8
247
+ },
248
+ "model.layers.9.mlp.shared_experts.down_proj": {
249
+ "bits": 8
250
+ },
251
+ "model.layers.10.self_attn.q_a_proj": {
252
+ "bits": 8
253
+ },
254
+ "model.layers.10.self_attn.q_b_proj": {
255
+ "bits": 8
256
+ },
257
+ "model.layers.10.self_attn.kv_a_proj_with_mqa": {
258
+ "bits": 8
259
+ },
260
+ "model.layers.10.self_attn.kv_b_proj": {
261
+ "bits": 8
262
+ },
263
+ "model.layers.10.self_attn.o_proj": {
264
+ "bits": 8
265
+ },
266
+ "model.layers.10.mlp.shared_experts.gate_proj": {
267
+ "bits": 8
268
+ },
269
+ "model.layers.10.mlp.shared_experts.up_proj": {
270
+ "bits": 8
271
+ },
272
+ "model.layers.10.mlp.shared_experts.down_proj": {
273
+ "bits": 8
274
+ },
275
+ "model.layers.11.self_attn.q_a_proj": {
276
+ "bits": 8
277
+ },
278
+ "model.layers.11.self_attn.q_b_proj": {
279
+ "bits": 8
280
+ },
281
+ "model.layers.11.self_attn.kv_a_proj_with_mqa": {
282
+ "bits": 8
283
+ },
284
+ "model.layers.11.self_attn.kv_b_proj": {
285
+ "bits": 8
286
+ },
287
+ "model.layers.11.self_attn.o_proj": {
288
+ "bits": 8
289
+ },
290
+ "model.layers.11.mlp.shared_experts.gate_proj": {
291
+ "bits": 8
292
+ },
293
+ "model.layers.11.mlp.shared_experts.up_proj": {
294
+ "bits": 8
295
+ },
296
+ "model.layers.11.mlp.shared_experts.down_proj": {
297
+ "bits": 8
298
+ },
299
+ "model.layers.12.self_attn.q_a_proj": {
300
+ "bits": 8
301
+ },
302
+ "model.layers.12.self_attn.q_b_proj": {
303
+ "bits": 8
304
+ },
305
+ "model.layers.12.self_attn.kv_a_proj_with_mqa": {
306
+ "bits": 8
307
+ },
308
+ "model.layers.12.self_attn.kv_b_proj": {
309
+ "bits": 8
310
+ },
311
+ "model.layers.12.self_attn.o_proj": {
312
+ "bits": 8
313
+ },
314
+ "model.layers.12.mlp.shared_experts.gate_proj": {
315
+ "bits": 8
316
+ },
317
+ "model.layers.12.mlp.shared_experts.up_proj": {
318
+ "bits": 8
319
+ },
320
+ "model.layers.12.mlp.shared_experts.down_proj": {
321
+ "bits": 8
322
+ },
323
+ "model.layers.13.self_attn.q_a_proj": {
324
+ "bits": 8
325
+ },
326
+ "model.layers.13.self_attn.q_b_proj": {
327
+ "bits": 8
328
+ },
329
+ "model.layers.13.self_attn.kv_a_proj_with_mqa": {
330
+ "bits": 8
331
+ },
332
+ "model.layers.13.self_attn.kv_b_proj": {
333
+ "bits": 8
334
+ },
335
+ "model.layers.13.self_attn.o_proj": {
336
+ "bits": 8
337
+ },
338
+ "model.layers.13.mlp.shared_experts.gate_proj": {
339
+ "bits": 8
340
+ },
341
+ "model.layers.13.mlp.shared_experts.up_proj": {
342
+ "bits": 8
343
+ },
344
+ "model.layers.13.mlp.shared_experts.down_proj": {
345
+ "bits": 8
346
+ },
347
+ "model.layers.14.self_attn.q_a_proj": {
348
+ "bits": 8
349
+ },
350
+ "model.layers.14.self_attn.q_b_proj": {
351
+ "bits": 8
352
+ },
353
+ "model.layers.14.self_attn.kv_a_proj_with_mqa": {
354
+ "bits": 8
355
+ },
356
+ "model.layers.14.self_attn.kv_b_proj": {
357
+ "bits": 8
358
+ },
359
+ "model.layers.14.self_attn.o_proj": {
360
+ "bits": 8
361
+ },
362
+ "model.layers.14.mlp.shared_experts.gate_proj": {
363
+ "bits": 8
364
+ },
365
+ "model.layers.14.mlp.shared_experts.up_proj": {
366
+ "bits": 8
367
+ },
368
+ "model.layers.14.mlp.shared_experts.down_proj": {
369
+ "bits": 8
370
+ },
371
+ "model.layers.15.self_attn.q_a_proj": {
372
+ "bits": 8
373
+ },
374
+ "model.layers.15.self_attn.q_b_proj": {
375
+ "bits": 8
376
+ },
377
+ "model.layers.15.self_attn.kv_a_proj_with_mqa": {
378
+ "bits": 8
379
+ },
380
+ "model.layers.15.self_attn.kv_b_proj": {
381
+ "bits": 8
382
+ },
383
+ "model.layers.15.self_attn.o_proj": {
384
+ "bits": 8
385
+ },
386
+ "model.layers.15.mlp.shared_experts.gate_proj": {
387
+ "bits": 8
388
+ },
389
+ "model.layers.15.mlp.shared_experts.up_proj": {
390
+ "bits": 8
391
+ },
392
+ "model.layers.15.mlp.shared_experts.down_proj": {
393
+ "bits": 8
394
+ },
395
+ "model.layers.16.self_attn.q_a_proj": {
396
+ "bits": 8
397
+ },
398
+ "model.layers.16.self_attn.q_b_proj": {
399
+ "bits": 8
400
+ },
401
+ "model.layers.16.self_attn.kv_a_proj_with_mqa": {
402
+ "bits": 8
403
+ },
404
+ "model.layers.16.self_attn.kv_b_proj": {
405
+ "bits": 8
406
+ },
407
+ "model.layers.16.self_attn.o_proj": {
408
+ "bits": 8
409
+ },
410
+ "model.layers.16.mlp.shared_experts.gate_proj": {
411
+ "bits": 8
412
+ },
413
+ "model.layers.16.mlp.shared_experts.up_proj": {
414
+ "bits": 8
415
+ },
416
+ "model.layers.16.mlp.shared_experts.down_proj": {
417
+ "bits": 8
418
+ },
419
+ "model.layers.17.self_attn.q_a_proj": {
420
+ "bits": 8
421
+ },
422
+ "model.layers.17.self_attn.q_b_proj": {
423
+ "bits": 8
424
+ },
425
+ "model.layers.17.self_attn.kv_a_proj_with_mqa": {
426
+ "bits": 8
427
+ },
428
+ "model.layers.17.self_attn.kv_b_proj": {
429
+ "bits": 8
430
+ },
431
+ "model.layers.17.self_attn.o_proj": {
432
+ "bits": 8
433
+ },
434
+ "model.layers.17.mlp.shared_experts.gate_proj": {
435
+ "bits": 8
436
+ },
437
+ "model.layers.17.mlp.shared_experts.up_proj": {
438
+ "bits": 8
439
+ },
440
+ "model.layers.17.mlp.shared_experts.down_proj": {
441
+ "bits": 8
442
+ },
443
+ "model.layers.18.self_attn.q_a_proj": {
444
+ "bits": 8
445
+ },
446
+ "model.layers.18.self_attn.q_b_proj": {
447
+ "bits": 8
448
+ },
449
+ "model.layers.18.self_attn.kv_a_proj_with_mqa": {
450
+ "bits": 8
451
+ },
452
+ "model.layers.18.self_attn.kv_b_proj": {
453
+ "bits": 8
454
+ },
455
+ "model.layers.18.self_attn.o_proj": {
456
+ "bits": 8
457
+ },
458
+ "model.layers.18.mlp.shared_experts.gate_proj": {
459
+ "bits": 8
460
+ },
461
+ "model.layers.18.mlp.shared_experts.up_proj": {
462
+ "bits": 8
463
+ },
464
+ "model.layers.18.mlp.shared_experts.down_proj": {
465
+ "bits": 8
466
+ },
467
+ "model.layers.19.self_attn.q_a_proj": {
468
+ "bits": 8
469
+ },
470
+ "model.layers.19.self_attn.q_b_proj": {
471
+ "bits": 8
472
+ },
473
+ "model.layers.19.self_attn.kv_a_proj_with_mqa": {
474
+ "bits": 8
475
+ },
476
+ "model.layers.19.self_attn.kv_b_proj": {
477
+ "bits": 8
478
+ },
479
+ "model.layers.19.self_attn.o_proj": {
480
+ "bits": 8
481
+ },
482
+ "model.layers.19.mlp.shared_experts.gate_proj": {
483
+ "bits": 8
484
+ },
485
+ "model.layers.19.mlp.shared_experts.up_proj": {
486
+ "bits": 8
487
+ },
488
+ "model.layers.19.mlp.shared_experts.down_proj": {
489
+ "bits": 8
490
+ },
491
+ "model.layers.20.self_attn.q_a_proj": {
492
+ "bits": 8
493
+ },
494
+ "model.layers.20.self_attn.q_b_proj": {
495
+ "bits": 8
496
+ },
497
+ "model.layers.20.self_attn.kv_a_proj_with_mqa": {
498
+ "bits": 8
499
+ },
500
+ "model.layers.20.self_attn.kv_b_proj": {
501
+ "bits": 8
502
+ },
503
+ "model.layers.20.self_attn.o_proj": {
504
+ "bits": 8
505
+ },
506
+ "model.layers.20.mlp.shared_experts.gate_proj": {
507
+ "bits": 8
508
+ },
509
+ "model.layers.20.mlp.shared_experts.up_proj": {
510
+ "bits": 8
511
+ },
512
+ "model.layers.20.mlp.shared_experts.down_proj": {
513
+ "bits": 8
514
+ },
515
+ "model.layers.21.self_attn.q_a_proj": {
516
+ "bits": 8
517
+ },
518
+ "model.layers.21.self_attn.q_b_proj": {
519
+ "bits": 8
520
+ },
521
+ "model.layers.21.self_attn.kv_a_proj_with_mqa": {
522
+ "bits": 8
523
+ },
524
+ "model.layers.21.self_attn.kv_b_proj": {
525
+ "bits": 8
526
+ },
527
+ "model.layers.21.self_attn.o_proj": {
528
+ "bits": 8
529
+ },
530
+ "model.layers.21.mlp.shared_experts.gate_proj": {
531
+ "bits": 8
532
+ },
533
+ "model.layers.21.mlp.shared_experts.up_proj": {
534
+ "bits": 8
535
+ },
536
+ "model.layers.21.mlp.shared_experts.down_proj": {
537
+ "bits": 8
538
+ },
539
+ "model.layers.22.self_attn.q_a_proj": {
540
+ "bits": 8
541
+ },
542
+ "model.layers.22.self_attn.q_b_proj": {
543
+ "bits": 8
544
+ },
545
+ "model.layers.22.self_attn.kv_a_proj_with_mqa": {
546
+ "bits": 8
547
+ },
548
+ "model.layers.22.self_attn.kv_b_proj": {
549
+ "bits": 8
550
+ },
551
+ "model.layers.22.self_attn.o_proj": {
552
+ "bits": 8
553
+ },
554
+ "model.layers.22.mlp.shared_experts.gate_proj": {
555
+ "bits": 8
556
+ },
557
+ "model.layers.22.mlp.shared_experts.up_proj": {
558
+ "bits": 8
559
+ },
560
+ "model.layers.22.mlp.shared_experts.down_proj": {
561
+ "bits": 8
562
+ },
563
+ "model.layers.23.self_attn.q_a_proj": {
564
+ "bits": 8
565
+ },
566
+ "model.layers.23.self_attn.q_b_proj": {
567
+ "bits": 8
568
+ },
569
+ "model.layers.23.self_attn.kv_a_proj_with_mqa": {
570
+ "bits": 8
571
+ },
572
+ "model.layers.23.self_attn.kv_b_proj": {
573
+ "bits": 8
574
+ },
575
+ "model.layers.23.self_attn.o_proj": {
576
+ "bits": 8
577
+ },
578
+ "model.layers.23.mlp.shared_experts.gate_proj": {
579
+ "bits": 8
580
+ },
581
+ "model.layers.23.mlp.shared_experts.up_proj": {
582
+ "bits": 8
583
+ },
584
+ "model.layers.23.mlp.shared_experts.down_proj": {
585
+ "bits": 8
586
+ },
587
+ "model.layers.24.self_attn.q_a_proj": {
588
+ "bits": 8
589
+ },
590
+ "model.layers.24.self_attn.q_b_proj": {
591
+ "bits": 8
592
+ },
593
+ "model.layers.24.self_attn.kv_a_proj_with_mqa": {
594
+ "bits": 8
595
+ },
596
+ "model.layers.24.self_attn.kv_b_proj": {
597
+ "bits": 8
598
+ },
599
+ "model.layers.24.self_attn.o_proj": {
600
+ "bits": 8
601
+ },
602
+ "model.layers.24.mlp.shared_experts.gate_proj": {
603
+ "bits": 8
604
+ },
605
+ "model.layers.24.mlp.shared_experts.up_proj": {
606
+ "bits": 8
607
+ },
608
+ "model.layers.24.mlp.shared_experts.down_proj": {
609
+ "bits": 8
610
+ },
611
+ "model.layers.25.self_attn.q_a_proj": {
612
+ "bits": 8
613
+ },
614
+ "model.layers.25.self_attn.q_b_proj": {
615
+ "bits": 8
616
+ },
617
+ "model.layers.25.self_attn.kv_a_proj_with_mqa": {
618
+ "bits": 8
619
+ },
620
+ "model.layers.25.self_attn.kv_b_proj": {
621
+ "bits": 8
622
+ },
623
+ "model.layers.25.self_attn.o_proj": {
624
+ "bits": 8
625
+ },
626
+ "model.layers.25.mlp.shared_experts.gate_proj": {
627
+ "bits": 8
628
+ },
629
+ "model.layers.25.mlp.shared_experts.up_proj": {
630
+ "bits": 8
631
+ },
632
+ "model.layers.25.mlp.shared_experts.down_proj": {
633
+ "bits": 8
634
+ },
635
+ "model.layers.26.self_attn.q_a_proj": {
636
+ "bits": 8
637
+ },
638
+ "model.layers.26.self_attn.q_b_proj": {
639
+ "bits": 8
640
+ },
641
+ "model.layers.26.self_attn.kv_a_proj_with_mqa": {
642
+ "bits": 8
643
+ },
644
+ "model.layers.26.self_attn.kv_b_proj": {
645
+ "bits": 8
646
+ },
647
+ "model.layers.26.self_attn.o_proj": {
648
+ "bits": 8
649
+ },
650
+ "model.layers.26.mlp.shared_experts.gate_proj": {
651
+ "bits": 8
652
+ },
653
+ "model.layers.26.mlp.shared_experts.up_proj": {
654
+ "bits": 8
655
+ },
656
+ "model.layers.26.mlp.shared_experts.down_proj": {
657
+ "bits": 8
658
+ },
659
+ "model.layers.27.self_attn.q_a_proj": {
660
+ "bits": 8
661
+ },
662
+ "model.layers.27.self_attn.q_b_proj": {
663
+ "bits": 8
664
+ },
665
+ "model.layers.27.self_attn.kv_a_proj_with_mqa": {
666
+ "bits": 8
667
+ },
668
+ "model.layers.27.self_attn.kv_b_proj": {
669
+ "bits": 8
670
+ },
671
+ "model.layers.27.self_attn.o_proj": {
672
+ "bits": 8
673
+ },
674
+ "model.layers.27.mlp.shared_experts.gate_proj": {
675
+ "bits": 8
676
+ },
677
+ "model.layers.27.mlp.shared_experts.up_proj": {
678
+ "bits": 8
679
+ },
680
+ "model.layers.27.mlp.shared_experts.down_proj": {
681
+ "bits": 8
682
+ },
683
+ "model.layers.28.self_attn.q_a_proj": {
684
+ "bits": 8
685
+ },
686
+ "model.layers.28.self_attn.q_b_proj": {
687
+ "bits": 8
688
+ },
689
+ "model.layers.28.self_attn.kv_a_proj_with_mqa": {
690
+ "bits": 8
691
+ },
692
+ "model.layers.28.self_attn.kv_b_proj": {
693
+ "bits": 8
694
+ },
695
+ "model.layers.28.self_attn.o_proj": {
696
+ "bits": 8
697
+ },
698
+ "model.layers.28.mlp.shared_experts.gate_proj": {
699
+ "bits": 8
700
+ },
701
+ "model.layers.28.mlp.shared_experts.up_proj": {
702
+ "bits": 8
703
+ },
704
+ "model.layers.28.mlp.shared_experts.down_proj": {
705
+ "bits": 8
706
+ },
707
+ "model.layers.29.self_attn.q_a_proj": {
708
+ "bits": 8
709
+ },
710
+ "model.layers.29.self_attn.q_b_proj": {
711
+ "bits": 8
712
+ },
713
+ "model.layers.29.self_attn.kv_a_proj_with_mqa": {
714
+ "bits": 8
715
+ },
716
+ "model.layers.29.self_attn.kv_b_proj": {
717
+ "bits": 8
718
+ },
719
+ "model.layers.29.self_attn.o_proj": {
720
+ "bits": 8
721
+ },
722
+ "model.layers.29.mlp.shared_experts.gate_proj": {
723
+ "bits": 8
724
+ },
725
+ "model.layers.29.mlp.shared_experts.up_proj": {
726
+ "bits": 8
727
+ },
728
+ "model.layers.29.mlp.shared_experts.down_proj": {
729
+ "bits": 8
730
+ },
731
+ "model.layers.30.self_attn.q_a_proj": {
732
+ "bits": 8
733
+ },
734
+ "model.layers.30.self_attn.q_b_proj": {
735
+ "bits": 8
736
+ },
737
+ "model.layers.30.self_attn.kv_a_proj_with_mqa": {
738
+ "bits": 8
739
+ },
740
+ "model.layers.30.self_attn.kv_b_proj": {
741
+ "bits": 8
742
+ },
743
+ "model.layers.30.self_attn.o_proj": {
744
+ "bits": 8
745
+ },
746
+ "model.layers.30.mlp.shared_experts.gate_proj": {
747
+ "bits": 8
748
+ },
749
+ "model.layers.30.mlp.shared_experts.up_proj": {
750
+ "bits": 8
751
+ },
752
+ "model.layers.30.mlp.shared_experts.down_proj": {
753
+ "bits": 8
754
+ },
755
+ "model.layers.31.self_attn.q_a_proj": {
756
+ "bits": 8
757
+ },
758
+ "model.layers.31.self_attn.q_b_proj": {
759
+ "bits": 8
760
+ },
761
+ "model.layers.31.self_attn.kv_a_proj_with_mqa": {
762
+ "bits": 8
763
+ },
764
+ "model.layers.31.self_attn.kv_b_proj": {
765
+ "bits": 8
766
+ },
767
+ "model.layers.31.self_attn.o_proj": {
768
+ "bits": 8
769
+ },
770
+ "model.layers.31.mlp.shared_experts.gate_proj": {
771
+ "bits": 8
772
+ },
773
+ "model.layers.31.mlp.shared_experts.up_proj": {
774
+ "bits": 8
775
+ },
776
+ "model.layers.31.mlp.shared_experts.down_proj": {
777
+ "bits": 8
778
+ },
779
+ "model.layers.32.self_attn.q_a_proj": {
780
+ "bits": 8
781
+ },
782
+ "model.layers.32.self_attn.q_b_proj": {
783
+ "bits": 8
784
+ },
785
+ "model.layers.32.self_attn.kv_a_proj_with_mqa": {
786
+ "bits": 8
787
+ },
788
+ "model.layers.32.self_attn.kv_b_proj": {
789
+ "bits": 8
790
+ },
791
+ "model.layers.32.self_attn.o_proj": {
792
+ "bits": 8
793
+ },
794
+ "model.layers.32.mlp.shared_experts.gate_proj": {
795
+ "bits": 8
796
+ },
797
+ "model.layers.32.mlp.shared_experts.up_proj": {
798
+ "bits": 8
799
+ },
800
+ "model.layers.32.mlp.shared_experts.down_proj": {
801
+ "bits": 8
802
+ },
803
+ "model.layers.33.self_attn.q_a_proj": {
804
+ "bits": 8
805
+ },
806
+ "model.layers.33.self_attn.q_b_proj": {
807
+ "bits": 8
808
+ },
809
+ "model.layers.33.self_attn.kv_a_proj_with_mqa": {
810
+ "bits": 8
811
+ },
812
+ "model.layers.33.self_attn.kv_b_proj": {
813
+ "bits": 8
814
+ },
815
+ "model.layers.33.self_attn.o_proj": {
816
+ "bits": 8
817
+ },
818
+ "model.layers.33.mlp.shared_experts.gate_proj": {
819
+ "bits": 8
820
+ },
821
+ "model.layers.33.mlp.shared_experts.up_proj": {
822
+ "bits": 8
823
+ },
824
+ "model.layers.33.mlp.shared_experts.down_proj": {
825
+ "bits": 8
826
+ },
827
+ "model.layers.34.self_attn.q_a_proj": {
828
+ "bits": 8
829
+ },
830
+ "model.layers.34.self_attn.q_b_proj": {
831
+ "bits": 8
832
+ },
833
+ "model.layers.34.self_attn.kv_a_proj_with_mqa": {
834
+ "bits": 8
835
+ },
836
+ "model.layers.34.self_attn.kv_b_proj": {
837
+ "bits": 8
838
+ },
839
+ "model.layers.34.self_attn.o_proj": {
840
+ "bits": 8
841
+ },
842
+ "model.layers.34.mlp.shared_experts.gate_proj": {
843
+ "bits": 8
844
+ },
845
+ "model.layers.34.mlp.shared_experts.up_proj": {
846
+ "bits": 8
847
+ },
848
+ "model.layers.34.mlp.shared_experts.down_proj": {
849
+ "bits": 8
850
+ },
851
+ "model.layers.35.self_attn.q_a_proj": {
852
+ "bits": 8
853
+ },
854
+ "model.layers.35.self_attn.q_b_proj": {
855
+ "bits": 8
856
+ },
857
+ "model.layers.35.self_attn.kv_a_proj_with_mqa": {
858
+ "bits": 8
859
+ },
860
+ "model.layers.35.self_attn.kv_b_proj": {
861
+ "bits": 8
862
+ },
863
+ "model.layers.35.self_attn.o_proj": {
864
+ "bits": 8
865
+ },
866
+ "model.layers.35.mlp.shared_experts.gate_proj": {
867
+ "bits": 8
868
+ },
869
+ "model.layers.35.mlp.shared_experts.up_proj": {
870
+ "bits": 8
871
+ },
872
+ "model.layers.35.mlp.shared_experts.down_proj": {
873
+ "bits": 8
874
+ },
875
+ "model.layers.36.self_attn.q_a_proj": {
876
+ "bits": 8
877
+ },
878
+ "model.layers.36.self_attn.q_b_proj": {
879
+ "bits": 8
880
+ },
881
+ "model.layers.36.self_attn.kv_a_proj_with_mqa": {
882
+ "bits": 8
883
+ },
884
+ "model.layers.36.self_attn.kv_b_proj": {
885
+ "bits": 8
886
+ },
887
+ "model.layers.36.self_attn.o_proj": {
888
+ "bits": 8
889
+ },
890
+ "model.layers.36.mlp.shared_experts.gate_proj": {
891
+ "bits": 8
892
+ },
893
+ "model.layers.36.mlp.shared_experts.up_proj": {
894
+ "bits": 8
895
+ },
896
+ "model.layers.36.mlp.shared_experts.down_proj": {
897
+ "bits": 8
898
+ },
899
+ "model.layers.37.self_attn.q_a_proj": {
900
+ "bits": 8
901
+ },
902
+ "model.layers.37.self_attn.q_b_proj": {
903
+ "bits": 8
904
+ },
905
+ "model.layers.37.self_attn.kv_a_proj_with_mqa": {
906
+ "bits": 8
907
+ },
908
+ "model.layers.37.self_attn.kv_b_proj": {
909
+ "bits": 8
910
+ },
911
+ "model.layers.37.self_attn.o_proj": {
912
+ "bits": 8
913
+ },
914
+ "model.layers.37.mlp.shared_experts.gate_proj": {
915
+ "bits": 8
916
+ },
917
+ "model.layers.37.mlp.shared_experts.up_proj": {
918
+ "bits": 8
919
+ },
920
+ "model.layers.37.mlp.shared_experts.down_proj": {
921
+ "bits": 8
922
+ },
923
+ "model.layers.38.self_attn.q_a_proj": {
924
+ "bits": 8
925
+ },
926
+ "model.layers.38.self_attn.q_b_proj": {
927
+ "bits": 8
928
+ },
929
+ "model.layers.38.self_attn.kv_a_proj_with_mqa": {
930
+ "bits": 8
931
+ },
932
+ "model.layers.38.self_attn.kv_b_proj": {
933
+ "bits": 8
934
+ },
935
+ "model.layers.38.self_attn.o_proj": {
936
+ "bits": 8
937
+ },
938
+ "model.layers.38.mlp.shared_experts.gate_proj": {
939
+ "bits": 8
940
+ },
941
+ "model.layers.38.mlp.shared_experts.up_proj": {
942
+ "bits": 8
943
+ },
944
+ "model.layers.38.mlp.shared_experts.down_proj": {
945
+ "bits": 8
946
+ },
947
+ "model.layers.39.self_attn.q_a_proj": {
948
+ "bits": 8
949
+ },
950
+ "model.layers.39.self_attn.q_b_proj": {
951
+ "bits": 8
952
+ },
953
+ "model.layers.39.self_attn.kv_a_proj_with_mqa": {
954
+ "bits": 8
955
+ },
956
+ "model.layers.39.self_attn.kv_b_proj": {
957
+ "bits": 8
958
+ },
959
+ "model.layers.39.self_attn.o_proj": {
960
+ "bits": 8
961
+ },
962
+ "model.layers.39.mlp.shared_experts.gate_proj": {
963
+ "bits": 8
964
+ },
965
+ "model.layers.39.mlp.shared_experts.up_proj": {
966
+ "bits": 8
967
+ },
968
+ "model.layers.39.mlp.shared_experts.down_proj": {
969
+ "bits": 8
970
+ },
971
+ "model.layers.40.self_attn.q_a_proj": {
972
+ "bits": 8
973
+ },
974
+ "model.layers.40.self_attn.q_b_proj": {
975
+ "bits": 8
976
+ },
977
+ "model.layers.40.self_attn.kv_a_proj_with_mqa": {
978
+ "bits": 8
979
+ },
980
+ "model.layers.40.self_attn.kv_b_proj": {
981
+ "bits": 8
982
+ },
983
+ "model.layers.40.self_attn.o_proj": {
984
+ "bits": 8
985
+ },
986
+ "model.layers.40.mlp.shared_experts.gate_proj": {
987
+ "bits": 8
988
+ },
989
+ "model.layers.40.mlp.shared_experts.up_proj": {
990
+ "bits": 8
991
+ },
992
+ "model.layers.40.mlp.shared_experts.down_proj": {
993
+ "bits": 8
994
+ },
995
+ "model.layers.41.self_attn.q_a_proj": {
996
+ "bits": 8
997
+ },
998
+ "model.layers.41.self_attn.q_b_proj": {
999
+ "bits": 8
1000
+ },
1001
+ "model.layers.41.self_attn.kv_a_proj_with_mqa": {
1002
+ "bits": 8
1003
+ },
1004
+ "model.layers.41.self_attn.kv_b_proj": {
1005
+ "bits": 8
1006
+ },
1007
+ "model.layers.41.self_attn.o_proj": {
1008
+ "bits": 8
1009
+ },
1010
+ "model.layers.41.mlp.shared_experts.gate_proj": {
1011
+ "bits": 8
1012
+ },
1013
+ "model.layers.41.mlp.shared_experts.up_proj": {
1014
+ "bits": 8
1015
+ },
1016
+ "model.layers.41.mlp.shared_experts.down_proj": {
1017
+ "bits": 8
1018
+ },
1019
+ "model.layers.42.self_attn.q_a_proj": {
1020
+ "bits": 8
1021
+ },
1022
+ "model.layers.42.self_attn.q_b_proj": {
1023
+ "bits": 8
1024
+ },
1025
+ "model.layers.42.self_attn.kv_a_proj_with_mqa": {
1026
+ "bits": 8
1027
+ },
1028
+ "model.layers.42.self_attn.kv_b_proj": {
1029
+ "bits": 8
1030
+ },
1031
+ "model.layers.42.self_attn.o_proj": {
1032
+ "bits": 8
1033
+ },
1034
+ "model.layers.42.mlp.shared_experts.gate_proj": {
1035
+ "bits": 8
1036
+ },
1037
+ "model.layers.42.mlp.shared_experts.up_proj": {
1038
+ "bits": 8
1039
+ },
1040
+ "model.layers.42.mlp.shared_experts.down_proj": {
1041
+ "bits": 8
1042
+ },
1043
+ "model.layers.43.self_attn.q_a_proj": {
1044
+ "bits": 8
1045
+ },
1046
+ "model.layers.43.self_attn.q_b_proj": {
1047
+ "bits": 8
1048
+ },
1049
+ "model.layers.43.self_attn.kv_a_proj_with_mqa": {
1050
+ "bits": 8
1051
+ },
1052
+ "model.layers.43.self_attn.kv_b_proj": {
1053
+ "bits": 8
1054
+ },
1055
+ "model.layers.43.self_attn.o_proj": {
1056
+ "bits": 8
1057
+ },
1058
+ "model.layers.43.mlp.shared_experts.gate_proj": {
1059
+ "bits": 8
1060
+ },
1061
+ "model.layers.43.mlp.shared_experts.up_proj": {
1062
+ "bits": 8
1063
+ },
1064
+ "model.layers.43.mlp.shared_experts.down_proj": {
1065
+ "bits": 8
1066
+ },
1067
+ "model.layers.44.self_attn.q_a_proj": {
1068
+ "bits": 8
1069
+ },
1070
+ "model.layers.44.self_attn.q_b_proj": {
1071
+ "bits": 8
1072
+ },
1073
+ "model.layers.44.self_attn.kv_a_proj_with_mqa": {
1074
+ "bits": 8
1075
+ },
1076
+ "model.layers.44.self_attn.kv_b_proj": {
1077
+ "bits": 8
1078
+ },
1079
+ "model.layers.44.self_attn.o_proj": {
1080
+ "bits": 8
1081
+ },
1082
+ "model.layers.44.mlp.shared_experts.gate_proj": {
1083
+ "bits": 8
1084
+ },
1085
+ "model.layers.44.mlp.shared_experts.up_proj": {
1086
+ "bits": 8
1087
+ },
1088
+ "model.layers.44.mlp.shared_experts.down_proj": {
1089
+ "bits": 8
1090
+ },
1091
+ "model.layers.45.self_attn.q_a_proj": {
1092
+ "bits": 8
1093
+ },
1094
+ "model.layers.45.self_attn.q_b_proj": {
1095
+ "bits": 8
1096
+ },
1097
+ "model.layers.45.self_attn.kv_a_proj_with_mqa": {
1098
+ "bits": 8
1099
+ },
1100
+ "model.layers.45.self_attn.kv_b_proj": {
1101
+ "bits": 8
1102
+ },
1103
+ "model.layers.45.self_attn.o_proj": {
1104
+ "bits": 8
1105
+ },
1106
+ "model.layers.45.mlp.shared_experts.gate_proj": {
1107
+ "bits": 8
1108
+ },
1109
+ "model.layers.45.mlp.shared_experts.up_proj": {
1110
+ "bits": 8
1111
+ },
1112
+ "model.layers.45.mlp.shared_experts.down_proj": {
1113
+ "bits": 8
1114
+ },
1115
+ "model.layers.46.self_attn.q_a_proj": {
1116
+ "bits": 8
1117
+ },
1118
+ "model.layers.46.self_attn.q_b_proj": {
1119
+ "bits": 8
1120
+ },
1121
+ "model.layers.46.self_attn.kv_a_proj_with_mqa": {
1122
+ "bits": 8
1123
+ },
1124
+ "model.layers.46.self_attn.kv_b_proj": {
1125
+ "bits": 8
1126
+ },
1127
+ "model.layers.46.self_attn.o_proj": {
1128
+ "bits": 8
1129
+ },
1130
+ "model.layers.46.mlp.shared_experts.gate_proj": {
1131
+ "bits": 8
1132
+ },
1133
+ "model.layers.46.mlp.shared_experts.up_proj": {
1134
+ "bits": 8
1135
+ },
1136
+ "model.layers.46.mlp.shared_experts.down_proj": {
1137
+ "bits": 8
1138
+ },
1139
+ "model.layers.47.self_attn.q_a_proj": {
1140
+ "bits": 8
1141
+ },
1142
+ "model.layers.47.self_attn.q_b_proj": {
1143
+ "bits": 8
1144
+ },
1145
+ "model.layers.47.self_attn.kv_a_proj_with_mqa": {
1146
+ "bits": 8
1147
+ },
1148
+ "model.layers.47.self_attn.kv_b_proj": {
1149
+ "bits": 8
1150
+ },
1151
+ "model.layers.47.self_attn.o_proj": {
1152
+ "bits": 8
1153
+ },
1154
+ "model.layers.47.mlp.shared_experts.gate_proj": {
1155
+ "bits": 8
1156
+ },
1157
+ "model.layers.47.mlp.shared_experts.up_proj": {
1158
+ "bits": 8
1159
+ },
1160
+ "model.layers.47.mlp.shared_experts.down_proj": {
1161
+ "bits": 8
1162
+ },
1163
+ "model.layers.48.self_attn.q_a_proj": {
1164
+ "bits": 8
1165
+ },
1166
+ "model.layers.48.self_attn.q_b_proj": {
1167
+ "bits": 8
1168
+ },
1169
+ "model.layers.48.self_attn.kv_a_proj_with_mqa": {
1170
+ "bits": 8
1171
+ },
1172
+ "model.layers.48.self_attn.kv_b_proj": {
1173
+ "bits": 8
1174
+ },
1175
+ "model.layers.48.self_attn.o_proj": {
1176
+ "bits": 8
1177
+ },
1178
+ "model.layers.48.mlp.shared_experts.gate_proj": {
1179
+ "bits": 8
1180
+ },
1181
+ "model.layers.48.mlp.shared_experts.up_proj": {
1182
+ "bits": 8
1183
+ },
1184
+ "model.layers.48.mlp.shared_experts.down_proj": {
1185
+ "bits": 8
1186
+ },
1187
+ "model.layers.49.self_attn.q_a_proj": {
1188
+ "bits": 8
1189
+ },
1190
+ "model.layers.49.self_attn.q_b_proj": {
1191
+ "bits": 8
1192
+ },
1193
+ "model.layers.49.self_attn.kv_a_proj_with_mqa": {
1194
+ "bits": 8
1195
+ },
1196
+ "model.layers.49.self_attn.kv_b_proj": {
1197
+ "bits": 8
1198
+ },
1199
+ "model.layers.49.self_attn.o_proj": {
1200
+ "bits": 8
1201
+ },
1202
+ "model.layers.49.mlp.shared_experts.gate_proj": {
1203
+ "bits": 8
1204
+ },
1205
+ "model.layers.49.mlp.shared_experts.up_proj": {
1206
+ "bits": 8
1207
+ },
1208
+ "model.layers.49.mlp.shared_experts.down_proj": {
1209
+ "bits": 8
1210
+ },
1211
+ "model.layers.50.self_attn.q_a_proj": {
1212
+ "bits": 8
1213
+ },
1214
+ "model.layers.50.self_attn.q_b_proj": {
1215
+ "bits": 8
1216
+ },
1217
+ "model.layers.50.self_attn.kv_a_proj_with_mqa": {
1218
+ "bits": 8
1219
+ },
1220
+ "model.layers.50.self_attn.kv_b_proj": {
1221
+ "bits": 8
1222
+ },
1223
+ "model.layers.50.self_attn.o_proj": {
1224
+ "bits": 8
1225
+ },
1226
+ "model.layers.50.mlp.shared_experts.gate_proj": {
1227
+ "bits": 8
1228
+ },
1229
+ "model.layers.50.mlp.shared_experts.up_proj": {
1230
+ "bits": 8
1231
+ },
1232
+ "model.layers.50.mlp.shared_experts.down_proj": {
1233
+ "bits": 8
1234
+ },
1235
+ "model.layers.51.self_attn.q_a_proj": {
1236
+ "bits": 8
1237
+ },
1238
+ "model.layers.51.self_attn.q_b_proj": {
1239
+ "bits": 8
1240
+ },
1241
+ "model.layers.51.self_attn.kv_a_proj_with_mqa": {
1242
+ "bits": 8
1243
+ },
1244
+ "model.layers.51.self_attn.kv_b_proj": {
1245
+ "bits": 8
1246
+ },
1247
+ "model.layers.51.self_attn.o_proj": {
1248
+ "bits": 8
1249
+ },
1250
+ "model.layers.51.mlp.shared_experts.gate_proj": {
1251
+ "bits": 8
1252
+ },
1253
+ "model.layers.51.mlp.shared_experts.up_proj": {
1254
+ "bits": 8
1255
+ },
1256
+ "model.layers.51.mlp.shared_experts.down_proj": {
1257
+ "bits": 8
1258
+ },
1259
+ "model.layers.52.self_attn.q_a_proj": {
1260
+ "bits": 8
1261
+ },
1262
+ "model.layers.52.self_attn.q_b_proj": {
1263
+ "bits": 8
1264
+ },
1265
+ "model.layers.52.self_attn.kv_a_proj_with_mqa": {
1266
+ "bits": 8
1267
+ },
1268
+ "model.layers.52.self_attn.kv_b_proj": {
1269
+ "bits": 8
1270
+ },
1271
+ "model.layers.52.self_attn.o_proj": {
1272
+ "bits": 8
1273
+ },
1274
+ "model.layers.52.mlp.shared_experts.gate_proj": {
1275
+ "bits": 8
1276
+ },
1277
+ "model.layers.52.mlp.shared_experts.up_proj": {
1278
+ "bits": 8
1279
+ },
1280
+ "model.layers.52.mlp.shared_experts.down_proj": {
1281
+ "bits": 8
1282
+ },
1283
+ "model.layers.53.self_attn.q_a_proj": {
1284
+ "bits": 8
1285
+ },
1286
+ "model.layers.53.self_attn.q_b_proj": {
1287
+ "bits": 8
1288
+ },
1289
+ "model.layers.53.self_attn.kv_a_proj_with_mqa": {
1290
+ "bits": 8
1291
+ },
1292
+ "model.layers.53.self_attn.kv_b_proj": {
1293
+ "bits": 8
1294
+ },
1295
+ "model.layers.53.self_attn.o_proj": {
1296
+ "bits": 8
1297
+ },
1298
+ "model.layers.53.mlp.shared_experts.gate_proj": {
1299
+ "bits": 8
1300
+ },
1301
+ "model.layers.53.mlp.shared_experts.up_proj": {
1302
+ "bits": 8
1303
+ },
1304
+ "model.layers.53.mlp.shared_experts.down_proj": {
1305
+ "bits": 8
1306
+ },
1307
+ "model.layers.54.self_attn.q_a_proj": {
1308
+ "bits": 8
1309
+ },
1310
+ "model.layers.54.self_attn.q_b_proj": {
1311
+ "bits": 8
1312
+ },
1313
+ "model.layers.54.self_attn.kv_a_proj_with_mqa": {
1314
+ "bits": 8
1315
+ },
1316
+ "model.layers.54.self_attn.kv_b_proj": {
1317
+ "bits": 8
1318
+ },
1319
+ "model.layers.54.self_attn.o_proj": {
1320
+ "bits": 8
1321
+ },
1322
+ "model.layers.54.mlp.shared_experts.gate_proj": {
1323
+ "bits": 8
1324
+ },
1325
+ "model.layers.54.mlp.shared_experts.up_proj": {
1326
+ "bits": 8
1327
+ },
1328
+ "model.layers.54.mlp.shared_experts.down_proj": {
1329
+ "bits": 8
1330
+ },
1331
+ "model.layers.55.self_attn.q_a_proj": {
1332
+ "bits": 8
1333
+ },
1334
+ "model.layers.55.self_attn.q_b_proj": {
1335
+ "bits": 8
1336
+ },
1337
+ "model.layers.55.self_attn.kv_a_proj_with_mqa": {
1338
+ "bits": 8
1339
+ },
1340
+ "model.layers.55.self_attn.kv_b_proj": {
1341
+ "bits": 8
1342
+ },
1343
+ "model.layers.55.self_attn.o_proj": {
1344
+ "bits": 8
1345
+ },
1346
+ "model.layers.55.mlp.shared_experts.gate_proj": {
1347
+ "bits": 8
1348
+ },
1349
+ "model.layers.55.mlp.shared_experts.up_proj": {
1350
+ "bits": 8
1351
+ },
1352
+ "model.layers.55.mlp.shared_experts.down_proj": {
1353
+ "bits": 8
1354
+ },
1355
+ "model.layers.56.self_attn.q_a_proj": {
1356
+ "bits": 8
1357
+ },
1358
+ "model.layers.56.self_attn.q_b_proj": {
1359
+ "bits": 8
1360
+ },
1361
+ "model.layers.56.self_attn.kv_a_proj_with_mqa": {
1362
+ "bits": 8
1363
+ },
1364
+ "model.layers.56.self_attn.kv_b_proj": {
1365
+ "bits": 8
1366
+ },
1367
+ "model.layers.56.self_attn.o_proj": {
1368
+ "bits": 8
1369
+ },
1370
+ "model.layers.56.mlp.shared_experts.gate_proj": {
1371
+ "bits": 8
1372
+ },
1373
+ "model.layers.56.mlp.shared_experts.up_proj": {
1374
+ "bits": 8
1375
+ },
1376
+ "model.layers.56.mlp.shared_experts.down_proj": {
1377
+ "bits": 8
1378
+ },
1379
+ "model.layers.57.self_attn.q_a_proj": {
1380
+ "bits": 8
1381
+ },
1382
+ "model.layers.57.self_attn.q_b_proj": {
1383
+ "bits": 8
1384
+ },
1385
+ "model.layers.57.self_attn.kv_a_proj_with_mqa": {
1386
+ "bits": 8
1387
+ },
1388
+ "model.layers.57.self_attn.kv_b_proj": {
1389
+ "bits": 8
1390
+ },
1391
+ "model.layers.57.self_attn.o_proj": {
1392
+ "bits": 8
1393
+ },
1394
+ "model.layers.57.mlp.shared_experts.gate_proj": {
1395
+ "bits": 8
1396
+ },
1397
+ "model.layers.57.mlp.shared_experts.up_proj": {
1398
+ "bits": 8
1399
+ },
1400
+ "model.layers.57.mlp.shared_experts.down_proj": {
1401
+ "bits": 8
1402
+ },
1403
+ "model.layers.58.self_attn.q_a_proj": {
1404
+ "bits": 8
1405
+ },
1406
+ "model.layers.58.self_attn.q_b_proj": {
1407
+ "bits": 8
1408
+ },
1409
+ "model.layers.58.self_attn.kv_a_proj_with_mqa": {
1410
+ "bits": 8
1411
+ },
1412
+ "model.layers.58.self_attn.kv_b_proj": {
1413
+ "bits": 8
1414
+ },
1415
+ "model.layers.58.self_attn.o_proj": {
1416
+ "bits": 8
1417
+ },
1418
+ "model.layers.58.mlp.shared_experts.gate_proj": {
1419
+ "bits": 8
1420
+ },
1421
+ "model.layers.58.mlp.shared_experts.up_proj": {
1422
+ "bits": 8
1423
+ },
1424
+ "model.layers.58.mlp.shared_experts.down_proj": {
1425
+ "bits": 8
1426
+ },
1427
+ "model.layers.59.self_attn.q_a_proj": {
1428
+ "bits": 8
1429
+ },
1430
+ "model.layers.59.self_attn.q_b_proj": {
1431
+ "bits": 8
1432
+ },
1433
+ "model.layers.59.self_attn.kv_a_proj_with_mqa": {
1434
+ "bits": 8
1435
+ },
1436
+ "model.layers.59.self_attn.kv_b_proj": {
1437
+ "bits": 8
1438
+ },
1439
+ "model.layers.59.self_attn.o_proj": {
1440
+ "bits": 8
1441
+ },
1442
+ "model.layers.59.mlp.shared_experts.gate_proj": {
1443
+ "bits": 8
1444
+ },
1445
+ "model.layers.59.mlp.shared_experts.up_proj": {
1446
+ "bits": 8
1447
+ },
1448
+ "model.layers.59.mlp.shared_experts.down_proj": {
1449
+ "bits": 8
1450
+ },
1451
+ "model.layers.60.self_attn.q_a_proj": {
1452
+ "bits": 8
1453
+ },
1454
+ "model.layers.60.self_attn.q_b_proj": {
1455
+ "bits": 8
1456
+ },
1457
+ "model.layers.60.self_attn.kv_a_proj_with_mqa": {
1458
+ "bits": 8
1459
+ },
1460
+ "model.layers.60.self_attn.kv_b_proj": {
1461
+ "bits": 8
1462
+ },
1463
+ "model.layers.60.self_attn.o_proj": {
1464
+ "bits": 8
1465
+ },
1466
+ "model.layers.60.mlp.shared_experts.gate_proj": {
1467
+ "bits": 8
1468
+ },
1469
+ "model.layers.60.mlp.shared_experts.up_proj": {
1470
+ "bits": 8
1471
+ },
1472
+ "model.layers.60.mlp.shared_experts.down_proj": {
1473
+ "bits": 8
1474
+ }
1475
+ }
1476
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|begin▁of▁sentence|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|end▁of▁sentence|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|end▁of▁sentence|>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff