Upload Qwen2ForCausalLM
Browse files- config.json +0 -2
- model-00001-of-00002.safetensors +2 -2
- model-00002-of-00002.safetensors +2 -2
- model.safetensors.index.json +148 -113
config.json
CHANGED
|
@@ -39,9 +39,7 @@
|
|
| 39 |
"model.layers.2.mlp",
|
| 40 |
"model.layers.3.self_attn",
|
| 41 |
"model.layers.3.mlp",
|
| 42 |
-
"model.layers.4.self_attn",
|
| 43 |
"model.layers.4.mlp",
|
| 44 |
-
"model.layers.6.mlp",
|
| 45 |
"model.layers.11.mlp",
|
| 46 |
"model.layers.24.mlp",
|
| 47 |
"model.layers.25.mlp",
|
|
|
|
| 39 |
"model.layers.2.mlp",
|
| 40 |
"model.layers.3.self_attn",
|
| 41 |
"model.layers.3.mlp",
|
|
|
|
| 42 |
"model.layers.4.mlp",
|
|
|
|
| 43 |
"model.layers.11.mlp",
|
| 44 |
"model.layers.24.mlp",
|
| 45 |
"model.layers.25.mlp",
|
model-00001-of-00002.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d89951edcbded32eea6718bf5957114c7151be76c8be277a381fa98f7105209b
|
| 3 |
+
size 4967113023
|
model-00002-of-00002.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:32152fcce250c61031324ce54a600de4e483becc307df492b0c910314b5cf2cd
|
| 3 |
+
size 3518593782
|
model.safetensors.index.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
| 3 |
-
"total_size":
|
| 4 |
},
|
| 5 |
"weight_map": {
|
| 6 |
"lm_head.weight": "model-00002-of-00002.safetensors",
|
|
@@ -155,26 +155,26 @@
|
|
| 155 |
"model.layers.12.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 156 |
"model.layers.12.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 157 |
"model.layers.12.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 158 |
-
"model.layers.13.input_layernorm.weight": "model-
|
| 159 |
-
"model.layers.13.mlp.down_proj.weight": "model-
|
| 160 |
-
"model.layers.13.mlp.down_proj.weight.absmax": "model-
|
| 161 |
-
"model.layers.13.mlp.down_proj.weight.nested_absmax": "model-
|
| 162 |
-
"model.layers.13.mlp.down_proj.weight.nested_quant_map": "model-
|
| 163 |
-
"model.layers.13.mlp.down_proj.weight.quant_map": "model-
|
| 164 |
-
"model.layers.13.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-
|
| 165 |
-
"model.layers.13.mlp.gate_proj.weight": "model-
|
| 166 |
-
"model.layers.13.mlp.gate_proj.weight.absmax": "model-
|
| 167 |
-
"model.layers.13.mlp.gate_proj.weight.nested_absmax": "model-
|
| 168 |
-
"model.layers.13.mlp.gate_proj.weight.nested_quant_map": "model-
|
| 169 |
-
"model.layers.13.mlp.gate_proj.weight.quant_map": "model-
|
| 170 |
-
"model.layers.13.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-
|
| 171 |
-
"model.layers.13.mlp.up_proj.weight": "model-
|
| 172 |
-
"model.layers.13.mlp.up_proj.weight.absmax": "model-
|
| 173 |
-
"model.layers.13.mlp.up_proj.weight.nested_absmax": "model-
|
| 174 |
-
"model.layers.13.mlp.up_proj.weight.nested_quant_map": "model-
|
| 175 |
-
"model.layers.13.mlp.up_proj.weight.quant_map": "model-
|
| 176 |
-
"model.layers.13.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-
|
| 177 |
-
"model.layers.13.post_attention_layernorm.weight": "model-
|
| 178 |
"model.layers.13.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
| 179 |
"model.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
| 180 |
"model.layers.13.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
|
@@ -182,12 +182,12 @@
|
|
| 182 |
"model.layers.13.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 183 |
"model.layers.13.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 184 |
"model.layers.13.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 185 |
-
"model.layers.13.self_attn.o_proj.weight": "model-
|
| 186 |
-
"model.layers.13.self_attn.o_proj.weight.absmax": "model-
|
| 187 |
-
"model.layers.13.self_attn.o_proj.weight.nested_absmax": "model-
|
| 188 |
-
"model.layers.13.self_attn.o_proj.weight.nested_quant_map": "model-
|
| 189 |
-
"model.layers.13.self_attn.o_proj.weight.quant_map": "model-
|
| 190 |
-
"model.layers.13.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-
|
| 191 |
"model.layers.13.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
| 192 |
"model.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
| 193 |
"model.layers.13.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
|
@@ -202,53 +202,53 @@
|
|
| 202 |
"model.layers.13.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 203 |
"model.layers.13.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 204 |
"model.layers.13.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 205 |
-
"model.layers.14.input_layernorm.weight": "model-
|
| 206 |
-
"model.layers.14.mlp.down_proj.weight": "model-
|
| 207 |
-
"model.layers.14.mlp.down_proj.weight.absmax": "model-
|
| 208 |
-
"model.layers.14.mlp.down_proj.weight.nested_absmax": "model-
|
| 209 |
-
"model.layers.14.mlp.down_proj.weight.nested_quant_map": "model-
|
| 210 |
-
"model.layers.14.mlp.down_proj.weight.quant_map": "model-
|
| 211 |
-
"model.layers.14.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-
|
| 212 |
-
"model.layers.14.mlp.gate_proj.weight": "model-
|
| 213 |
-
"model.layers.14.mlp.gate_proj.weight.absmax": "model-
|
| 214 |
-
"model.layers.14.mlp.gate_proj.weight.nested_absmax": "model-
|
| 215 |
-
"model.layers.14.mlp.gate_proj.weight.nested_quant_map": "model-
|
| 216 |
-
"model.layers.14.mlp.gate_proj.weight.quant_map": "model-
|
| 217 |
-
"model.layers.14.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-
|
| 218 |
-
"model.layers.14.mlp.up_proj.weight": "model-
|
| 219 |
-
"model.layers.14.mlp.up_proj.weight.absmax": "model-
|
| 220 |
-
"model.layers.14.mlp.up_proj.weight.nested_absmax": "model-
|
| 221 |
-
"model.layers.14.mlp.up_proj.weight.nested_quant_map": "model-
|
| 222 |
-
"model.layers.14.mlp.up_proj.weight.quant_map": "model-
|
| 223 |
-
"model.layers.14.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-
|
| 224 |
-
"model.layers.14.post_attention_layernorm.weight": "model-
|
| 225 |
-
"model.layers.14.self_attn.k_proj.bias": "model-
|
| 226 |
-
"model.layers.14.self_attn.k_proj.weight": "model-
|
| 227 |
-
"model.layers.14.self_attn.k_proj.weight.absmax": "model-
|
| 228 |
-
"model.layers.14.self_attn.k_proj.weight.nested_absmax": "model-
|
| 229 |
-
"model.layers.14.self_attn.k_proj.weight.nested_quant_map": "model-
|
| 230 |
-
"model.layers.14.self_attn.k_proj.weight.quant_map": "model-
|
| 231 |
-
"model.layers.14.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-
|
| 232 |
-
"model.layers.14.self_attn.o_proj.weight": "model-
|
| 233 |
-
"model.layers.14.self_attn.o_proj.weight.absmax": "model-
|
| 234 |
-
"model.layers.14.self_attn.o_proj.weight.nested_absmax": "model-
|
| 235 |
-
"model.layers.14.self_attn.o_proj.weight.nested_quant_map": "model-
|
| 236 |
-
"model.layers.14.self_attn.o_proj.weight.quant_map": "model-
|
| 237 |
-
"model.layers.14.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-
|
| 238 |
-
"model.layers.14.self_attn.q_proj.bias": "model-
|
| 239 |
-
"model.layers.14.self_attn.q_proj.weight": "model-
|
| 240 |
-
"model.layers.14.self_attn.q_proj.weight.absmax": "model-
|
| 241 |
-
"model.layers.14.self_attn.q_proj.weight.nested_absmax": "model-
|
| 242 |
-
"model.layers.14.self_attn.q_proj.weight.nested_quant_map": "model-
|
| 243 |
-
"model.layers.14.self_attn.q_proj.weight.quant_map": "model-
|
| 244 |
-
"model.layers.14.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-
|
| 245 |
-
"model.layers.14.self_attn.v_proj.bias": "model-
|
| 246 |
-
"model.layers.14.self_attn.v_proj.weight": "model-
|
| 247 |
-
"model.layers.14.self_attn.v_proj.weight.absmax": "model-
|
| 248 |
-
"model.layers.14.self_attn.v_proj.weight.nested_absmax": "model-
|
| 249 |
-
"model.layers.14.self_attn.v_proj.weight.nested_quant_map": "model-
|
| 250 |
-
"model.layers.14.self_attn.v_proj.weight.quant_map": "model-
|
| 251 |
-
"model.layers.14.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-
|
| 252 |
"model.layers.15.input_layernorm.weight": "model-00002-of-00002.safetensors",
|
| 253 |
"model.layers.15.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
|
| 254 |
"model.layers.15.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
|
|
@@ -256,46 +256,46 @@
|
|
| 256 |
"model.layers.15.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
|
| 257 |
"model.layers.15.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
|
| 258 |
"model.layers.15.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
|
| 259 |
-
"model.layers.15.mlp.gate_proj.weight": "model-
|
| 260 |
-
"model.layers.15.mlp.gate_proj.weight.absmax": "model-
|
| 261 |
-
"model.layers.15.mlp.gate_proj.weight.nested_absmax": "model-
|
| 262 |
-
"model.layers.15.mlp.gate_proj.weight.nested_quant_map": "model-
|
| 263 |
-
"model.layers.15.mlp.gate_proj.weight.quant_map": "model-
|
| 264 |
-
"model.layers.15.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-
|
| 265 |
-
"model.layers.15.mlp.up_proj.weight": "model-
|
| 266 |
-
"model.layers.15.mlp.up_proj.weight.absmax": "model-
|
| 267 |
-
"model.layers.15.mlp.up_proj.weight.nested_absmax": "model-
|
| 268 |
-
"model.layers.15.mlp.up_proj.weight.nested_quant_map": "model-
|
| 269 |
-
"model.layers.15.mlp.up_proj.weight.quant_map": "model-
|
| 270 |
-
"model.layers.15.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-
|
| 271 |
"model.layers.15.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
|
| 272 |
-
"model.layers.15.self_attn.k_proj.bias": "model-
|
| 273 |
-
"model.layers.15.self_attn.k_proj.weight": "model-
|
| 274 |
-
"model.layers.15.self_attn.k_proj.weight.absmax": "model-
|
| 275 |
-
"model.layers.15.self_attn.k_proj.weight.nested_absmax": "model-
|
| 276 |
-
"model.layers.15.self_attn.k_proj.weight.nested_quant_map": "model-
|
| 277 |
-
"model.layers.15.self_attn.k_proj.weight.quant_map": "model-
|
| 278 |
-
"model.layers.15.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-
|
| 279 |
-
"model.layers.15.self_attn.o_proj.weight": "model-
|
| 280 |
-
"model.layers.15.self_attn.o_proj.weight.absmax": "model-
|
| 281 |
-
"model.layers.15.self_attn.o_proj.weight.nested_absmax": "model-
|
| 282 |
-
"model.layers.15.self_attn.o_proj.weight.nested_quant_map": "model-
|
| 283 |
-
"model.layers.15.self_attn.o_proj.weight.quant_map": "model-
|
| 284 |
-
"model.layers.15.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-
|
| 285 |
-
"model.layers.15.self_attn.q_proj.bias": "model-
|
| 286 |
-
"model.layers.15.self_attn.q_proj.weight": "model-
|
| 287 |
-
"model.layers.15.self_attn.q_proj.weight.absmax": "model-
|
| 288 |
-
"model.layers.15.self_attn.q_proj.weight.nested_absmax": "model-
|
| 289 |
-
"model.layers.15.self_attn.q_proj.weight.nested_quant_map": "model-
|
| 290 |
-
"model.layers.15.self_attn.q_proj.weight.quant_map": "model-
|
| 291 |
-
"model.layers.15.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-
|
| 292 |
-
"model.layers.15.self_attn.v_proj.bias": "model-
|
| 293 |
-
"model.layers.15.self_attn.v_proj.weight": "model-
|
| 294 |
-
"model.layers.15.self_attn.v_proj.weight.absmax": "model-
|
| 295 |
-
"model.layers.15.self_attn.v_proj.weight.nested_absmax": "model-
|
| 296 |
-
"model.layers.15.self_attn.v_proj.weight.nested_quant_map": "model-
|
| 297 |
-
"model.layers.15.self_attn.v_proj.weight.quant_map": "model-
|
| 298 |
-
"model.layers.15.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-
|
| 299 |
"model.layers.16.input_layernorm.weight": "model-00002-of-00002.safetensors",
|
| 300 |
"model.layers.16.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
|
| 301 |
"model.layers.16.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
|
|
@@ -826,11 +826,31 @@
|
|
| 826 |
"model.layers.4.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
|
| 827 |
"model.layers.4.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
| 828 |
"model.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 829 |
"model.layers.4.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 830 |
"model.layers.4.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
| 831 |
"model.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 832 |
"model.layers.4.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
| 833 |
"model.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 834 |
"model.layers.5.input_layernorm.weight": "model-00001-of-00002.safetensors",
|
| 835 |
"model.layers.5.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
|
| 836 |
"model.layers.5.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
|
@@ -880,8 +900,23 @@
|
|
| 880 |
"model.layers.5.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 881 |
"model.layers.6.input_layernorm.weight": "model-00001-of-00002.safetensors",
|
| 882 |
"model.layers.6.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 883 |
"model.layers.6.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 884 |
"model.layers.6.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 885 |
"model.layers.6.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
|
| 886 |
"model.layers.6.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
| 887 |
"model.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
|
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
| 3 |
+
"total_size": 8485578781
|
| 4 |
},
|
| 5 |
"weight_map": {
|
| 6 |
"lm_head.weight": "model-00002-of-00002.safetensors",
|
|
|
|
| 155 |
"model.layers.12.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 156 |
"model.layers.12.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 157 |
"model.layers.12.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 158 |
+
"model.layers.13.input_layernorm.weight": "model-00001-of-00002.safetensors",
|
| 159 |
+
"model.layers.13.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
|
| 160 |
+
"model.layers.13.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 161 |
+
"model.layers.13.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 162 |
+
"model.layers.13.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 163 |
+
"model.layers.13.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 164 |
+
"model.layers.13.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 165 |
+
"model.layers.13.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
|
| 166 |
+
"model.layers.13.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 167 |
+
"model.layers.13.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 168 |
+
"model.layers.13.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 169 |
+
"model.layers.13.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 170 |
+
"model.layers.13.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 171 |
+
"model.layers.13.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
|
| 172 |
+
"model.layers.13.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 173 |
+
"model.layers.13.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 174 |
+
"model.layers.13.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 175 |
+
"model.layers.13.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 176 |
+
"model.layers.13.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 177 |
+
"model.layers.13.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
|
| 178 |
"model.layers.13.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
| 179 |
"model.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
| 180 |
"model.layers.13.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
|
|
|
| 182 |
"model.layers.13.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 183 |
"model.layers.13.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 184 |
"model.layers.13.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 185 |
+
"model.layers.13.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
|
| 186 |
+
"model.layers.13.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 187 |
+
"model.layers.13.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 188 |
+
"model.layers.13.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 189 |
+
"model.layers.13.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 190 |
+
"model.layers.13.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 191 |
"model.layers.13.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
| 192 |
"model.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
| 193 |
"model.layers.13.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
|
|
|
| 202 |
"model.layers.13.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 203 |
"model.layers.13.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 204 |
"model.layers.13.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 205 |
+
"model.layers.14.input_layernorm.weight": "model-00001-of-00002.safetensors",
|
| 206 |
+
"model.layers.14.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
|
| 207 |
+
"model.layers.14.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 208 |
+
"model.layers.14.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 209 |
+
"model.layers.14.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 210 |
+
"model.layers.14.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 211 |
+
"model.layers.14.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 212 |
+
"model.layers.14.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
|
| 213 |
+
"model.layers.14.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 214 |
+
"model.layers.14.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 215 |
+
"model.layers.14.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 216 |
+
"model.layers.14.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 217 |
+
"model.layers.14.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 218 |
+
"model.layers.14.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
|
| 219 |
+
"model.layers.14.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 220 |
+
"model.layers.14.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 221 |
+
"model.layers.14.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 222 |
+
"model.layers.14.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 223 |
+
"model.layers.14.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 224 |
+
"model.layers.14.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
|
| 225 |
+
"model.layers.14.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
| 226 |
+
"model.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
| 227 |
+
"model.layers.14.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 228 |
+
"model.layers.14.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 229 |
+
"model.layers.14.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 230 |
+
"model.layers.14.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 231 |
+
"model.layers.14.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 232 |
+
"model.layers.14.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
|
| 233 |
+
"model.layers.14.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 234 |
+
"model.layers.14.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 235 |
+
"model.layers.14.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 236 |
+
"model.layers.14.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 237 |
+
"model.layers.14.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 238 |
+
"model.layers.14.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
| 239 |
+
"model.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
| 240 |
+
"model.layers.14.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 241 |
+
"model.layers.14.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 242 |
+
"model.layers.14.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 243 |
+
"model.layers.14.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 244 |
+
"model.layers.14.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 245 |
+
"model.layers.14.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
| 246 |
+
"model.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
| 247 |
+
"model.layers.14.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 248 |
+
"model.layers.14.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 249 |
+
"model.layers.14.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 250 |
+
"model.layers.14.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 251 |
+
"model.layers.14.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 252 |
"model.layers.15.input_layernorm.weight": "model-00002-of-00002.safetensors",
|
| 253 |
"model.layers.15.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
|
| 254 |
"model.layers.15.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
|
|
|
|
| 256 |
"model.layers.15.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
|
| 257 |
"model.layers.15.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
|
| 258 |
"model.layers.15.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
|
| 259 |
+
"model.layers.15.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
|
| 260 |
+
"model.layers.15.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 261 |
+
"model.layers.15.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 262 |
+
"model.layers.15.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 263 |
+
"model.layers.15.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 264 |
+
"model.layers.15.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 265 |
+
"model.layers.15.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
|
| 266 |
+
"model.layers.15.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 267 |
+
"model.layers.15.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 268 |
+
"model.layers.15.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 269 |
+
"model.layers.15.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 270 |
+
"model.layers.15.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 271 |
"model.layers.15.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
|
| 272 |
+
"model.layers.15.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
| 273 |
+
"model.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
| 274 |
+
"model.layers.15.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 275 |
+
"model.layers.15.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 276 |
+
"model.layers.15.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 277 |
+
"model.layers.15.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 278 |
+
"model.layers.15.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 279 |
+
"model.layers.15.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
|
| 280 |
+
"model.layers.15.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 281 |
+
"model.layers.15.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 282 |
+
"model.layers.15.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 283 |
+
"model.layers.15.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 284 |
+
"model.layers.15.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 285 |
+
"model.layers.15.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
| 286 |
+
"model.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
| 287 |
+
"model.layers.15.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 288 |
+
"model.layers.15.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 289 |
+
"model.layers.15.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 290 |
+
"model.layers.15.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 291 |
+
"model.layers.15.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 292 |
+
"model.layers.15.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
| 293 |
+
"model.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
| 294 |
+
"model.layers.15.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 295 |
+
"model.layers.15.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 296 |
+
"model.layers.15.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 297 |
+
"model.layers.15.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 298 |
+
"model.layers.15.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 299 |
"model.layers.16.input_layernorm.weight": "model-00002-of-00002.safetensors",
|
| 300 |
"model.layers.16.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
|
| 301 |
"model.layers.16.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
|
|
|
|
| 826 |
"model.layers.4.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
|
| 827 |
"model.layers.4.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
| 828 |
"model.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
| 829 |
+
"model.layers.4.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 830 |
+
"model.layers.4.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 831 |
+
"model.layers.4.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 832 |
+
"model.layers.4.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 833 |
+
"model.layers.4.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 834 |
"model.layers.4.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
|
| 835 |
+
"model.layers.4.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 836 |
+
"model.layers.4.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 837 |
+
"model.layers.4.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 838 |
+
"model.layers.4.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 839 |
+
"model.layers.4.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 840 |
"model.layers.4.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
|
| 841 |
"model.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
|
| 842 |
+
"model.layers.4.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 843 |
+
"model.layers.4.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 844 |
+
"model.layers.4.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 845 |
+
"model.layers.4.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 846 |
+
"model.layers.4.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 847 |
"model.layers.4.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
|
| 848 |
"model.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
| 849 |
+
"model.layers.4.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 850 |
+
"model.layers.4.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 851 |
+
"model.layers.4.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 852 |
+
"model.layers.4.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 853 |
+
"model.layers.4.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 854 |
"model.layers.5.input_layernorm.weight": "model-00001-of-00002.safetensors",
|
| 855 |
"model.layers.5.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
|
| 856 |
"model.layers.5.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
|
|
|
| 900 |
"model.layers.5.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 901 |
"model.layers.6.input_layernorm.weight": "model-00001-of-00002.safetensors",
|
| 902 |
"model.layers.6.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
|
| 903 |
+
"model.layers.6.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 904 |
+
"model.layers.6.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 905 |
+
"model.layers.6.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 906 |
+
"model.layers.6.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 907 |
+
"model.layers.6.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 908 |
"model.layers.6.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
|
| 909 |
+
"model.layers.6.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 910 |
+
"model.layers.6.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 911 |
+
"model.layers.6.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 912 |
+
"model.layers.6.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 913 |
+
"model.layers.6.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 914 |
"model.layers.6.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
|
| 915 |
+
"model.layers.6.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
|
| 916 |
+
"model.layers.6.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
|
| 917 |
+
"model.layers.6.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
|
| 918 |
+
"model.layers.6.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
|
| 919 |
+
"model.layers.6.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
|
| 920 |
"model.layers.6.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
|
| 921 |
"model.layers.6.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
|
| 922 |
"model.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|