Qubitium commited on
Commit
b364160
·
verified ·
1 Parent(s): 4d92fa0

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
chat_template.jinja ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0]['role'] == 'system' %}
4
+ {{- messages[0]['content'] }}
5
+ {%- else %}
6
+ {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}
7
+ {%- endif %}
8
+ {{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
9
+ {%- for tool in tools %}
10
+ {{- "\n" }}
11
+ {{- tool | tojson }}
12
+ {%- endfor %}
13
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
14
+ {%- else %}
15
+ {%- if messages[0]['role'] == 'system' %}
16
+ {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
17
+ {%- else %}
18
+ {{- '<|im_start|>system\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\n' }}
19
+ {%- endif %}
20
+ {%- endif %}
21
+ {%- for message in messages %}
22
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) or (message.role == "assistant" and not message.tool_calls) %}
23
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
24
+ {%- elif message.role == "assistant" %}
25
+ {{- '<|im_start|>' + message.role }}
26
+ {%- if message.content %}
27
+ {{- '\n' + message.content }}
28
+ {%- endif %}
29
+ {%- for tool_call in message.tool_calls %}
30
+ {%- if tool_call.function is defined %}
31
+ {%- set tool_call = tool_call.function %}
32
+ {%- endif %}
33
+ {{- '\n<tool_call>\n{"name": "' }}
34
+ {{- tool_call.name }}
35
+ {{- '", "arguments": ' }}
36
+ {{- tool_call.arguments | tojson }}
37
+ {{- '}\n</tool_call>' }}
38
+ {%- endfor %}
39
+ {{- '<|im_end|>\n' }}
40
+ {%- elif message.role == "tool" %}
41
+ {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != "tool") %}
42
+ {{- '<|im_start|>user' }}
43
+ {%- endif %}
44
+ {{- '\n<tool_response>\n' }}
45
+ {{- message.content }}
46
+ {{- '\n</tool_response>' }}
47
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
48
+ {{- '<|im_end|>\n' }}
49
+ {%- endif %}
50
+ {%- endif %}
51
+ {%- endfor %}
52
+ {%- if add_generation_prompt %}
53
+ {{- '<|im_start|>assistant\n' }}
54
+ {%- endif %}
config.json CHANGED
@@ -1,39 +1,70 @@
1
  {
2
- "_attn_implementation_autoset": true,
3
- "_name_or_path": "/monster/data/model/Qwen2.5-0.5B-Instruct",
4
  "architectures": [
5
  "Qwen2ForCausalLM"
6
  ],
7
  "attention_dropout": 0.0,
8
  "bos_token_id": 151643,
 
9
  "eos_token_id": 151645,
10
  "hidden_act": "silu",
11
  "hidden_size": 896,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 4864,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
  "max_position_embeddings": 32768,
15
  "max_window_layers": 21,
16
  "model_type": "qwen2",
17
  "num_attention_heads": 14,
18
  "num_hidden_layers": 24,
19
  "num_key_value_heads": 2,
 
20
  "quantization_config": {
21
  "bits": 4,
22
  "checkpoint_format": "gptq",
23
- "desc_act": true,
24
- "dynamic": null,
25
  "group_size": 128,
26
  "lm_head": false,
27
  "meta": {
28
- "damp_auto_increment": 0.0025,
29
- "damp_percent": 0.1,
 
 
30
  "quantizer": [
31
- "gptqmodel:1.7.0"
32
  ],
33
  "static_groups": false,
34
  "true_sequential": true,
35
- "uri": "https://github.com/modelcloud/gptqmodel"
 
 
36
  },
 
 
37
  "quant_method": "gptq",
38
  "sym": true
39
  },
@@ -42,8 +73,7 @@
42
  "rope_theta": 1000000.0,
43
  "sliding_window": null,
44
  "tie_word_embeddings": true,
45
- "torch_dtype": "bfloat16",
46
- "transformers_version": "4.49.0.dev0",
47
  "use_cache": true,
48
  "use_sliding_window": false,
49
  "vocab_size": 151936
 
1
  {
 
 
2
  "architectures": [
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
  "bos_token_id": 151643,
7
+ "dtype": "bfloat16",
8
  "eos_token_id": 151645,
9
  "hidden_act": "silu",
10
  "hidden_size": 896,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 4864,
13
+ "layer_types": [
14
+ "full_attention",
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention"
38
+ ],
39
  "max_position_embeddings": 32768,
40
  "max_window_layers": 21,
41
  "model_type": "qwen2",
42
  "num_attention_heads": 14,
43
  "num_hidden_layers": 24,
44
  "num_key_value_heads": 2,
45
+ "pad_token_id": 151662,
46
  "quantization_config": {
47
  "bits": 4,
48
  "checkpoint_format": "gptq",
49
+ "desc_act": false,
 
50
  "group_size": 128,
51
  "lm_head": false,
52
  "meta": {
53
+ "act_group_aware": true,
54
+ "damp_auto_increment": 0.01,
55
+ "damp_percent": 0.05,
56
+ "mse": 0.0,
57
  "quantizer": [
58
+ "gptqmodel:5.0.0-dev0"
59
  ],
60
  "static_groups": false,
61
  "true_sequential": true,
62
+ "uri": "https://github.com/modelcloud/gptqmodel",
63
+ "v2": false,
64
+ "v2_alpha": 0.25
65
  },
66
+ "pack_dtype": "int32",
67
+ "pack_impl": "cpu",
68
  "quant_method": "gptq",
69
  "sym": true
70
  },
 
73
  "rope_theta": 1000000.0,
74
  "sliding_window": null,
75
  "tie_word_embeddings": true,
76
+ "transformers_version": "4.57.1",
 
77
  "use_cache": true,
78
  "use_sliding_window": false,
79
  "vocab_size": 151936
generation_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "repetition_penalty": 1.1,
9
+ "top_k": 20,
10
+ "transformers_version": "4.57.1"
11
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:71f77eda26499bb3b6aa6a889f6852b0dd0e42fdbbe17ecee7fbfdf7ec631611
3
- size 731653040
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05d1802c2690026b6326f91a35e2d7740c44b526c602eb84d651c787ee358218
3
+ size 459384459
quant_log.csv CHANGED
@@ -1,169 +1,169 @@
1
- layer,module,loss,damp,time
2
- 0,self_attn.k_proj,0.00669,0.10000,0.572
3
- 0,self_attn.v_proj,0.00015,0.10000,0.366
4
- 0,self_attn.q_proj,0.04556,0.10000,0.369
5
- 0,self_attn.o_proj,0.00003,0.10000,0.365
6
- 0,mlp.up_proj,0.57997,0.10000,0.394
7
- 0,mlp.gate_proj,0.91966,0.10000,0.393
8
- 0,mlp.down_proj,0.01179,0.10000,1.944
9
- 1,self_attn.k_proj,0.06515,0.10000,0.361
10
- 1,self_attn.v_proj,0.00657,0.10000,0.371
11
- 1,self_attn.q_proj,0.22341,0.10000,0.374
12
- 1,self_attn.o_proj,0.00404,0.10000,0.374
13
- 1,mlp.up_proj,0.78053,0.10000,0.401
14
- 1,mlp.gate_proj,1.47476,0.10000,0.398
15
- 1,mlp.down_proj,0.01116,0.10000,1.960
16
- 2,self_attn.k_proj,0.13747,0.10000,0.371
17
- 2,self_attn.v_proj,0.01578,0.10000,0.367
18
- 2,self_attn.q_proj,0.46999,0.10000,0.372
19
- 2,self_attn.o_proj,0.00497,0.10000,0.370
20
- 2,mlp.up_proj,1.05533,0.10000,0.400
21
- 2,mlp.gate_proj,1.95990,0.10000,0.395
22
- 2,mlp.down_proj,0.54380,0.10000,2.036
23
- 3,self_attn.k_proj,0.14006,0.10000,0.365
24
- 3,self_attn.v_proj,0.03060,0.10000,0.370
25
- 3,self_attn.q_proj,0.59166,0.10000,0.380
26
- 3,self_attn.o_proj,0.00675,0.10000,0.380
27
- 3,mlp.up_proj,1.98003,0.10000,0.401
28
- 3,mlp.gate_proj,3.62075,0.10000,0.396
29
- 3,mlp.down_proj,0.64975,0.10000,2.032
30
- 4,self_attn.k_proj,0.09814,0.10000,0.366
31
- 4,self_attn.v_proj,0.04382,0.10000,0.366
32
- 4,self_attn.q_proj,0.46673,0.10000,0.373
33
- 4,self_attn.o_proj,0.01211,0.10000,0.377
34
- 4,mlp.up_proj,1.58077,0.10000,0.401
35
- 4,mlp.gate_proj,2.47939,0.10000,0.397
36
- 4,mlp.down_proj,0.03189,0.10000,1.972
37
- 5,self_attn.k_proj,0.12246,0.10000,0.368
38
- 5,self_attn.v_proj,0.05167,0.10000,0.368
39
- 5,self_attn.q_proj,0.61106,0.10000,0.373
40
- 5,self_attn.o_proj,0.01123,0.10000,0.375
41
- 5,mlp.up_proj,2.01079,0.10000,0.406
42
- 5,mlp.gate_proj,4.08619,0.10000,0.396
43
- 5,mlp.down_proj,0.03081,0.10000,1.992
44
- 6,self_attn.k_proj,0.11447,0.10000,0.363
45
- 6,self_attn.v_proj,0.03224,0.10000,0.361
46
- 6,self_attn.q_proj,0.50002,0.10000,0.371
47
- 6,self_attn.o_proj,0.01039,0.10000,0.371
48
- 6,mlp.up_proj,1.59279,0.10000,0.403
49
- 6,mlp.gate_proj,2.33140,0.10000,0.424
50
- 6,mlp.down_proj,0.03129,0.10000,1.943
51
- 7,self_attn.k_proj,0.17277,0.10000,0.363
52
- 7,self_attn.v_proj,0.04741,0.10000,0.364
53
- 7,self_attn.q_proj,0.70939,0.10000,0.370
54
- 7,self_attn.o_proj,0.01649,0.10000,0.373
55
- 7,mlp.up_proj,1.71943,0.10000,0.401
56
- 7,mlp.gate_proj,2.22081,0.10000,0.397
57
- 7,mlp.down_proj,0.03903,0.10000,1.962
58
- 8,self_attn.k_proj,0.11656,0.10000,0.360
59
- 8,self_attn.v_proj,0.04624,0.10000,0.365
60
- 8,self_attn.q_proj,0.79887,0.10000,0.371
61
- 8,self_attn.o_proj,0.02273,0.10000,0.368
62
- 8,mlp.up_proj,1.43981,0.10000,0.413
63
- 8,mlp.gate_proj,1.95721,0.10000,0.397
64
- 8,mlp.down_proj,0.02604,0.10000,1.942
65
- 9,self_attn.k_proj,0.33948,0.10000,0.362
66
- 9,self_attn.v_proj,0.08356,0.10000,0.362
67
- 9,self_attn.q_proj,1.38756,0.10000,0.373
68
- 9,self_attn.o_proj,0.01633,0.10000,0.374
69
- 9,mlp.up_proj,1.48946,0.10000,0.400
70
- 9,mlp.gate_proj,1.75232,0.10000,0.396
71
- 9,mlp.down_proj,0.02843,0.10000,1.953
72
- 10,self_attn.k_proj,0.10033,0.10000,0.365
73
- 10,self_attn.v_proj,0.06627,0.10000,0.364
74
- 10,self_attn.q_proj,0.50354,0.10000,0.375
75
- 10,self_attn.o_proj,0.02601,0.10000,0.373
76
- 10,mlp.up_proj,1.28366,0.10000,0.400
77
- 10,mlp.gate_proj,1.68165,0.10000,0.395
78
- 10,mlp.down_proj,0.02453,0.10000,1.970
79
- 11,self_attn.k_proj,0.35640,0.10000,0.364
80
- 11,self_attn.v_proj,0.08797,0.10000,0.363
81
- 11,self_attn.q_proj,1.44697,0.10000,0.373
82
- 11,self_attn.o_proj,0.01708,0.10000,0.374
83
- 11,mlp.up_proj,1.46764,0.10000,0.400
84
- 11,mlp.gate_proj,1.59477,0.10000,0.403
85
- 11,mlp.down_proj,0.03929,0.10000,1.959
86
- 12,self_attn.k_proj,0.09292,0.10000,0.366
87
- 12,self_attn.v_proj,0.06008,0.10000,0.366
88
- 12,self_attn.q_proj,0.50941,0.10000,0.376
89
- 12,self_attn.o_proj,0.02493,0.10000,0.407
90
- 12,mlp.up_proj,1.36413,0.10000,0.420
91
- 12,mlp.gate_proj,1.50805,0.10000,0.423
92
- 12,mlp.down_proj,0.02825,0.10000,2.090
93
- 13,self_attn.k_proj,0.17860,0.10000,0.391
94
- 13,self_attn.v_proj,0.06740,0.10000,0.392
95
- 13,self_attn.q_proj,0.91921,0.10000,0.375
96
- 13,self_attn.o_proj,0.02215,0.10000,0.405
97
- 13,mlp.up_proj,1.60880,0.10000,0.428
98
- 13,mlp.gate_proj,1.61348,0.10000,0.419
99
- 13,mlp.down_proj,0.03996,0.10000,1.946
100
- 14,self_attn.k_proj,0.12914,0.10000,0.363
101
- 14,self_attn.v_proj,0.10303,0.10000,0.360
102
- 14,self_attn.q_proj,0.79523,0.10000,0.370
103
- 14,self_attn.o_proj,0.04685,0.10000,0.372
104
- 14,mlp.up_proj,1.58363,0.10000,0.395
105
- 14,mlp.gate_proj,1.77731,0.10000,0.392
106
- 14,mlp.down_proj,0.04859,0.10000,1.937
107
- 15,self_attn.k_proj,0.13345,0.10000,0.362
108
- 15,self_attn.v_proj,0.08482,0.10000,0.361
109
- 15,self_attn.q_proj,0.72791,0.10000,0.372
110
- 15,self_attn.o_proj,0.03135,0.10000,0.371
111
- 15,mlp.up_proj,2.07139,0.10000,0.398
112
- 15,mlp.gate_proj,2.14117,0.10000,0.393
113
- 15,mlp.down_proj,0.06882,0.10000,2.049
114
- 16,self_attn.k_proj,0.25867,0.10000,0.366
115
- 16,self_attn.v_proj,0.23036,0.10000,0.368
116
- 16,self_attn.q_proj,1.52680,0.10000,0.372
117
- 16,self_attn.o_proj,0.03056,0.10000,0.404
118
- 16,mlp.up_proj,2.59174,0.10000,0.404
119
- 16,mlp.gate_proj,3.33505,0.10000,0.403
120
- 16,mlp.down_proj,0.12967,0.10000,2.131
121
- 17,self_attn.k_proj,0.15921,0.10000,0.364
122
- 17,self_attn.v_proj,0.14702,0.10000,0.363
123
- 17,self_attn.q_proj,1.03912,0.10000,0.379
124
- 17,self_attn.o_proj,0.02504,0.10000,0.373
125
- 17,mlp.up_proj,2.94703,0.10000,0.400
126
- 17,mlp.gate_proj,4.60301,0.10000,0.399
127
- 17,mlp.down_proj,0.10250,0.10000,2.064
128
- 18,self_attn.k_proj,0.17395,0.10000,0.362
129
- 18,self_attn.v_proj,0.13174,0.10000,0.361
130
- 18,self_attn.q_proj,1.01365,0.10000,0.377
131
- 18,self_attn.o_proj,0.02879,0.10000,0.382
132
- 18,mlp.up_proj,2.89602,0.10000,0.416
133
- 18,mlp.gate_proj,3.61816,0.10000,0.394
134
- 18,mlp.down_proj,0.11206,0.10000,2.045
135
- 19,self_attn.k_proj,0.15185,0.10000,0.365
136
- 19,self_attn.v_proj,0.12662,0.10000,0.366
137
- 19,self_attn.q_proj,0.91668,0.10000,0.372
138
- 19,self_attn.o_proj,0.04191,0.10000,0.377
139
- 19,mlp.up_proj,4.22402,0.10000,0.402
140
- 19,mlp.gate_proj,5.07732,0.10000,0.397
141
- 19,mlp.down_proj,0.21600,0.10000,1.942
142
- 20,self_attn.k_proj,0.21450,0.10000,0.368
143
- 20,self_attn.v_proj,0.46156,0.10000,0.375
144
- 20,self_attn.q_proj,1.34480,0.10000,0.383
145
- 20,self_attn.o_proj,0.07874,0.10000,0.373
146
- 20,mlp.up_proj,4.72404,0.10000,0.405
147
- 20,mlp.gate_proj,5.23060,0.10000,0.420
148
- 20,mlp.down_proj,0.30142,0.10000,1.931
149
- 21,self_attn.k_proj,0.25363,0.10000,0.363
150
- 21,self_attn.v_proj,0.68723,0.10000,0.361
151
- 21,self_attn.q_proj,1.54567,0.10000,0.371
152
- 21,self_attn.o_proj,0.19063,0.10000,0.373
153
- 21,mlp.up_proj,4.74235,0.10000,0.399
154
- 21,mlp.gate_proj,5.02750,0.10000,0.393
155
- 21,mlp.down_proj,1.48262,0.10000,1.988
156
- 22,self_attn.k_proj,0.24376,0.10000,0.361
157
- 22,self_attn.v_proj,0.69571,0.10000,0.368
158
- 22,self_attn.q_proj,1.45133,0.10000,0.380
159
- 22,self_attn.o_proj,0.05970,0.10000,0.381
160
- 22,mlp.up_proj,4.15480,0.10000,0.406
161
- 22,mlp.gate_proj,3.93745,0.10000,0.398
162
- 22,mlp.down_proj,0.47177,0.10000,1.975
163
- 23,self_attn.k_proj,0.30426,0.10000,0.368
164
- 23,self_attn.v_proj,0.64836,0.10000,0.365
165
- 23,self_attn.q_proj,1.41738,0.10000,0.374
166
- 23,self_attn.o_proj,0.18444,0.10000,0.374
167
- 23,mlp.up_proj,4.43292,0.10000,0.401
168
- 23,mlp.gate_proj,4.59121,0.10000,0.397
169
- 23,mlp.down_proj,0.96985,0.10000,1.943
 
1
+ layer,module,loss,samples,damp,time
2
+ 0,self_attn.k_proj,0.0000001846,0.05000,0.421
3
+ 0,self_attn.v_proj,0.0000000044,0.05000,0.623
4
+ 0,self_attn.q_proj,0.0000012105,0.05000,0.651
5
+ 0,self_attn.o_proj,0.0000000008,0.05000,0.272
6
+ 0,mlp.up_proj,0.0000175088,0.05000,0.287
7
+ 0,mlp.gate_proj,0.0000274922,0.05000,0.304
8
+ 0,mlp.down_proj,0.0000003773,0.05000,1.938
9
+ 1,self_attn.k_proj,0.0000020258,0.05000,0.378
10
+ 1,self_attn.v_proj,0.0000002284,0.05000,0.662
11
+ 1,self_attn.q_proj,0.0000073043,0.05000,0.671
12
+ 1,self_attn.o_proj,0.0000001099,0.05000,0.616
13
+ 1,mlp.up_proj,0.0000257946,0.05000,0.674
14
+ 1,mlp.gate_proj,0.0000495168,0.05000,0.693
15
+ 1,mlp.down_proj,0.0000003927,0.05000,1.795
16
+ 2,self_attn.k_proj,0.0000044331,0.05000,0.442
17
+ 2,self_attn.v_proj,0.0000006006,0.05000,0.808
18
+ 2,self_attn.q_proj,0.0000161915,0.05000,0.832
19
+ 2,self_attn.o_proj,0.0000001450,0.05000,0.437
20
+ 2,mlp.gate_proj,0.0000690778,0.05000,0.259
21
+ 2,mlp.up_proj,0.0000374052,0.05000,0.706
22
+ 2,mlp.down_proj,0.0000096427,0.05000,1.414
23
+ 3,self_attn.k_proj,0.0000051747,0.05000,0.492
24
+ 3,self_attn.v_proj,0.0000012535,0.05000,0.897
25
+ 3,self_attn.q_proj,0.0000214496,0.05000,0.907
26
+ 3,self_attn.o_proj,0.0000002033,0.05000,0.422
27
+ 3,mlp.gate_proj,0.0001435908,0.05000,0.493
28
+ 3,mlp.up_proj,0.0000814998,0.05000,0.537
29
+ 3,mlp.down_proj,0.0000165190,0.05000,1.751
30
+ 4,self_attn.v_proj,0.0000015579,0.05000,0.470
31
+ 4,self_attn.q_proj,0.0000169404,0.05000,0.474
32
+ 4,self_attn.k_proj,0.0000033608,0.05000,0.710
33
+ 4,self_attn.o_proj,0.0000004269,0.05000,0.545
34
+ 4,mlp.up_proj,0.0000605056,0.05000,0.257
35
+ 4,mlp.gate_proj,0.0000924927,0.05000,0.268
36
+ 4,mlp.down_proj,0.0000012452,0.05000,2.330
37
+ 5,self_attn.k_proj,0.0000042852,0.05000,0.248
38
+ 5,self_attn.v_proj,0.0000020662,0.05000,0.388
39
+ 5,self_attn.q_proj,0.0000212661,0.05000,0.393
40
+ 5,self_attn.o_proj,0.0000003989,0.05000,0.239
41
+ 5,mlp.gate_proj,0.0001488598,0.05000,0.466
42
+ 5,mlp.up_proj,0.0000753015,0.05000,0.479
43
+ 5,mlp.down_proj,0.0000011859,0.05000,2.195
44
+ 6,self_attn.k_proj,0.0000039774,0.05000,0.668
45
+ 6,self_attn.q_proj,0.0000173516,0.05000,1.329
46
+ 6,self_attn.v_proj,0.0000011332,0.05000,1.336
47
+ 6,self_attn.o_proj,0.0000002758,0.05000,0.556
48
+ 6,mlp.up_proj,0.0000618804,0.05000,0.244
49
+ 6,mlp.gate_proj,0.0000890550,0.05000,0.289
50
+ 6,mlp.down_proj,0.0000013619,0.05000,3.632
51
+ 7,self_attn.k_proj,0.0000060954,0.05000,0.255
52
+ 7,self_attn.q_proj,0.0000250038,0.05000,1.341
53
+ 7,self_attn.v_proj,0.0000019706,0.05000,1.345
54
+ 7,self_attn.o_proj,0.0000006489,0.05000,0.269
55
+ 7,mlp.up_proj,0.0000707421,0.05000,0.258
56
+ 7,mlp.gate_proj,0.0000907406,0.05000,0.683
57
+ 7,mlp.down_proj,0.0000017962,0.05000,3.559
58
+ 8,self_attn.k_proj,0.0000045065,0.05000,0.372
59
+ 8,self_attn.v_proj,0.0000016776,0.05000,0.670
60
+ 8,self_attn.q_proj,0.0000237909,0.05000,0.693
61
+ 8,self_attn.o_proj,0.0000007376,0.05000,0.356
62
+ 8,mlp.up_proj,0.0000616151,0.05000,0.469
63
+ 8,mlp.gate_proj,0.0000799935,0.05000,0.557
64
+ 8,mlp.down_proj,0.0000011966,0.05000,3.384
65
+ 9,self_attn.k_proj,0.0000126081,0.05000,0.343
66
+ 9,self_attn.q_proj,0.0000496791,0.05000,0.533
67
+ 9,self_attn.v_proj,0.0000033377,0.05000,0.570
68
+ 9,self_attn.o_proj,0.0000005403,0.05000,0.411
69
+ 9,mlp.gate_proj,0.0000756505,0.05000,0.502
70
+ 9,mlp.up_proj,0.0000649747,0.05000,0.518
71
+ 9,mlp.down_proj,0.0000014291,0.05000,1.314
72
+ 10,self_attn.k_proj,0.0000040391,0.05000,0.408
73
+ 10,self_attn.q_proj,0.0000211797,0.05000,0.704
74
+ 10,self_attn.v_proj,0.0000029188,0.05000,0.713
75
+ 10,self_attn.o_proj,0.0000008407,0.05000,0.387
76
+ 10,mlp.gate_proj,0.0000737496,0.05000,0.456
77
+ 10,mlp.up_proj,0.0000574715,0.05000,0.465
78
+ 10,mlp.down_proj,0.0000011740,0.05000,3.047
79
+ 11,self_attn.q_proj,0.0000515761,0.05000,0.552
80
+ 11,self_attn.v_proj,0.0000037881,0.05000,0.567
81
+ 11,self_attn.k_proj,0.0000141380,0.05000,0.632
82
+ 11,self_attn.o_proj,0.0000005721,0.05000,0.704
83
+ 11,mlp.gate_proj,0.0000717162,0.05000,0.263
84
+ 11,mlp.up_proj,0.0000662663,0.05000,0.535
85
+ 11,mlp.down_proj,0.0000017422,0.05000,2.340
86
+ 12,self_attn.q_proj,0.0000219778,0.05000,0.404
87
+ 12,self_attn.v_proj,0.0000025610,0.05000,0.417
88
+ 12,self_attn.k_proj,0.0000039857,0.05000,0.663
89
+ 12,self_attn.o_proj,0.0000009195,0.05000,0.678
90
+ 12,mlp.gate_proj,0.0000669683,0.05000,0.263
91
+ 12,mlp.up_proj,0.0000601216,0.05000,0.676
92
+ 12,mlp.down_proj,0.0000013678,0.05000,2.078
93
+ 13,self_attn.q_proj,0.0000371196,0.05000,0.511
94
+ 13,self_attn.v_proj,0.0000030710,0.05000,0.517
95
+ 13,self_attn.k_proj,0.0000074217,0.05000,0.612
96
+ 13,self_attn.o_proj,0.0000007384,0.05000,0.669
97
+ 13,mlp.gate_proj,0.0000725489,0.05000,0.232
98
+ 13,mlp.up_proj,0.0000715233,0.05000,0.550
99
+ 13,mlp.down_proj,0.0000019356,0.05000,3.418
100
+ 14,self_attn.k_proj,0.0000051945,0.05000,0.446
101
+ 14,self_attn.q_proj,0.0000330241,0.05000,0.719
102
+ 14,self_attn.v_proj,0.0000047070,0.05000,0.733
103
+ 14,self_attn.o_proj,0.0000018053,0.05000,0.223
104
+ 14,mlp.up_proj,0.0000730560,0.05000,0.249
105
+ 14,mlp.gate_proj,0.0000833288,0.05000,0.251
106
+ 14,mlp.down_proj,0.0000024172,0.05000,1.230
107
+ 15,self_attn.k_proj,0.0000056382,0.05000,0.430
108
+ 15,self_attn.v_proj,0.0000039362,0.05000,0.856
109
+ 15,self_attn.q_proj,0.0000289588,0.05000,0.860
110
+ 15,self_attn.o_proj,0.0000010586,0.05000,0.353
111
+ 15,mlp.gate_proj,0.0000967614,0.05000,0.617
112
+ 15,mlp.up_proj,0.0000926957,0.05000,0.702
113
+ 15,mlp.down_proj,0.0000031419,0.05000,2.435
114
+ 16,self_attn.k_proj,0.0000109810,0.05000,0.443
115
+ 16,self_attn.v_proj,0.0000096741,0.05000,0.734
116
+ 16,self_attn.q_proj,0.0000628785,0.05000,0.748
117
+ 16,self_attn.o_proj,0.0000011271,0.05000,0.389
118
+ 16,mlp.gate_proj,0.0001423836,0.05000,0.274
119
+ 16,mlp.up_proj,0.0001109843,0.05000,0.730
120
+ 16,mlp.down_proj,0.0000060371,0.05000,1.237
121
+ 17,self_attn.k_proj,0.0000060039,0.05000,0.309
122
+ 17,self_attn.q_proj,0.0000400760,0.05000,0.681
123
+ 17,self_attn.v_proj,0.0000059069,0.05000,0.682
124
+ 17,self_attn.o_proj,0.0000008255,0.05000,0.386
125
+ 17,mlp.up_proj,0.0001233144,0.05000,0.384
126
+ 17,mlp.gate_proj,0.0001934214,0.05000,0.412
127
+ 17,mlp.down_proj,0.0000050482,0.05000,2.300
128
+ 18,self_attn.q_proj,0.0000411873,0.05000,0.357
129
+ 18,self_attn.v_proj,0.0000059423,0.05000,0.369
130
+ 18,self_attn.k_proj,0.0000077203,0.05000,0.678
131
+ 18,self_attn.o_proj,0.0000010869,0.05000,0.599
132
+ 18,mlp.up_proj,0.0001260548,0.05000,0.258
133
+ 18,mlp.gate_proj,0.0001582911,0.05000,0.264
134
+ 18,mlp.down_proj,0.0000059834,0.05000,2.446
135
+ 19,self_attn.k_proj,0.0000063490,0.05000,0.272
136
+ 19,self_attn.q_proj,0.0000380589,0.05000,0.666
137
+ 19,self_attn.v_proj,0.0000057355,0.05000,0.670
138
+ 19,self_attn.o_proj,0.0000013602,0.05000,0.247
139
+ 19,mlp.up_proj,0.0001832222,0.05000,0.367
140
+ 19,mlp.gate_proj,0.0002190037,0.05000,0.372
141
+ 19,mlp.down_proj,0.0000113082,0.05000,2.235
142
+ 20,self_attn.k_proj,0.0000078660,0.05000,0.290
143
+ 20,self_attn.q_proj,0.0000533177,0.05000,0.443
144
+ 20,self_attn.v_proj,0.0000177706,0.05000,0.459
145
+ 20,self_attn.o_proj,0.0000024812,0.05000,0.569
146
+ 20,mlp.up_proj,0.0002066750,0.05000,0.240
147
+ 20,mlp.gate_proj,0.0002273082,0.05000,0.717
148
+ 20,mlp.down_proj,0.0000155725,0.05000,2.551
149
+ 21,self_attn.k_proj,0.0000082325,0.05000,0.232
150
+ 21,self_attn.q_proj,0.0000596139,0.05000,1.457
151
+ 21,self_attn.v_proj,0.0000242640,0.05000,1.457
152
+ 21,self_attn.o_proj,0.0000054829,0.05000,0.257
153
+ 21,mlp.up_proj,0.0002139384,0.05000,0.277
154
+ 21,mlp.gate_proj,0.0002226604,0.05000,0.695
155
+ 21,mlp.down_proj,0.0000279960,0.05000,2.661
156
+ 22,self_attn.k_proj,0.0000075570,0.05000,0.458
157
+ 22,self_attn.v_proj,0.0000210302,0.05000,0.811
158
+ 22,self_attn.q_proj,0.0000556213,0.05000,0.815
159
+ 22,self_attn.o_proj,0.0000020455,0.05000,0.256
160
+ 22,mlp.up_proj,0.0001812818,0.05000,0.505
161
+ 22,mlp.gate_proj,0.0001728906,0.05000,0.536
162
+ 22,mlp.down_proj,0.0000204518,0.05000,1.919
163
+ 23,self_attn.k_proj,0.0000085684,0.05000,0.263
164
+ 23,self_attn.v_proj,0.0000168603,0.05000,0.415
165
+ 23,self_attn.q_proj,0.0000655303,0.05000,0.421
166
+ 23,self_attn.o_proj,0.0000053465,0.05000,0.242
167
+ 23,mlp.gate_proj,0.0001968066,0.05000,0.261
168
+ 23,mlp.up_proj,0.0001895685,0.05000,0.267
169
+ 23,mlp.down_proj,0.0000390435,0.05000,1.745
quantize_config.json CHANGED
@@ -1,21 +1,25 @@
1
  {
2
  "bits": 4,
3
- "dynamic": null,
4
  "group_size": 128,
5
- "desc_act": true,
6
  "sym": true,
7
  "lm_head": false,
8
  "quant_method": "gptq",
9
  "checkpoint_format": "gptq",
 
10
  "meta": {
11
  "quantizer": [
12
- "gptqmodel:1.7.0"
13
  ],
14
  "uri": "https://github.com/modelcloud/gptqmodel",
15
- "damp_percent": 0.1,
16
- "damp_auto_increment": 0.0025,
17
  "static_groups": false,
18
  "true_sequential": true,
19
- "mse": 2.4
20
- }
 
 
 
 
21
  }
 
1
  {
2
  "bits": 4,
 
3
  "group_size": 128,
4
+ "desc_act": false,
5
  "sym": true,
6
  "lm_head": false,
7
  "quant_method": "gptq",
8
  "checkpoint_format": "gptq",
9
+ "pack_dtype": "int32",
10
  "meta": {
11
  "quantizer": [
12
+ "gptqmodel:5.0.0-dev0"
13
  ],
14
  "uri": "https://github.com/modelcloud/gptqmodel",
15
+ "damp_percent": 0.05,
16
+ "damp_auto_increment": 0.01,
17
  "static_groups": false,
18
  "true_sequential": true,
19
+ "mse": 0.0,
20
+ "v2": false,
21
+ "v2_alpha": 0.25,
22
+ "act_group_aware": true
23
+ },
24
+ "pack_impl": "cpu"
25
  }
special_tokens_map.json CHANGED
@@ -21,5 +21,5 @@
21
  "rstrip": false,
22
  "single_word": false
23
  },
24
- "pad_token": "<unk>"
25
  }
 
21
  "rstrip": false,
22
  "single_word": false
23
  },
24
+ "pad_token": "<|fim_pad|>"
25
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -160,7 +160,7 @@
160
  "normalized": false,
161
  "rstrip": false,
162
  "single_word": false,
163
- "special": false
164
  },
165
  "151663": {
166
  "content": "<|repo_name|>",
@@ -195,13 +195,13 @@
195
  "<|video_pad|>"
196
  ],
197
  "bos_token": null,
198
- "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
199
  "clean_up_tokenization_spaces": false,
200
  "eos_token": "<|im_end|>",
201
  "errors": "replace",
 
202
  "model_max_length": 131072,
203
- "pad_token": "<|endoftext|>",
204
  "split_special_tokens": false,
205
  "tokenizer_class": "Qwen2Tokenizer",
206
  "unk_token": null
207
- }
 
160
  "normalized": false,
161
  "rstrip": false,
162
  "single_word": false,
163
+ "special": true
164
  },
165
  "151663": {
166
  "content": "<|repo_name|>",
 
195
  "<|video_pad|>"
196
  ],
197
  "bos_token": null,
 
198
  "clean_up_tokenization_spaces": false,
199
  "eos_token": "<|im_end|>",
200
  "errors": "replace",
201
+ "extra_special_tokens": {},
202
  "model_max_length": 131072,
203
+ "pad_token": "<|fim_pad|>",
204
  "split_special_tokens": false,
205
  "tokenizer_class": "Qwen2Tokenizer",
206
  "unk_token": null
207
+ }