Rob P Smith commited on
Commit
42ae119
·
1 Parent(s): 1ac471a

Configure Git LFS for large model files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,10 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ *.index.json filter=lfs diff=lfs merge=lfs -text
37
+ merges.txt filter=lfs diff=lfs merge=lfs -text
38
+ README.md filter=lfs diff=lfs merge=lfs -text
39
+ recipe.yaml filter=lfs diff=lfs merge=lfs -text
40
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
41
+ chat_template.jinja filter=lfs diff=lfs merge=lfs -text
42
+ tokenizer_config.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,4 +1,3 @@
1
- ---
2
- license: apache-2.0
3
- ---
4
- Temp
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d436dd151f7f319ce8580292bdf548302a5cbcffa8760cc9324f73a6417ee5ba
3
+ size 32
 
added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }
chat_template.jinja ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:672e747c77e990320152343b0a4951222e40de5645297905d89afba05586d827
3
+ size 6722
config.json ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen3MoeForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "decoder_sparse_step": 1,
8
+ "dtype": "bfloat16",
9
+ "eos_token_id": 151645,
10
+ "head_dim": 128,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 2048,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 5472,
15
+ "max_position_embeddings": 524288,
16
+ "max_window_layers": 28,
17
+ "mlp_only_layers": [],
18
+ "model_type": "qwen3_moe",
19
+ "moe_intermediate_size": 768,
20
+ "norm_topk_prob": true,
21
+ "num_attention_heads": 32,
22
+ "num_experts": 128,
23
+ "num_experts_per_tok": 8,
24
+ "num_hidden_layers": 67,
25
+ "num_key_value_heads": 4,
26
+ "output_router_logits": false,
27
+ "qkv_bias": false,
28
+ "quantization_config": {
29
+ "config_groups": {
30
+ "group_0": {
31
+ "format": "pack-quantized",
32
+ "input_activations": null,
33
+ "output_activations": null,
34
+ "targets": [
35
+ "Linear"
36
+ ],
37
+ "weights": {
38
+ "actorder": "static",
39
+ "block_structure": null,
40
+ "dynamic": false,
41
+ "group_size": 128,
42
+ "num_bits": 4,
43
+ "observer": "minmax",
44
+ "observer_kwargs": {},
45
+ "strategy": "group",
46
+ "symmetric": true,
47
+ "type": "int"
48
+ }
49
+ }
50
+ },
51
+ "format": "pack-quantized",
52
+ "global_compression_ratio": null,
53
+ "ignore": [
54
+ "model.layers.0.mlp.gate",
55
+ "model.layers.1.mlp.gate",
56
+ "model.layers.2.mlp.gate",
57
+ "model.layers.3.mlp.gate",
58
+ "model.layers.4.mlp.gate",
59
+ "model.layers.5.mlp.gate",
60
+ "model.layers.6.mlp.gate",
61
+ "model.layers.7.mlp.gate",
62
+ "model.layers.8.mlp.gate",
63
+ "model.layers.9.mlp.gate",
64
+ "model.layers.10.mlp.gate",
65
+ "model.layers.11.mlp.gate",
66
+ "model.layers.12.mlp.gate",
67
+ "model.layers.13.mlp.gate",
68
+ "model.layers.14.mlp.gate",
69
+ "model.layers.15.mlp.gate",
70
+ "model.layers.16.mlp.gate",
71
+ "model.layers.17.mlp.gate",
72
+ "model.layers.18.mlp.gate",
73
+ "model.layers.19.mlp.gate",
74
+ "model.layers.20.mlp.gate",
75
+ "model.layers.21.mlp.gate",
76
+ "model.layers.22.mlp.gate",
77
+ "model.layers.23.mlp.gate",
78
+ "model.layers.24.mlp.gate",
79
+ "model.layers.25.mlp.gate",
80
+ "model.layers.26.mlp.gate",
81
+ "model.layers.27.mlp.gate",
82
+ "model.layers.28.mlp.gate",
83
+ "model.layers.29.mlp.gate",
84
+ "model.layers.30.mlp.gate",
85
+ "model.layers.31.mlp.gate",
86
+ "model.layers.32.mlp.gate",
87
+ "model.layers.33.mlp.gate",
88
+ "model.layers.34.mlp.gate",
89
+ "model.layers.35.mlp.gate",
90
+ "model.layers.36.mlp.gate",
91
+ "model.layers.37.mlp.gate",
92
+ "model.layers.38.mlp.gate",
93
+ "model.layers.39.mlp.gate",
94
+ "model.layers.40.mlp.gate",
95
+ "model.layers.41.mlp.gate",
96
+ "model.layers.42.mlp.gate",
97
+ "model.layers.43.mlp.gate",
98
+ "model.layers.44.mlp.gate",
99
+ "model.layers.45.mlp.gate",
100
+ "model.layers.46.mlp.gate",
101
+ "model.layers.47.mlp.gate",
102
+ "model.layers.48.mlp.gate",
103
+ "model.layers.49.mlp.gate",
104
+ "model.layers.50.mlp.gate",
105
+ "model.layers.51.mlp.gate",
106
+ "model.layers.52.mlp.gate",
107
+ "model.layers.53.mlp.gate",
108
+ "model.layers.54.mlp.gate",
109
+ "model.layers.55.mlp.gate",
110
+ "model.layers.56.mlp.gate",
111
+ "model.layers.57.mlp.gate",
112
+ "model.layers.58.mlp.gate",
113
+ "model.layers.59.mlp.gate",
114
+ "model.layers.60.mlp.gate",
115
+ "model.layers.61.mlp.gate",
116
+ "model.layers.62.mlp.gate",
117
+ "model.layers.63.mlp.gate",
118
+ "model.layers.64.mlp.gate",
119
+ "model.layers.65.mlp.gate",
120
+ "model.layers.66.mlp.gate",
121
+ "lm_head"
122
+ ],
123
+ "kv_cache_scheme": null,
124
+ "quant_method": "compressed-tensors",
125
+ "quantization_status": "compressed",
126
+ "sparsity_config": {},
127
+ "transform_config": {},
128
+ "version": "0.12.2"
129
+ },
130
+ "rms_norm_eps": 1e-06,
131
+ "rope_scaling": {
132
+ "factor": 2.0,
133
+ "original_max_position_embeddings": 262144,
134
+ "rope_type": "yarn",
135
+ "type": "yarn"
136
+ },
137
+ "rope_theta": 10000000,
138
+ "router_aux_loss_coef": 0.0,
139
+ "shared_expert_intermediate_size": 0,
140
+ "sliding_window": null,
141
+ "tie_word_embeddings": false,
142
+ "transformers_version": "4.57.1",
143
+ "use_cache": true,
144
+ "use_qk_norm": true,
145
+ "use_sliding_window": false,
146
+ "vocab_size": 151936
147
+ }
generation_config.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "eos_token_id": 151645,
4
+ "transformers_version": "4.57.1"
5
+ }
merges.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8831e4f1a044471340f7c0a83d7bd71306a5b867e95fd870f74d0c5308a904d5
3
+ size 1671853
model-00001-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56e975a98c7e83050076b9d7f70793ed57f7441197d1b753635470a7611c1660
3
+ size 5001524144
model-00002-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0810b0ac49377cd8f893f1cb6214cd1e4ad2fdaf1b2923c60e88a194a5b9cd9
3
+ size 5001803304
model-00003-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1142c828c7e313b9f8102ab6398e3152f90426b48d48f68dd41bd9e1f9a096e1
3
+ size 5002084152
model-00004-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02375548d249bfcdd586106c44682fb5969344058084060b5ea4690931f268c1
3
+ size 5001803304
model-00005-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62bd670173edc9e9bb1471b37dad3daa3e5d735610d9912693604ad25e470d89
3
+ size 2800869152
model.safetensors.index.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1a90757ed987528ee18c7ddbf4c44d252fef7a7553d38c07ad2487423e1537f
3
+ size 7560146
recipe.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9da07e9dca21134a77d2ceb08e5d25701958fd61fd8d08b95603b5f705a6cfcc
3
+ size 883
special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0acdaba32b920d640afb36af4396c91974e074735636e4016d17a8ed9c03730
3
+ size 11422753
tokenizer_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05a2d456eb807c8368fe6b60101b1d1160d6889071a751ed08576b96e77e01cd
3
+ size 5405
vocab.json ADDED
The diff for this file is too large to render. See raw diff