awni commited on
Commit
ba0babe
·
verified ·
1 Parent(s): d88402e

Add files using upload-large-folder tool

Browse files
Files changed (4) hide show
  1. README.md +25 -5
  2. config.json +4 -2
  3. model.safetensors +1 -1
  4. tokenizer_config.json +2 -1
README.md CHANGED
@@ -1,21 +1,41 @@
1
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  library_name: mlx
3
- tags:
4
- - falcon-h1
5
- - mlx
6
  license: other
7
  license_name: falcon-llm-license
8
  license_link: https://falconllm.tii.ae/falcon-terms-and-conditions.html
9
- base_model: tiiuae/Falcon-H1-1.5B-Instruct
 
 
10
  inference: true
11
  pipeline_tag: text-generation
 
12
  ---
13
 
14
  # mlx-community/Falcon-H1-1.5B-Instruct-6bit
15
 
16
  This model [mlx-community/Falcon-H1-1.5B-Instruct-6bit](https://huggingface.co/mlx-community/Falcon-H1-1.5B-Instruct-6bit) was
17
  converted to MLX format from [tiiuae/Falcon-H1-1.5B-Instruct](https://huggingface.co/tiiuae/Falcon-H1-1.5B-Instruct)
18
- using mlx-lm version **0.25.2**.
19
 
20
  ## Use with mlx
21
 
 
1
  ---
2
+ base_model: tiiuae/Falcon-H1-1.5B-Instruct
3
+ language:
4
+ - ar
5
+ - cs
6
+ - de
7
+ - en
8
+ - es
9
+ - fr
10
+ - hi
11
+ - it
12
+ - ja
13
+ - ko
14
+ - nl
15
+ - pl
16
+ - pt
17
+ - ro
18
+ - ru
19
+ - sv
20
+ - ur
21
+ - zh
22
  library_name: mlx
 
 
 
23
  license: other
24
  license_name: falcon-llm-license
25
  license_link: https://falconllm.tii.ae/falcon-terms-and-conditions.html
26
+ tags:
27
+ - falcon-h1
28
+ - mlx
29
  inference: true
30
  pipeline_tag: text-generation
31
+ paper: tiiuae/falcon-h1
32
  ---
33
 
34
  # mlx-community/Falcon-H1-1.5B-Instruct-6bit
35
 
36
  This model [mlx-community/Falcon-H1-1.5B-Instruct-6bit](https://huggingface.co/mlx-community/Falcon-H1-1.5B-Instruct-6bit) was
37
  converted to MLX format from [tiiuae/Falcon-H1-1.5B-Instruct](https://huggingface.co/tiiuae/Falcon-H1-1.5B-Instruct)
38
+ using mlx-lm version **0.28.0**.
39
 
40
  ## Use with mlx
41
 
config.json CHANGED
@@ -46,11 +46,13 @@
46
  "projectors_bias": false,
47
  "quantization": {
48
  "group_size": 64,
49
- "bits": 6
 
50
  },
51
  "quantization_config": {
52
  "group_size": 64,
53
- "bits": 6
 
54
  },
55
  "rms_norm_eps": 1e-05,
56
  "rope_scaling": null,
 
46
  "projectors_bias": false,
47
  "quantization": {
48
  "group_size": 64,
49
+ "bits": 6,
50
+ "mode": "affine"
51
  },
52
  "quantization_config": {
53
  "group_size": 64,
54
+ "bits": 6,
55
+ "mode": "affine"
56
  },
57
  "rms_norm_eps": 1e-05,
58
  "rope_scaling": null,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f5e5eace4ce3117bfb09700782f74244057686a00b8e9921a044fad1790134a9
3
  size 1264143237
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:954a3da6448f2338cab4c788e5a340978521ca4f54c4df23c510611a93c1e1f9
3
  size 1264143237
tokenizer_config.json CHANGED
@@ -4535,6 +4535,7 @@
4535
  ">>UNUSED_511<<"
4536
  ],
4537
  "bos_token": "<|begin_of_text|>",
 
4538
  "clean_up_tokenization_spaces": true,
4539
  "eos_token": "<|end_of_text|>",
4540
  "extra_special_tokens": {},
@@ -4544,5 +4545,5 @@
4544
  ],
4545
  "model_max_length": 1000000000000000019884624838656,
4546
  "pad_token": "<pad>",
4547
- "tokenizer_class": "PreTrainedTokenizerFast"
4548
  }
 
4535
  ">>UNUSED_511<<"
4536
  ],
4537
  "bos_token": "<|begin_of_text|>",
4538
+ "chat_template": "{{bos_token}}\n{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"You are a function calling AI model. You are provided with function signature within <tools> </tools> XML tags. You may call one or more functions to assist with the user query. Don't make assumptions about what values to plug into functions.\\n<tools>\\n\" }}\n {%- for tool in tools %}[{{- tool | tojson }}]{%- endfor %}\n {{- \"\\n</tools>\\nFor each function call, return a json object with function name and arguments within <tool_call> </tool_call> tags with the following schema:\\n<tool_call>\\n{'arguments': <args-dict>, 'name': <function-name>}\\n</tool_call>\\n\" }}\n{%- else %}\n {%- if messages[0].role == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0].content + '<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}{% for message in messages %}{%- if message.role != 'system' %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{%- endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
4539
  "clean_up_tokenization_spaces": true,
4540
  "eos_token": "<|end_of_text|>",
4541
  "extra_special_tokens": {},
 
4545
  ],
4546
  "model_max_length": 1000000000000000019884624838656,
4547
  "pad_token": "<pad>",
4548
+ "tokenizer_class": "PreTrainedTokenizer"
4549
  }