lbourdois commited on
Commit
a9e5ad4
·
verified ·
1 Parent(s): 7e40d00

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +92 -80
README.md CHANGED
@@ -1,80 +1,92 @@
1
- ---
2
- license: creativeml-openrail-m
3
- datasets:
4
- - prithivMLmods/Math-IIO-68K-Mini
5
- language:
6
- - en
7
- base_model:
8
- - Qwen/Qwen2.5-7B-Instruct
9
- pipeline_tag: text-generation
10
- library_name: transformers
11
- tags:
12
- - safetensors
13
- - qwen2.5
14
- - 7B
15
- - Instruct
16
- - Math
17
- - CoT
18
- - one-shot
19
- ---
20
- ![aaa.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/faLfR-doaWP_BLUkOQrbq.png)
21
-
22
- ### **Math IIO 7B Instruct**
23
-
24
- The **Math IIO 7B Instruct** is a fine-tuned language model based on the robust **Qwen2.5-7B-Instruct** architecture. This model has been specifically trained to excel in single-shot mathematical reasoning and instruction-based tasks, making it a reliable choice for educational, analytical, and problem-solving applications.
25
-
26
- ### **Key Features:**
27
-
28
- 1. **Math-Optimized Capabilities:**
29
- The model is designed to handle complex mathematical problems, step-by-step calculations, and reasoning tasks.
30
-
31
- 2. **Instruction-Tuned:**
32
- Fine-tuned for better adherence to structured queries and task-oriented prompts, enabling clear and concise outputs.
33
-
34
- 3. **Large Vocabulary:**
35
- Equipped with an extensive tokenizer configuration and custom tokens to ensure precise mathematical notation support.
36
-
37
- ### Single Shot Answers
38
-
39
- ![solution.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/4Zq6crBrbFLDqfKlDwBMU.png)
40
-
41
- ### Math-IIO File Structure
42
-
43
- | File Name [ Uploaded file ] | Size | Description | Upload Status |
44
- |------------------------------------|------------|-----------------------------------------------|----------------|
45
- | `.gitattributes` | 1.57 kB | Git attributes configuration file | Uploaded |
46
- | `README.md` | 263 Bytes | README file with minimal details | Updated |
47
- | `added_tokens.json` | 657 Bytes | Custom added tokens for tokenizer | Uploaded |
48
- | `config.json` | 861 Bytes | Model configuration file | Uploaded |
49
- | `generation_config.json` | 281 Bytes | Configuration for text generation settings | Uploaded |
50
- | `merges.txt` | 1.82 MB | Merge rules for byte pair encoding tokenizer | Uploaded |
51
- | `pytorch_model-00001-of-00004.bin` | 4.88 GB | First part of model weights (PyTorch) | Uploaded (LFS) |
52
- | `pytorch_model-00002-of-00004.bin` | 4.93 GB | Second part of model weights (PyTorch) | Uploaded (LFS) |
53
- | `pytorch_model-00003-of-00004.bin` | 4.33 GB | Third part of model weights (PyTorch) | Uploaded (LFS) |
54
- | `pytorch_model-00004-of-00004.bin` | 1.09 GB | Fourth part of model weights (PyTorch) | Uploaded (LFS) |
55
- | `pytorch_model.bin.index.json` | 28.1 kB | Index JSON file for model weights | Uploaded |
56
- | `special_tokens_map.json` | 644 Bytes | Map of special tokens used by the tokenizer | Uploaded |
57
- | `tokenizer.json` | 11.4 MB | Tokenizer settings and vocab | Uploaded (LFS) |
58
- | `tokenizer_config.json` | 7.73 kB | Configuration for tokenizer | Uploaded |
59
- | `vocab.json` | 2.78 MB | Vocabulary for tokenizer | Uploaded |
60
-
61
- | Model Type | Size | Context Length | Link |
62
- |------------|------|----------------|------|
63
- | GGUF | 7B | - | [🤗 Math-IIO-7B-Instruct-GGUF](https://huggingface.co/prithivMLmods/Math-IIO-7B-Instruct-GGUF) |
64
-
65
- ### **Training Details:**
66
- - **Base Model:** [Qwen/Qwen2.5-7B-Instruct](#)
67
- - **Dataset:** Trained on **Math-IIO-68K-Mini**, a curated dataset with 68.8k high-quality examples focusing on mathematical instructions, equations, and logic-based queries.
68
-
69
- ### **Capabilities:**
70
- - **Problem-Solving:** Solves mathematical problems ranging from basic arithmetic to advanced calculus and linear algebra.
71
- - **Educational Use:** Explains solutions step-by-step, making it a valuable teaching assistant.
72
- - **Analysis & Reasoning:** Handles logical reasoning tasks and computational queries effectively.
73
-
74
- ### **How to Use:**
75
- 1. Download all model files, ensuring the PyTorch weights and tokenizer configurations are included.
76
- 2. Load the model in your Python environment using frameworks like PyTorch or Hugging Face Transformers.
77
- 3. Use the provided configurations (`config.json` and `generation_config.json`) for optimal inference.
78
-
79
- ---
80
-
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: creativeml-openrail-m
3
+ datasets:
4
+ - prithivMLmods/Math-IIO-68K-Mini
5
+ language:
6
+ - zho
7
+ - eng
8
+ - fra
9
+ - spa
10
+ - por
11
+ - deu
12
+ - ita
13
+ - rus
14
+ - jpn
15
+ - kor
16
+ - vie
17
+ - tha
18
+ - ara
19
+ base_model:
20
+ - Qwen/Qwen2.5-7B-Instruct
21
+ pipeline_tag: text-generation
22
+ library_name: transformers
23
+ tags:
24
+ - safetensors
25
+ - qwen2.5
26
+ - 7B
27
+ - Instruct
28
+ - Math
29
+ - CoT
30
+ - one-shot
31
+ ---
32
+ ![aaa.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/faLfR-doaWP_BLUkOQrbq.png)
33
+
34
+ ### **Math IIO 7B Instruct**
35
+
36
+ The **Math IIO 7B Instruct** is a fine-tuned language model based on the robust **Qwen2.5-7B-Instruct** architecture. This model has been specifically trained to excel in single-shot mathematical reasoning and instruction-based tasks, making it a reliable choice for educational, analytical, and problem-solving applications.
37
+
38
+ ### **Key Features:**
39
+
40
+ 1. **Math-Optimized Capabilities:**
41
+ The model is designed to handle complex mathematical problems, step-by-step calculations, and reasoning tasks.
42
+
43
+ 2. **Instruction-Tuned:**
44
+ Fine-tuned for better adherence to structured queries and task-oriented prompts, enabling clear and concise outputs.
45
+
46
+ 3. **Large Vocabulary:**
47
+ Equipped with an extensive tokenizer configuration and custom tokens to ensure precise mathematical notation support.
48
+
49
+ ### Single Shot Answers
50
+
51
+ ![solution.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/4Zq6crBrbFLDqfKlDwBMU.png)
52
+
53
+ ### Math-IIO File Structure
54
+
55
+ | File Name [ Uploaded file ] | Size | Description | Upload Status |
56
+ |------------------------------------|------------|-----------------------------------------------|----------------|
57
+ | `.gitattributes` | 1.57 kB | Git attributes configuration file | Uploaded |
58
+ | `README.md` | 263 Bytes | README file with minimal details | Updated |
59
+ | `added_tokens.json` | 657 Bytes | Custom added tokens for tokenizer | Uploaded |
60
+ | `config.json` | 861 Bytes | Model configuration file | Uploaded |
61
+ | `generation_config.json` | 281 Bytes | Configuration for text generation settings | Uploaded |
62
+ | `merges.txt` | 1.82 MB | Merge rules for byte pair encoding tokenizer | Uploaded |
63
+ | `pytorch_model-00001-of-00004.bin` | 4.88 GB | First part of model weights (PyTorch) | Uploaded (LFS) |
64
+ | `pytorch_model-00002-of-00004.bin` | 4.93 GB | Second part of model weights (PyTorch) | Uploaded (LFS) |
65
+ | `pytorch_model-00003-of-00004.bin` | 4.33 GB | Third part of model weights (PyTorch) | Uploaded (LFS) |
66
+ | `pytorch_model-00004-of-00004.bin` | 1.09 GB | Fourth part of model weights (PyTorch) | Uploaded (LFS) |
67
+ | `pytorch_model.bin.index.json` | 28.1 kB | Index JSON file for model weights | Uploaded |
68
+ | `special_tokens_map.json` | 644 Bytes | Map of special tokens used by the tokenizer | Uploaded |
69
+ | `tokenizer.json` | 11.4 MB | Tokenizer settings and vocab | Uploaded (LFS) |
70
+ | `tokenizer_config.json` | 7.73 kB | Configuration for tokenizer | Uploaded |
71
+ | `vocab.json` | 2.78 MB | Vocabulary for tokenizer | Uploaded |
72
+
73
+ | Model Type | Size | Context Length | Link |
74
+ |------------|------|----------------|------|
75
+ | GGUF | 7B | - | [🤗 Math-IIO-7B-Instruct-GGUF](https://huggingface.co/prithivMLmods/Math-IIO-7B-Instruct-GGUF) |
76
+
77
+ ### **Training Details:**
78
+ - **Base Model:** [Qwen/Qwen2.5-7B-Instruct](#)
79
+ - **Dataset:** Trained on **Math-IIO-68K-Mini**, a curated dataset with 68.8k high-quality examples focusing on mathematical instructions, equations, and logic-based queries.
80
+
81
+ ### **Capabilities:**
82
+ - **Problem-Solving:** Solves mathematical problems ranging from basic arithmetic to advanced calculus and linear algebra.
83
+ - **Educational Use:** Explains solutions step-by-step, making it a valuable teaching assistant.
84
+ - **Analysis & Reasoning:** Handles logical reasoning tasks and computational queries effectively.
85
+
86
+ ### **How to Use:**
87
+ 1. Download all model files, ensuring the PyTorch weights and tokenizer configurations are included.
88
+ 2. Load the model in your Python environment using frameworks like PyTorch or Hugging Face Transformers.
89
+ 3. Use the provided configurations (`config.json` and `generation_config.json`) for optimal inference.
90
+
91
+ ---
92
+