ML4SE23_G1_WizardCoder-SCoT-1B-V1.0
IN4334 ML4SE
Group1 WizardCoder
This model is the result of the fine-tunign of the WizardCoder-1B-V1.0 model using Structured Chain-of-Though (S-CoT) enhanced instructions. S-CoT is used to enhance a sample of about 1200 entries from the Evol-Instruct 80k dataset. The resulting dataset is then used for the training task. The current WizardCoder model and the new S-CoT fine-tuned one are compared on both versions of HumanEval and MBPP (S-CoT enhanced and not) on the pass@1 metric. The S-CoT enhancement of the evaluation datasets allows to study its effect when used just as a prompting technique, independently of the S-CoT fine-tuning of the model.
Fine-tuning Details
| Hyperparameter | WizardCoder-1B-V1.0 |
|---|---|
| Batch size | 16 |
| Learning rate | 2e-5 |
| Epochs | 3 |
| Max length | 2048 |
| Warmup step | 30 |
| LR scheduler | cosine |
| Dataset | ML4SE23_G1_EvolInstruct-SCoT-1k |
The hardware consisted on a GPU instance rented from DataCrunch with the following specifications:
| NVidia RTX A6000 48GB 1A6000.10V |
|---|
| 2 GPUs |
| 48GB VRAM per GPU |
| 60 GB RAM |
| 10 CPUs |
| 100GB SSD Storage |
| Ubuntu 20.04 |
| CUDA 11.6 |
Results
Results of pass@1(%) on HumanEval and MBPP compared to HumanEval-SCoT and MBPP-SCoT using WizardCoder-1B, WizardCoder-SCoT-1B and WizardCoder-15B.
| Dataset | WizardCoder-1B-V1.0 | WizardCoder-SCoT-1B-V1.0 | WizardCoder-15B-V1.0 |
|---|---|---|---|
| HumanEval | 23.78 | 17.68 | 57.3 |
| HumanEval-SCoT | 44.51 | 27.44 | 57.3 |
| MBPP | 23.4 | 19.4 | 51.8 |
| MBPP-SCoT | 40 | 28 | 45.6 |
- Downloads last month
- 2