README.md
Browse files
README.md
CHANGED
|
@@ -59,7 +59,6 @@ This model serves as the **core component** of a full-stack **AI engineering and
|
|
| 59 |
| **Method** | QLoRA (Quantized LoRA Fine-Tuning) |
|
| 60 |
| **Language** | English only |
|
| 61 |
| **Precision** | 4-bit (NF4) |
|
| 62 |
-
| **LoRA Config** | r=16, alpha=32, Dropout=0.07 |
|
| 63 |
| **Optimizer** | Paged AdamW 8-bit |
|
| 64 |
| **Frameworks** | `transformers`, `peft`, `bitsandbytes`, `fastapi` |
|
| 65 |
|
|
@@ -94,8 +93,7 @@ This repository demonstrates how to integrate this adapter with:
|
|
| 94 |
|
| 95 |
The fine tuning consist of multiple stage experiment
|
| 96 |
|
| 97 |
-
Stage 1:
|
| 98 |
-
|
| 99 |
| Phase | Summary | Runtime |
|
| 100 |
|--------|----------|----------|
|
| 101 |
| **1A** | Initial fine-tune (canceled due to overfitting) | 11h 50m |
|
|
@@ -104,12 +102,13 @@ Stage 1:
|
|
| 104 |
| **1D / 1D-A / 1E** | Refinement attempts with packing & oversampling | ~3d total |
|
| 105 |
| **1F** | Final adapter re-train from **1B** (expanded persona dataset, balanced oversampling) | 1d 5h |
|
| 106 |
|
| 107 |
-
Stage 2:
|
|
|
|
| 108 |
After gathering all the insights from the initial experiments (1A-1F), fine-tuning was restarted completely from scratch. By applying all the lessons learned, this new training process achieved better and more balanced performance in just 1s 21h.
|
| 109 |
The adapter released in this repository is the result of this final, optimized training.
|
| 110 |
| Phase | Summary | Runtime |
|
| 111 |
|--------|----------|----------|
|
| 112 |
-
| **1** | Fine-tune again from scratch by applying all the insights from previous experiments. | 1d 21h |
|
| 113 |
|
| 114 |
π **W&B Log (Phase 1F):** [wandb.ai/VoidNova/.../runs/bpju3d09](https://wandb.ai/VoidNova/phi-2-2.7B_qlora_alpaca-51.8k_identity-model-232_squadv2-15k/runs/bpju3d09?nw=nwuseradhafajp)
|
| 115 |
π **W&B Log (Final):** [wandb.ai/VoidNova/.../runs/rx5fih5v](https://wandb.ai/VoidNova/phi-2_qlora_ZeroChat/runs/rx5fih5v?nw=nwuseradhafajp)
|
|
|
|
| 59 |
| **Method** | QLoRA (Quantized LoRA Fine-Tuning) |
|
| 60 |
| **Language** | English only |
|
| 61 |
| **Precision** | 4-bit (NF4) |
|
|
|
|
| 62 |
| **Optimizer** | Paged AdamW 8-bit |
|
| 63 |
| **Frameworks** | `transformers`, `peft`, `bitsandbytes`, `fastapi` |
|
| 64 |
|
|
|
|
| 93 |
|
| 94 |
The fine tuning consist of multiple stage experiment
|
| 95 |
|
| 96 |
+
#### Stage 1:
|
|
|
|
| 97 |
| Phase | Summary | Runtime |
|
| 98 |
|--------|----------|----------|
|
| 99 |
| **1A** | Initial fine-tune (canceled due to overfitting) | 11h 50m |
|
|
|
|
| 102 |
| **1D / 1D-A / 1E** | Refinement attempts with packing & oversampling | ~3d total |
|
| 103 |
| **1F** | Final adapter re-train from **1B** (expanded persona dataset, balanced oversampling) | 1d 5h |
|
| 104 |
|
| 105 |
+
#### Stage 2:
|
| 106 |
+
|
| 107 |
After gathering all the insights from the initial experiments (1A-1F), fine-tuning was restarted completely from scratch. By applying all the lessons learned, this new training process achieved better and more balanced performance in just 1s 21h.
|
| 108 |
The adapter released in this repository is the result of this final, optimized training.
|
| 109 |
| Phase | Summary | Runtime |
|
| 110 |
|--------|----------|----------|
|
| 111 |
+
| **1** | Fine-tune again from scratch(from base model) by applying all the insights from previous experiments. | 1d 21h |
|
| 112 |
|
| 113 |
π **W&B Log (Phase 1F):** [wandb.ai/VoidNova/.../runs/bpju3d09](https://wandb.ai/VoidNova/phi-2-2.7B_qlora_alpaca-51.8k_identity-model-232_squadv2-15k/runs/bpju3d09?nw=nwuseradhafajp)
|
| 114 |
π **W&B Log (Final):** [wandb.ai/VoidNova/.../runs/rx5fih5v](https://wandb.ai/VoidNova/phi-2_qlora_ZeroChat/runs/rx5fih5v?nw=nwuseradhafajp)
|