adhafajp commited on
Commit
5934c08
Β·
1 Parent(s): 2fc4a1a
Files changed (1) hide show
  1. README.md +4 -5
README.md CHANGED
@@ -59,7 +59,6 @@ This model serves as the **core component** of a full-stack **AI engineering and
59
  | **Method** | QLoRA (Quantized LoRA Fine-Tuning) |
60
  | **Language** | English only |
61
  | **Precision** | 4-bit (NF4) |
62
- | **LoRA Config** | r=16, alpha=32, Dropout=0.07 |
63
  | **Optimizer** | Paged AdamW 8-bit |
64
  | **Frameworks** | `transformers`, `peft`, `bitsandbytes`, `fastapi` |
65
 
@@ -94,8 +93,7 @@ This repository demonstrates how to integrate this adapter with:
94
 
95
  The fine tuning consist of multiple stage experiment
96
 
97
- Stage 1:
98
-
99
  | Phase | Summary | Runtime |
100
  |--------|----------|----------|
101
  | **1A** | Initial fine-tune (canceled due to overfitting) | 11h 50m |
@@ -104,12 +102,13 @@ Stage 1:
104
  | **1D / 1D-A / 1E** | Refinement attempts with packing & oversampling | ~3d total |
105
  | **1F** | Final adapter re-train from **1B** (expanded persona dataset, balanced oversampling) | 1d 5h |
106
 
107
- Stage 2:
 
108
  After gathering all the insights from the initial experiments (1A-1F), fine-tuning was restarted completely from scratch. By applying all the lessons learned, this new training process achieved better and more balanced performance in just 1s 21h.
109
  The adapter released in this repository is the result of this final, optimized training.
110
  | Phase | Summary | Runtime |
111
  |--------|----------|----------|
112
- | **1** | Fine-tune again from scratch by applying all the insights from previous experiments. | 1d 21h |
113
 
114
  πŸ“Š **W&B Log (Phase 1F):** [wandb.ai/VoidNova/.../runs/bpju3d09](https://wandb.ai/VoidNova/phi-2-2.7B_qlora_alpaca-51.8k_identity-model-232_squadv2-15k/runs/bpju3d09?nw=nwuseradhafajp)
115
  πŸ“Š **W&B Log (Final):** [wandb.ai/VoidNova/.../runs/rx5fih5v](https://wandb.ai/VoidNova/phi-2_qlora_ZeroChat/runs/rx5fih5v?nw=nwuseradhafajp)
 
59
  | **Method** | QLoRA (Quantized LoRA Fine-Tuning) |
60
  | **Language** | English only |
61
  | **Precision** | 4-bit (NF4) |
 
62
  | **Optimizer** | Paged AdamW 8-bit |
63
  | **Frameworks** | `transformers`, `peft`, `bitsandbytes`, `fastapi` |
64
 
 
93
 
94
  The fine tuning consist of multiple stage experiment
95
 
96
+ #### Stage 1:
 
97
  | Phase | Summary | Runtime |
98
  |--------|----------|----------|
99
  | **1A** | Initial fine-tune (canceled due to overfitting) | 11h 50m |
 
102
  | **1D / 1D-A / 1E** | Refinement attempts with packing & oversampling | ~3d total |
103
  | **1F** | Final adapter re-train from **1B** (expanded persona dataset, balanced oversampling) | 1d 5h |
104
 
105
+ #### Stage 2:
106
+
107
  After gathering all the insights from the initial experiments (1A-1F), fine-tuning was restarted completely from scratch. By applying all the lessons learned, this new training process achieved better and more balanced performance in just 1s 21h.
108
  The adapter released in this repository is the result of this final, optimized training.
109
  | Phase | Summary | Runtime |
110
  |--------|----------|----------|
111
+ | **1** | Fine-tune again from scratch(from base model) by applying all the insights from previous experiments. | 1d 21h |
112
 
113
  πŸ“Š **W&B Log (Phase 1F):** [wandb.ai/VoidNova/.../runs/bpju3d09](https://wandb.ai/VoidNova/phi-2-2.7B_qlora_alpaca-51.8k_identity-model-232_squadv2-15k/runs/bpju3d09?nw=nwuseradhafajp)
114
  πŸ“Š **W&B Log (Final):** [wandb.ai/VoidNova/.../runs/rx5fih5v](https://wandb.ai/VoidNova/phi-2_qlora_ZeroChat/runs/rx5fih5v?nw=nwuseradhafajp)