Update README.md
Browse files
README.md
CHANGED
|
@@ -6,4 +6,16 @@ metrics:
|
|
| 6 |
- accuracy
|
| 7 |
base_model:
|
| 8 |
- mistralai/Mistral-7B-Instruct-v0.3
|
| 9 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
- accuracy
|
| 7 |
base_model:
|
| 8 |
- mistralai/Mistral-7B-Instruct-v0.3
|
| 9 |
+
---
|
| 10 |
+
|
| 11 |
+
Lightweight LoRA adapters trained on the EdinburghNLP/XSum dataset for abstractive news summarization.
|
| 12 |
+
|
| 13 |
+
The adapters fine-tune Mistral-7B-Instruct-v0.3 using ~50k training and 2k validation samples prepared via a summarization prompt.
|
| 14 |
+
|
| 15 |
+
Training used the following parameters:
|
| 16 |
+
r=8,
|
| 17 |
+
lora_alpha=16,
|
| 18 |
+
lora_dropout=0.1,
|
| 19 |
+
target_modules=["q_proj", "k_proj", "v_proj", "o_proj"]
|
| 20 |
+
|
| 21 |
+
Adapters can be merged into the base model for inference using PeftModel.merge_and_unload().
|