Update README.md
Browse files
README.md
CHANGED
|
@@ -98,11 +98,9 @@ Sample output:
|
|
| 98 |
## Training Procedure
|
| 99 |
|
| 100 |
- **Base model:** [`answerdotai/ModernBERT-large`](https://huggingface.co/answerdotai/ModernBERT-large).
|
| 101 |
-
- **Dataset configuration:** training and validation splits pulled from `juanmcristobal/ner-ioc-dataset3` with column mapping `tokens` → tokens, `tags` → labels.
|
| 102 |
- **Hardware:** single Nvidia L40S instance (8 vCPU / 62 GB RAM / 48 GB VRAM).
|
| 103 |
- **Optimisation setup:** mixed precision `fp16`, optimiser `adamw_torch`, cosine learning-rate scheduler, gradient accumulation `1`.
|
| 104 |
- **Key hyperparameters:** learning rate `5e-5`, batch size `128`, epochs `5`, maximum sequence length `128`.
|
| 105 |
-
- **Checkpoint:** best-performing checkpoint automatically pushed to the Hub as `juanmcristobal/autotrain-sec4`.
|
| 106 |
|
| 107 |
| Parameter | Value |
|
| 108 |
|-----------|-------|
|
|
|
|
| 98 |
## Training Procedure
|
| 99 |
|
| 100 |
- **Base model:** [`answerdotai/ModernBERT-large`](https://huggingface.co/answerdotai/ModernBERT-large).
|
|
|
|
| 101 |
- **Hardware:** single Nvidia L40S instance (8 vCPU / 62 GB RAM / 48 GB VRAM).
|
| 102 |
- **Optimisation setup:** mixed precision `fp16`, optimiser `adamw_torch`, cosine learning-rate scheduler, gradient accumulation `1`.
|
| 103 |
- **Key hyperparameters:** learning rate `5e-5`, batch size `128`, epochs `5`, maximum sequence length `128`.
|
|
|
|
| 104 |
|
| 105 |
| Parameter | Value |
|
| 106 |
|-----------|-------|
|