Added banner.
Browse files
README.md
CHANGED
|
@@ -10,6 +10,8 @@ pipeline_tag: text-generation
|
|
| 10 |
library_name: transformers
|
| 11 |
---
|
| 12 |
# GPT4chan 8B FP8
|
|
|
|
|
|
|
| 13 |
This model is [v2ray/GPT4chan-8B](https://huggingface.co/v2ray/GPT4chan-8B) quantized to fp8 using [vllm-project/llm-compressor](https://github.com/vllm-project/llm-compressor).
|
| 14 |
|
| 15 |
Trained using 8x H100 with global batch size 64, using 2e-4 learning rate, for 4000 steps, which is approximately 5 epochs.
|
|
|
|
| 10 |
library_name: transformers
|
| 11 |
---
|
| 12 |
# GPT4chan 8B FP8
|
| 13 |
+

|
| 14 |
+
|
| 15 |
This model is [v2ray/GPT4chan-8B](https://huggingface.co/v2ray/GPT4chan-8B) quantized to fp8 using [vllm-project/llm-compressor](https://github.com/vllm-project/llm-compressor).
|
| 16 |
|
| 17 |
Trained using 8x H100 with global batch size 64, using 2e-4 learning rate, for 4000 steps, which is approximately 5 epochs.
|