Swallow-8B-it-v05-gguf-q6_k-mixed-v1
- Quantization Type: Mixed Precision (
q5_K, q6_K, q8_0)
- Bits Per Weight (BPW):
7.13
Swallow-8B-it-v05-gguf-q6_k-mixed-v2
- Quantization Type: Mixed Precision (
q6_K, q8_0)
- Bits Per Weight (BPW):
7.50
Swallow-8B-it-v05-gguf-q8_0-mixed-v1
- Quantization Type: Mixed Precision (
bf16, q4_K, q5_K, q6_K, q8_0)
- Bits Per Weight (BPW):
8.01
Swallow-8B-it-v05-gguf-q8_0-mixed-v2
- Quantization Type: Mixed Precision (
bf16, q5_K, q6_K, q8_0)
- Bits Per Weight (BPW):
9.31
Swallow-8B-it-v05-gguf-q8_0-mixed-v3
- Quantization Type: Mixed Precision (
bf16, q6_K, q8_0)
- Bits Per Weight (BPW):
11.44
Swallow-8B-it-v05-gguf-q8_0-mixed-v4
- Quantization Type: Mixed Precision (
bf16, q8_0)
- Bits Per Weight (BPW):
13.38
6-bit
8-bit
Model tree for marcelone/Llama-3.1-Swallow-8B-Instruct-v0.5-gguf
Collection including
marcelone/Llama-3.1-Swallow-8B-Instruct-v0.5-gguf