Upload ./README.md with huggingface_hub
Browse files
README.md
ADDED
|
@@ -0,0 +1,69 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
library_name: transformers
|
| 4 |
+
language:
|
| 5 |
+
- en
|
| 6 |
+
tags:
|
| 7 |
+
- chat
|
| 8 |
+
- conversational
|
| 9 |
+
base_model:
|
| 10 |
+
- maldv/Qwentile2.5-32B-Instruct
|
| 11 |
+
- a-m-team/AM-Thinking-v1
|
| 12 |
+
- nvidia/OpenCodeReasoning-Nemotron-32B
|
| 13 |
+
- maldv/Loqwqtus2.5-32B-Instruct
|
| 14 |
+
- trashpanda-org/QwQ-32B-Snowdrop-v0
|
| 15 |
+
- ArliAI/QwQ-32B-ArliAI-RpR-v3
|
| 16 |
+
pipeline_tags:
|
| 17 |
+
- text-generation
|
| 18 |
+
- conversational
|
| 19 |
+
- chat
|
| 20 |
+
|
| 21 |
+
---
|
| 22 |
+
### exl3 quant
|
| 23 |
+
---
|
| 24 |
+
### check revisions for quants
|
| 25 |
+
---
|
| 26 |
+
|
| 27 |
+
|
| 28 |
+

|
| 29 |
+
|
| 30 |
+
[GGUF](https://huggingface.co/mradermacher/QwentileLambda2.5-32B-Instruct-GGUF) [iMat](https://huggingface.co/mradermacher/QwentileLambda2.5-32B-Instruct-i1-GGUF)
|
| 31 |
+
|
| 32 |
+
# Qwentile Λ 2.5 32B Instruct
|
| 33 |
+
|
| 34 |
+
Qwentile Λ 2.5 32B Instruct is a *normalized denoised fourier interpolation* of the following models:
|
| 35 |
+
|
| 36 |
+
```yaml
|
| 37 |
+
output_base_model: "maldv/Qwentile2.5-32B-Instruct"
|
| 38 |
+
output_dtype: "bfloat16"
|
| 39 |
+
finetune_merge:
|
| 40 |
+
- { "model": "a-m-team/AM-Thinking-v1", "base": "Qwen/Qwen2.5-32B", "alpha": 0.9 }
|
| 41 |
+
- { "model": "nvidia/OpenCodeReasoning-Nemotron-32B", "base": "Qwen/Qwen2.5-32B", "alpha": 0.8, "is_input": true}
|
| 42 |
+
- { "model": "maldv/Loqwqtus2.5-32B-Instruct", "base": "Qwen/Qwen2.5-32B", "alpha": 0.9 }
|
| 43 |
+
- { "model": "trashpanda-org/QwQ-32B-Snowdrop-v0", "base": "Qwen/Qwen2.5-32B", "alpha": 0.9 }
|
| 44 |
+
- { "model": "ArliAI/QwQ-32B-ArliAI-RpR-v3", "base": "Qwen/Qwen2.5-32B", "alpha": 0.8 }
|
| 45 |
+
```
|
| 46 |
+
|
| 47 |
+
In other words, all of these models get warped and interpolated in signal space, and then jammed back on top of the base model (which in this case was Qwentile2.5-32B-Instruct); but with the Nemotron OpenCodeReasoning input layer.
|
| 48 |
+
|
| 49 |
+
### What is this?
|
| 50 |
+
|
| 51 |
+
The latest in my series of Qwen 2.5 merges. Some really good models have been released recently, so I folded them in with Qwentile as the base. It should exhibit superior thinking skills, and perhaps even some code ability. I was satisfied with QReasoner2.5-32B-Instruct for advanced reasoning, but I suspect this will be an improvement.
|
| 52 |
+
|
| 53 |
+
### A <think> model?
|
| 54 |
+
|
| 55 |
+
No, oddly enough, given it's lineage I thought for sure it would be a thought model, but instead it blends thought with it's creative output almost seamlessly. The combination is pretty powerful in my initial tests.
|
| 56 |
+
|
| 57 |
+
## Citation
|
| 58 |
+
|
| 59 |
+
If you find our work helpful, feel free to give us a cite.
|
| 60 |
+
|
| 61 |
+
```
|
| 62 |
+
@misc{qwentile-labmda-2.5-32b-instruct,
|
| 63 |
+
title = {Qwentile Λ 2.5 32B Instruct},
|
| 64 |
+
url = {https://huggingface.co/maldv/QwentileLambda2.5-32B-Instruct},
|
| 65 |
+
author = {Praxis Maldevide},
|
| 66 |
+
month = {May},
|
| 67 |
+
year = {2025}
|
| 68 |
+
}
|
| 69 |
+
```
|