lucyknada commited on
Commit
d820a96
·
verified ·
1 Parent(s): c5bb3f4

Upload ./README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +69 -0
README.md ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: transformers
4
+ language:
5
+ - en
6
+ tags:
7
+ - chat
8
+ - conversational
9
+ base_model:
10
+ - maldv/Qwentile2.5-32B-Instruct
11
+ - a-m-team/AM-Thinking-v1
12
+ - nvidia/OpenCodeReasoning-Nemotron-32B
13
+ - maldv/Loqwqtus2.5-32B-Instruct
14
+ - trashpanda-org/QwQ-32B-Snowdrop-v0
15
+ - ArliAI/QwQ-32B-ArliAI-RpR-v3
16
+ pipeline_tags:
17
+ - text-generation
18
+ - conversational
19
+ - chat
20
+
21
+ ---
22
+ ### exl3 quant
23
+ ---
24
+ ### check revisions for quants
25
+ ---
26
+
27
+
28
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65b19c1b098c85365af5a83e/nJ7JqRPXgIthnMl6YZ8zt.png)
29
+
30
+ [GGUF](https://huggingface.co/mradermacher/QwentileLambda2.5-32B-Instruct-GGUF) [iMat](https://huggingface.co/mradermacher/QwentileLambda2.5-32B-Instruct-i1-GGUF)
31
+
32
+ # Qwentile Λ 2.5 32B Instruct
33
+
34
+ Qwentile Λ 2.5 32B Instruct is a *normalized denoised fourier interpolation* of the following models:
35
+
36
+ ```yaml
37
+ output_base_model: "maldv/Qwentile2.5-32B-Instruct"
38
+ output_dtype: "bfloat16"
39
+ finetune_merge:
40
+ - { "model": "a-m-team/AM-Thinking-v1", "base": "Qwen/Qwen2.5-32B", "alpha": 0.9 }
41
+ - { "model": "nvidia/OpenCodeReasoning-Nemotron-32B", "base": "Qwen/Qwen2.5-32B", "alpha": 0.8, "is_input": true}
42
+ - { "model": "maldv/Loqwqtus2.5-32B-Instruct", "base": "Qwen/Qwen2.5-32B", "alpha": 0.9 }
43
+ - { "model": "trashpanda-org/QwQ-32B-Snowdrop-v0", "base": "Qwen/Qwen2.5-32B", "alpha": 0.9 }
44
+ - { "model": "ArliAI/QwQ-32B-ArliAI-RpR-v3", "base": "Qwen/Qwen2.5-32B", "alpha": 0.8 }
45
+ ```
46
+
47
+ In other words, all of these models get warped and interpolated in signal space, and then jammed back on top of the base model (which in this case was Qwentile2.5-32B-Instruct); but with the Nemotron OpenCodeReasoning input layer.
48
+
49
+ ### What is this?
50
+
51
+ The latest in my series of Qwen 2.5 merges. Some really good models have been released recently, so I folded them in with Qwentile as the base. It should exhibit superior thinking skills, and perhaps even some code ability. I was satisfied with QReasoner2.5-32B-Instruct for advanced reasoning, but I suspect this will be an improvement.
52
+
53
+ ### A <think> model?
54
+
55
+ No, oddly enough, given it's lineage I thought for sure it would be a thought model, but instead it blends thought with it's creative output almost seamlessly. The combination is pretty powerful in my initial tests.
56
+
57
+ ## Citation
58
+
59
+ If you find our work helpful, feel free to give us a cite.
60
+
61
+ ```
62
+ @misc{qwentile-labmda-2.5-32b-instruct,
63
+ title = {Qwentile Λ 2.5 32B Instruct},
64
+ url = {https://huggingface.co/maldv/QwentileLambda2.5-32B-Instruct},
65
+ author = {Praxis Maldevide},
66
+ month = {May},
67
+ year = {2025}
68
+ }
69
+ ```