ZeroXClem commited on
Commit
bcacbe0
Β·
verified Β·
1 Parent(s): a91c052

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +199 -7
README.md CHANGED
@@ -3,18 +3,41 @@ license: apache-2.0
3
  tags:
4
  - merge
5
  - mergekit
6
- - lazymergekit
 
 
 
 
 
 
 
 
 
 
7
  ---
8
 
9
- # ZeroXClem-Qwen3-8B-HoneyBadger-EXP
10
 
11
- ZeroXClem-Qwen3-8B-HoneyBadger-EXP is a merge of the following models using [mergekit](https://github.com/cg123/mergekit):
12
 
13
- ## 🧩 Configuration
14
 
15
- ```yaml
16
- # Merge configuration for ZeroXClem-Qwen3-8B-HoneyBadger-EXP using MODEL STOCK
 
 
 
 
 
 
 
 
 
 
 
 
17
 
 
18
  name: ZeroXClem-Qwen3-8B-HoneyBadger-EXP
19
  base_model: AXCXEPT/Qwen3-EZO-8B-beta
20
  dtype: bfloat16
@@ -25,4 +48,173 @@ models:
25
  - model: YOYO-AI/Qwen3-8B-YOYO
26
  - model: KaraKaraWitch/CavesOfQwen3-8b
27
  tokenizer_source: AXCXEPT/Qwen3-EZO-8B-beta
28
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  tags:
4
  - merge
5
  - mergekit
6
+ - model-stock
7
+ language:
8
+ - en
9
+ base_model:
10
+ - AXCXEPT/Qwen3-EZO-8B-beta
11
+ - YOYO-AI/Qwen3-8B-YOYO
12
+ - KaraKaraWitch/CavesOfQwen3-8b
13
+ - GreenerPastures/Bald-Beaver-8B
14
+ - taki555/Qwen3-8B-Shadow-FT-BAAI-2k
15
+ pipeline_tag: text-generation
16
+ library_name: transformers
17
  ---
18
 
19
+ # 🦑 ZeroXClem-Qwen3-8B-HoneyBadger-EXP
20
 
21
+ ## 🧬 Overview
22
 
23
+ **ZeroXClem-Qwen3-8B-HoneyBadger-EXP** is a fierce and expressive model fusion crafted using the **Model Stock** merge method from [MergeKit](https://github.com/cg123/mergekit). Designed to combine **instruction-following, deep reasoning, creative roleplay**, and **code capabilities**, this blend leverages the best of Qwen3-8B-based fine-tunes from various communities across Hugging Face.
24
 
25
+ This HoneyBadger doesn't just careβ€”it dominates symbolic reasoning, narrative immersion, and technical comprehension with sleek aggression.
26
+
27
+ **Be advised** Use the ollama modelfile, or customized prompt with default Qwen3 chat template for **optimal inference**.
28
+
29
+ ---
30
+
31
+ ## πŸ”§ Merge Configuration
32
+
33
+ - **Merge Method:** `model_stock`
34
+ - **Base Model:** [`AXCXEPT/Qwen3-EZO-8B-beta`](https://huggingface.co/AXCXEPT/Qwen3-EZO-8B-beta)
35
+ - **Dtype:** `bfloat16`
36
+ - **Tokenizer Source:** `AXCXEPT/Qwen3-EZO-8B-beta`
37
+
38
+ ### 🧾 YAML
39
 
40
+ ```yaml
41
  name: ZeroXClem-Qwen3-8B-HoneyBadger-EXP
42
  base_model: AXCXEPT/Qwen3-EZO-8B-beta
43
  dtype: bfloat16
 
48
  - model: YOYO-AI/Qwen3-8B-YOYO
49
  - model: KaraKaraWitch/CavesOfQwen3-8b
50
  tokenizer_source: AXCXEPT/Qwen3-EZO-8B-beta
51
+ ````
52
+
53
+ ---
54
+
55
+ ## πŸ’‘ Models Merged
56
+
57
+ | Model | Highlights |
58
+ | ------------------------------------------------------------------------------------------------- | --------------------------------------------------------------------- |
59
+ | [`KaraKaraWitch/CavesOfQwen3-8b`](https://huggingface.co/KaraKaraWitch/CavesOfQwen3-8b) | Loosens baked-in instruct bias for more natural RP and abstract depth |
60
+ | [`YOYO-AI/Qwen3-8B-YOYO`](https://huggingface.co/YOYO-AI/Qwen3-8B-YOYO) | Della-style merge optimized for rich conversational alignment |
61
+ | [`AXCXEPT/Qwen3-EZO-8B-beta`](https://huggingface.co/AXCXEPT/Qwen3-EZO-8B-beta) | MT-Bench 9.08, deep-thought prompting, vLLM friendly |
62
+ | [`GreenerPastures/Bald-Beaver-8B`](https://huggingface.co/GreenerPastures/Bald-Beaver-8B) | Uncensored storytelling and immersive character dialogue |
63
+ | [`taki555/Qwen3-8B-Shadow-FT-BAAI-2k`](https://huggingface.co/taki555/Qwen3-8B-Shadow-FT-BAAI-2k) | Shadow-FT tuned for precise instruction-following on BAAI-2k |
64
+
65
+ ---
66
+
67
+ ## πŸ§ͺ Capabilities
68
+
69
+ 🧠 **Deep Symbolic Reasoning** – Via Shadow-FT and DeepScaleR techniques from base models
70
+ 🎭 **Immersive Roleplay & Storytelling** – Injected from Bald-Beaver and CavesOfQwen merges
71
+ πŸ’» **Code Understanding & Generation** – Python, C++, JS supported from Bootes & Shadow paths
72
+ 🧾 **Structured Outputs** – Supports Markdown, JSON, LaTeX, and more
73
+ 🧡 **ChatML Friendly** – Full compatibility with ChatML-format prompts
74
+
75
+ ---
76
+
77
+ ## πŸ› οΈ Usage Instructions
78
+ **For Optimal Inference** Use the following ollama modelfile, create it as a file caled Modelfile.
79
+ <details>
80
+ <summary> Ollama Modelfile </summary>
81
+
82
+ ```
83
+ FROM https://hf.co/ZeroXClem/Qwen3-8B-HoneyBadger-EXP-Q4_K_M-GGUF:latest
84
+ PARAMETER temperature 0.6
85
+ PARAMETER top_p 0.95
86
+ PARAMETER repeat_penalty 1.05
87
+ PARAMETER top_k 20
88
+ TEMPLATE """"{{- if .Messages }}
89
+ {{- if or .System .Tools }}<|im_start|>system
90
+ {{- if .System }}
91
+ {{ .System }}
92
+ {{- end }}
93
+ {{- if .Tools }}
94
+
95
+ # Tools
96
+
97
+ You may call one or more functions to assist with the user query.
98
+
99
+ You are provided with function signatures within <tools></tools> XML tags:
100
+ <tools>
101
+ {{- range .Tools }}
102
+ {"type": "function", "function": {{ .Function }}}
103
+ {{- end }}
104
+ </tools>
105
+
106
+ For each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:
107
+ <tool_call>
108
+ {"name": <function-name>, "arguments": <args-json-object>}
109
+ </tool_call>
110
+ {{- end }}<|im_end|>
111
+ {{ end }}
112
+ {{- range $i, $_ := .Messages }}
113
+ {{- $last := eq (len (slice $.Messages $i)) 1 -}}
114
+ {{- if eq .Role "user" }}<|im_start|>user
115
+ {{ .Content }}<|im_end|>
116
+ {{ else if eq .Role "assistant" }}<|im_start|>assistant
117
+ {{ if .Content }}{{ .Content }}
118
+ {{- else if .ToolCalls }}<tool_call>
119
+ {{ range .ToolCalls }}{"name": "{{ .Function.Name }}", "arguments": {{ .Function.Arguments }}}
120
+ {{ end }}</tool_call>
121
+ {{- end }}{{ if not $last }}<|im_end|>
122
+ {{ end }}
123
+ {{- else if eq .Role "tool" }}<|im_start|>user
124
+ <tool_response>
125
+ {{ .Content }}
126
+ </tool_response><|im_end|>
127
+ {{ end }}
128
+ {{- if and (ne .Role "assistant") $last }}<|im_start|>assistant
129
+ {{ end }}
130
+ {{- end }}
131
+ {{- else }}
132
+ {{- if .System }}<|im_start|>system
133
+ {{ .System }}<|im_end|>
134
+ {{ end }}{{ if .Prompt }}<|im_start|>user
135
+ {{ .Prompt }}<|im_end|>
136
+ {{ end }}<|im_start|>assistant
137
+ {{ end }}{{ .Response }}{{ if .Response }}<|im_end|>{{ end }}"""
138
+ SYSTEM """# System Prompt: Universal Coder and DevOps Expert
139
+
140
+ You are an advanced AI assistant specializing in coding and DevOps. Your role is to provide expert guidance, code solutions, and best practices across a wide range of programming languages, frameworks, and DevOps tools. Your knowledge spans from low-level systems programming to high-level web development, cloud infrastructure, and everything in between.
141
+
142
+ ## Key responsibilities:
143
+ 1. Code analysis and optimization
144
+ 2. Debugging and troubleshooting
145
+ 3. Architecture design and system planning
146
+ 4. Version Control best practices (Git)
147
+ 5. Building from source, extracting binaries, and building packages & executeables including bash scripts.
148
+ 6. Security and implementation and auditing
149
+ 7. Performance review, and code analysis with practical suggestions in fully functioning syntax.
150
+
151
+ Be VERY selective on choosing how to respond based on the user query. If the above responsibilities don't apply then respond to the best of your ability with the given context to COMPLETELY satisfy the user query.
152
+
153
+ ### Guidance
154
+ When assisting users:
155
+ - Provide clear, concise, and well-commented code examples
156
+ - Explain complrex concepts in simple terms
157
+ - Offer multiple solutions when applicable, highlighting pros and cons
158
+ - Prioritize security, efficiency, scalability, and maintainability in all suggestions
159
+ - Adapt your communication style for expert users.
160
+
161
+ ### Helpful
162
+ Be EXTREMELY helpful, insightful, and lucid."""
163
+ ```
164
+
165
+
166
+
167
+ ---
168
+
169
+ ### πŸ¦™ Ollama Quickstart
170
+ This command downloads the pre-quantized GGUF version of the model and runs it locally, making it easy to experiment without extensive configuration.
171
+
172
+ ```bash
173
+ ollama run hf.co/ZeroXClem/Qwen3-8B-HoneyBadger-EXP-Q4_K_M-GGUF
174
+ ```
175
+
176
+ ### 🐍 Python Code Snippet
177
+ ```python
178
+ from transformers import AutoModelForCausalLM, AutoTokenizer
179
+
180
+ model_name = "ZeroXClem/Qwen3-8B-HoneyBadger-EXP"
181
+
182
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
183
+ model = AutoModelForCausalLM.from_pretrained(
184
+ model_name,
185
+ torch_dtype="auto",
186
+ device_map="auto"
187
+ )
188
+
189
+ prompt = "Write a short story about a detective solving a paradox in time."
190
+
191
+ inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
192
+ outputs = model.generate(**inputs, max_new_tokens=300)
193
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
194
+ ```
195
+ ---
196
+
197
+ ## ⚠️ Disclaimer
198
+
199
+ 🚧 **Experimental Merge**: This model is an early-stage **experimental prototype** and is **not ready for production**. It may contain unaligned or unfiltered behaviors. Use it for **research**, **prompt testing**, or further **fine-tuning workflows**.
200
+
201
+ ---
202
+
203
+ ## πŸ’– Special Thanks
204
+
205
+ To the brilliant developers and open-source pioneers who made this possible:
206
+
207
+ * πŸ§™β€β™€οΈ [KaraKaraWitch](https://huggingface.co/KaraKaraWitch) for CavesOfQwen3
208
+ * 🧠 [YOYO-AI](https://huggingface.co/YOYO-AI) for Della-style merges
209
+ * πŸ¦… [AXCXEPT](https://huggingface.co/AXCXEPT) for the exceptional Qwen3-EZO base
210
+ * 🌲 [GreenerPastures](https://huggingface.co/GreenerPastures) for uncensored RP excellence
211
+ * 🧩 [taki555](https://huggingface.co/taki555) for integrating Shadow-FT's cutting-edge research
212
+
213
+ πŸ”— Powered by [MergeKit](https://github.com/cg123/mergekit)
214
+
215
+ ---
216
+
217
+ **ZeroXClem Team | 2025** πŸͺ
218
+ *β€œBlending minds, one layer at a time.”*
219
+
220
+