kxdw2580 commited on
Commit
198ff26
·
verified ·
1 Parent(s): 20af869

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +74 -3
README.md CHANGED
@@ -1,3 +1,74 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - GPTQ
5
+ license: apache-2.0
6
+ datasets:
7
+ - kxdw2580/catgirl-dataset
8
+ language:
9
+ - zh
10
+ base_model:
11
+ - kxdw2580/DeepSeek-R1-0528-Qwen3-8B-catgirl-v2.5
12
+ ---
13
+
14
+ This model is the **GPTQ-v2 8-bit quantized version** of [`kxdw2580/DeepSeek-R1-0528-Qwen3-8B-catgirl-v2.5`](https://huggingface.co/kxdw2580/DeepSeek-R1-0528-Qwen3-8B-catgirl-v2.5). The quantization process resulted in minimal loss, with an average of **0.64539...**, which has been validated through internal few-shot benchmark tests.
15
+
16
+ Below is the original model's README:
17
+
18
+ ---
19
+ # kxdw2580/DeepSeek-R1-0528-Qwen3-8B-catgirl-v2.5
20
+
21
+ This new model series integrates updated datasets, base architectures, and fine-tuning methodologies. Based on **Qwen3**, it includes models with parameter counts of **8B** and **1.7B**.
22
+
23
+ Key updates focus on **daily conversations**, **creative generation**, **basic mathematics**, and **code generation**. Leveraging Qwen3's architecture, the model also supports **reasoning mode switching**.
24
+
25
+ 🔍 **Fine-tuning records** are available on **SwanLab**:
26
+ 1. [First Fine-tuning](https://swanlab.cn/@shadow01a/qwen-catgirl/runs/pcxfkgosz2e0cb430jk0a/overview)
27
+ 2. [Second Fine-tuning](https://swanlab.cn/@shadow01a/qwen-catgirl/runs/iuou1xratkvbiv7jxw16k/overview)
28
+ 3. [Third Fine-tuning](https://swanlab.cn/@shadow01a/qwen-catgirl/runs/9i2l4mc5qevmnlx2h51m0/overview)
29
+
30
+ ---
31
+
32
+ ## Evaluation
33
+
34
+ Due to the model's unique characteristics, we employed **human evaluation** for daily conversations and **DeepSeek-R1 scoring** (with reference answers provided in advance) for other domains to ensure character consistency and response validity.
35
+
36
+ ### Key Improvements (vs. internal test models "0501" and "0531-test-all"):
37
+ - **Stronger detail-awareness** in casual dialogue
38
+ - **More coherent storytelling** in creative tasks
39
+ - **Deeper reasoning** during thinking mode
40
+ - **Better persona adherence** in long-form conversations without explicit prompts
41
+ - **Significant gains** in math/code domains (internal 20-question benchmark):
42
+
43
+ | Model | Math (Single Attempt) | Code (Single Attempt) |
44
+ |-------|-----------------------|-----------------------|
45
+ | Internal Test Model-0501 | 10% | 0% |
46
+ | DeepSeek-R1-0528-Qwen3-8B-Catgirl-0531-test-all | 30% | 20% |
47
+ | **DeepSeek-R1-0528-Qwen3-8B-Catgirl-v2.5** | **70%** | **60%** |
48
+
49
+ ---
50
+
51
+ ## Usage Guidelines
52
+
53
+ ### Recommended Parameters:
54
+ - `temperature`: 0.7 (reasoning mode) / 0.6 (standard mode)
55
+ - `top_p`: 0.95
56
+
57
+ ### Critical Notes:
58
+ - **Avoid** using model's reasoning chains as conversation context
59
+ - Inherits base model's tendency for lengthy reasoning in some cases – allow completion even if intermediate steps seem unusual
60
+
61
+ ### English Mode:
62
+ Add this system prompt for English responses:
63
+ ```
64
+ You are a catgirl. Please speak English.
65
+ ```
66
+
67
+ ---
68
+
69
+ ## Acknowledgments
70
+
71
+ Special thanks to:
72
+ - **LLaMA-Factory** (fine-tuning framework)
73
+ - **Qwen Team** (base model provider)
74
+ - **DeepSeek Team** (DeepSeek-R1 evaluation support)