Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,74 @@
|
|
| 1 |
-
---
|
| 2 |
-
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
library_name: transformers
|
| 3 |
+
tags:
|
| 4 |
+
- GPTQ
|
| 5 |
+
license: apache-2.0
|
| 6 |
+
datasets:
|
| 7 |
+
- kxdw2580/catgirl-dataset
|
| 8 |
+
language:
|
| 9 |
+
- zh
|
| 10 |
+
base_model:
|
| 11 |
+
- kxdw2580/DeepSeek-R1-0528-Qwen3-8B-catgirl-v2.5
|
| 12 |
+
---
|
| 13 |
+
|
| 14 |
+
This model is the **GPTQ-v2 8-bit quantized version** of [`kxdw2580/DeepSeek-R1-0528-Qwen3-8B-catgirl-v2.5`](https://huggingface.co/kxdw2580/DeepSeek-R1-0528-Qwen3-8B-catgirl-v2.5). The quantization process resulted in minimal loss, with an average of **0.64539...**, which has been validated through internal few-shot benchmark tests.
|
| 15 |
+
|
| 16 |
+
Below is the original model's README:
|
| 17 |
+
|
| 18 |
+
---
|
| 19 |
+
# kxdw2580/DeepSeek-R1-0528-Qwen3-8B-catgirl-v2.5
|
| 20 |
+
|
| 21 |
+
This new model series integrates updated datasets, base architectures, and fine-tuning methodologies. Based on **Qwen3**, it includes models with parameter counts of **8B** and **1.7B**.
|
| 22 |
+
|
| 23 |
+
Key updates focus on **daily conversations**, **creative generation**, **basic mathematics**, and **code generation**. Leveraging Qwen3's architecture, the model also supports **reasoning mode switching**.
|
| 24 |
+
|
| 25 |
+
🔍 **Fine-tuning records** are available on **SwanLab**:
|
| 26 |
+
1. [First Fine-tuning](https://swanlab.cn/@shadow01a/qwen-catgirl/runs/pcxfkgosz2e0cb430jk0a/overview)
|
| 27 |
+
2. [Second Fine-tuning](https://swanlab.cn/@shadow01a/qwen-catgirl/runs/iuou1xratkvbiv7jxw16k/overview)
|
| 28 |
+
3. [Third Fine-tuning](https://swanlab.cn/@shadow01a/qwen-catgirl/runs/9i2l4mc5qevmnlx2h51m0/overview)
|
| 29 |
+
|
| 30 |
+
---
|
| 31 |
+
|
| 32 |
+
## Evaluation
|
| 33 |
+
|
| 34 |
+
Due to the model's unique characteristics, we employed **human evaluation** for daily conversations and **DeepSeek-R1 scoring** (with reference answers provided in advance) for other domains to ensure character consistency and response validity.
|
| 35 |
+
|
| 36 |
+
### Key Improvements (vs. internal test models "0501" and "0531-test-all"):
|
| 37 |
+
- **Stronger detail-awareness** in casual dialogue
|
| 38 |
+
- **More coherent storytelling** in creative tasks
|
| 39 |
+
- **Deeper reasoning** during thinking mode
|
| 40 |
+
- **Better persona adherence** in long-form conversations without explicit prompts
|
| 41 |
+
- **Significant gains** in math/code domains (internal 20-question benchmark):
|
| 42 |
+
|
| 43 |
+
| Model | Math (Single Attempt) | Code (Single Attempt) |
|
| 44 |
+
|-------|-----------------------|-----------------------|
|
| 45 |
+
| Internal Test Model-0501 | 10% | 0% |
|
| 46 |
+
| DeepSeek-R1-0528-Qwen3-8B-Catgirl-0531-test-all | 30% | 20% |
|
| 47 |
+
| **DeepSeek-R1-0528-Qwen3-8B-Catgirl-v2.5** | **70%** | **60%** |
|
| 48 |
+
|
| 49 |
+
---
|
| 50 |
+
|
| 51 |
+
## Usage Guidelines
|
| 52 |
+
|
| 53 |
+
### Recommended Parameters:
|
| 54 |
+
- `temperature`: 0.7 (reasoning mode) / 0.6 (standard mode)
|
| 55 |
+
- `top_p`: 0.95
|
| 56 |
+
|
| 57 |
+
### Critical Notes:
|
| 58 |
+
- **Avoid** using model's reasoning chains as conversation context
|
| 59 |
+
- Inherits base model's tendency for lengthy reasoning in some cases – allow completion even if intermediate steps seem unusual
|
| 60 |
+
|
| 61 |
+
### English Mode:
|
| 62 |
+
Add this system prompt for English responses:
|
| 63 |
+
```
|
| 64 |
+
You are a catgirl. Please speak English.
|
| 65 |
+
```
|
| 66 |
+
|
| 67 |
+
---
|
| 68 |
+
|
| 69 |
+
## Acknowledgments
|
| 70 |
+
|
| 71 |
+
Special thanks to:
|
| 72 |
+
- **LLaMA-Factory** (fine-tuning framework)
|
| 73 |
+
- **Qwen Team** (base model provider)
|
| 74 |
+
- **DeepSeek Team** (DeepSeek-R1 evaluation support)
|