Updating model files
Browse files
README.md
CHANGED
|
@@ -6,6 +6,17 @@ tags:
|
|
| 6 |
- llama
|
| 7 |
inference: false
|
| 8 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
|
| 10 |
# Wizard-Vicuna-13B-GPTQ
|
| 11 |
|
|
@@ -18,7 +29,7 @@ It is the result of quantising to 4bit using [GPTQ-for-LLaMa](https://github.com
|
|
| 18 |
* [4bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/wizard-vicuna-13B-GPTQ).
|
| 19 |
* [4bit and 5bit GGML models for CPU inference](https://huggingface.co/TheBloke/wizard-vicuna-13B-GGML).
|
| 20 |
* [float16 HF format model for GPU inference](https://huggingface.co/TheBloke/wizard-vicuna-13B-HF).
|
| 21 |
-
|
| 22 |
## How to easily download and use this model in text-generation-webui
|
| 23 |
|
| 24 |
Open the text-generation-webui UI as normal.
|
|
@@ -53,7 +64,18 @@ It was created without the `--act-order` parameter. It may have slightly lower i
|
|
| 53 |
```
|
| 54 |
CUDA_VISIBLE_DEVICES=0 python3 llama.py wizard-vicuna-13B-HF c4 --wbits 4 --true-sequential --groupsize 128 --save_safetensors wizard-vicuna-13B-GPTQ-4bit.compat.no-act-order.safetensors
|
| 55 |
```
|
| 56 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 57 |
# Original WizardVicuna-13B model card
|
| 58 |
|
| 59 |
Github page: https://github.com/melodysdreamj/WizardVicunaLM
|
|
@@ -68,7 +90,7 @@ I am a big fan of the ideas behind WizardLM and VicunaLM. I particularly like th
|
|
| 68 |

|
| 69 |
|
| 70 |
|
| 71 |
-
### Detail
|
| 72 |
|
| 73 |
The questions presented here are not from rigorous tests, but rather, I asked a few questions and requested GPT-4 to score them. The models compared were ChatGPT 3.5, WizardVicunaLM, VicunaLM, and WizardLM, in that order.
|
| 74 |
|
|
|
|
| 6 |
- llama
|
| 7 |
inference: false
|
| 8 |
---
|
| 9 |
+
<div style="width: 100%;">
|
| 10 |
+
<img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
|
| 11 |
+
</div>
|
| 12 |
+
<div style="display: flex; justify-content: space-between; width: 100%;">
|
| 13 |
+
<div style="display: flex; flex-direction: column; align-items: flex-start;">
|
| 14 |
+
<p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
|
| 15 |
+
</div>
|
| 16 |
+
<div style="display: flex; flex-direction: column; align-items: flex-end;">
|
| 17 |
+
<p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
|
| 18 |
+
</div>
|
| 19 |
+
</div>
|
| 20 |
|
| 21 |
# Wizard-Vicuna-13B-GPTQ
|
| 22 |
|
|
|
|
| 29 |
* [4bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/wizard-vicuna-13B-GPTQ).
|
| 30 |
* [4bit and 5bit GGML models for CPU inference](https://huggingface.co/TheBloke/wizard-vicuna-13B-GGML).
|
| 31 |
* [float16 HF format model for GPU inference](https://huggingface.co/TheBloke/wizard-vicuna-13B-HF).
|
| 32 |
+
|
| 33 |
## How to easily download and use this model in text-generation-webui
|
| 34 |
|
| 35 |
Open the text-generation-webui UI as normal.
|
|
|
|
| 64 |
```
|
| 65 |
CUDA_VISIBLE_DEVICES=0 python3 llama.py wizard-vicuna-13B-HF c4 --wbits 4 --true-sequential --groupsize 128 --save_safetensors wizard-vicuna-13B-GPTQ-4bit.compat.no-act-order.safetensors
|
| 66 |
```
|
| 67 |
+
|
| 68 |
+
## Want to support my work?
|
| 69 |
+
|
| 70 |
+
I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
|
| 71 |
+
|
| 72 |
+
So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
|
| 73 |
+
|
| 74 |
+
Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
|
| 75 |
+
|
| 76 |
+
* Patreon: coming soon! (just awaiting approval)
|
| 77 |
+
* Ko-Fi: https://ko-fi.com/TheBlokeAI
|
| 78 |
+
* Discord: https://discord.gg/UBgz4VXf
|
| 79 |
# Original WizardVicuna-13B model card
|
| 80 |
|
| 81 |
Github page: https://github.com/melodysdreamj/WizardVicunaLM
|
|
|
|
| 90 |

|
| 91 |
|
| 92 |
|
| 93 |
+
### Detail
|
| 94 |
|
| 95 |
The questions presented here are not from rigorous tests, but rather, I asked a few questions and requested GPT-4 to score them. The models compared were ChatGPT 3.5, WizardVicunaLM, VicunaLM, and WizardLM, in that order.
|
| 96 |
|