add perplexity chart for v0.2 quants
Browse files- README.md +3 -1
- images/perplexity.png +3 -0
README.md
CHANGED
|
@@ -32,13 +32,15 @@ Also thanks to all the folks in the quanting and inferencing community on [Beave
|
|
| 32 |
## *UPDATED RECIPES*
|
| 33 |
Updated new better lower perplexity recipes and worlds smallest Kimi-K2-Instruct-smol-IQ1_KT at 219.375 GIB (1.835) BPW. Please ask any questions in [this discussion here](https://huggingface.co/ubergarm/Kimi-K2-Instruct-GGUF/discussions/4), thanks!
|
| 34 |
|
| 35 |
-
|
| 36 |
|
| 37 |
## Quant Collection
|
| 38 |
Compare with Perplexity of full size `Q8_0` 1016.623 GiB (8.504 BPW):
|
| 39 |
|
| 40 |
Final estimate: PPL = 2.9507 +/- 0.01468
|
| 41 |
|
|
|
|
|
|
|
| 42 |
### * v0.2 `IQ4_KS` 554.421 GiB (4.638 BPW)
|
| 43 |
Final estimate: PPL = 2.9584 +/- 0.01473
|
| 44 |
|
|
|
|
| 32 |
## *UPDATED RECIPES*
|
| 33 |
Updated new better lower perplexity recipes and worlds smallest Kimi-K2-Instruct-smol-IQ1_KT at 219.375 GIB (1.835) BPW. Please ask any questions in [this discussion here](https://huggingface.co/ubergarm/Kimi-K2-Instruct-GGUF/discussions/4), thanks!
|
| 34 |
|
| 35 |
+
Old versions still available as described in the dicsussion at tag/revision v0.1.
|
| 36 |
|
| 37 |
## Quant Collection
|
| 38 |
Compare with Perplexity of full size `Q8_0` 1016.623 GiB (8.504 BPW):
|
| 39 |
|
| 40 |
Final estimate: PPL = 2.9507 +/- 0.01468
|
| 41 |
|
| 42 |
+

|
| 43 |
+
|
| 44 |
### * v0.2 `IQ4_KS` 554.421 GiB (4.638 BPW)
|
| 45 |
Final estimate: PPL = 2.9584 +/- 0.01473
|
| 46 |
|
images/perplexity.png
ADDED
|
Git LFS Details
|