Update README.md
Browse files
README.md
CHANGED
|
@@ -118,6 +118,40 @@ model-index:
|
|
| 118 |
This model was converted to GGUF format from [`anthracite-org/magnum-v4-22b`](https://huggingface.co/anthracite-org/magnum-v4-22b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
| 119 |
Refer to the [original model card](https://huggingface.co/anthracite-org/magnum-v4-22b) for more details on the model.
|
| 120 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 121 |
## Use with llama.cpp
|
| 122 |
Install llama.cpp through brew (works on Mac and Linux)
|
| 123 |
|
|
|
|
| 118 |
This model was converted to GGUF format from [`anthracite-org/magnum-v4-22b`](https://huggingface.co/anthracite-org/magnum-v4-22b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
| 119 |
Refer to the [original model card](https://huggingface.co/anthracite-org/magnum-v4-22b) for more details on the model.
|
| 120 |
|
| 121 |
+
---
|
| 122 |
+
Model details:
|
| 123 |
+
-
|
| 124 |
+
This is a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus.
|
| 125 |
+
|
| 126 |
+
This model is fine-tuned on top of Mistral-Small-Instruct-2409.
|
| 127 |
+
|
| 128 |
+
Prompting
|
| 129 |
+
-
|
| 130 |
+
A typical input would look like this:
|
| 131 |
+
|
| 132 |
+
<s>[INST] SYSTEM MESSAGE
|
| 133 |
+
USER MESSAGE[/INST] ASSISTANT MESSAGE</s>[INST] USER MESSAGE[/INST]
|
| 134 |
+
|
| 135 |
+
Credits
|
| 136 |
+
-
|
| 137 |
+
We'd like to thank Recursal / Featherless for sponsoring the compute for this train, Featherless has been hosting our Magnum models since the first 72 B and has given thousands of people access to our models and helped us grow.
|
| 138 |
+
|
| 139 |
+
We would also like to thank all members of Anthracite who made this finetune possible.
|
| 140 |
+
|
| 141 |
+
Datasets
|
| 142 |
+
-
|
| 143 |
+
anthracite-org/c2_logs_32k_mistral-v3_v1.2_no_system
|
| 144 |
+
anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system
|
| 145 |
+
anthracite-org/kalo-opus-instruct-3k-filtered-no-system
|
| 146 |
+
anthracite-org/nopm_claude_writing_fixed
|
| 147 |
+
anthracite-org/kalo_opus_misc_240827_no_system
|
| 148 |
+
anthracite-org/kalo_misc_part2_no_system
|
| 149 |
+
|
| 150 |
+
Training
|
| 151 |
+
-
|
| 152 |
+
The training was done for 2 epochs. We used 8xH100s GPUs graciously provided by Recursal AI / Featherless AI for the full-parameter fine-tuning of the model.
|
| 153 |
+
|
| 154 |
+
---
|
| 155 |
## Use with llama.cpp
|
| 156 |
Install llama.cpp through brew (works on Mac and Linux)
|
| 157 |
|