Lamapi commited on
Commit
9416209
·
verified ·
1 Parent(s): dbd5c40

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +97 -0
README.md ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - tr
4
+ - en
5
+ - de
6
+ - es
7
+ - fr
8
+ - ru
9
+ - zh
10
+ - ja
11
+ - ko
12
+ license: mit
13
+ tags:
14
+ - turkish
15
+ - türkiye
16
+ - reasoning
17
+ - ai
18
+ - lamapi
19
+ - gemma3
20
+ - next
21
+ - next-x1
22
+ - text-generation
23
+ - open-source
24
+ - 14b
25
+ - large-language-model
26
+ - llm
27
+ - transformer
28
+ - artificial-intelligence
29
+ - machine-learning
30
+ - nlp
31
+ - multilingual
32
+ - instruction-tuned
33
+ - chat
34
+ - generative-ai
35
+ - optimized
36
+ - trl
37
+ - sft
38
+ - cognitive
39
+ - analytical
40
+ - enterprise
41
+ - llama-cpp
42
+ - gguf-my-repo
43
+ pipeline_tag: text-generation
44
+ datasets:
45
+ - mlabonne/FineTome-100k
46
+ - CognitiveKernel/CognitiveKernel-Pro-SFT
47
+ - OpenSPG/KAG-Thinker-training-dataset
48
+ - Gryphe/ChatGPT-4o-Writing-Prompts
49
+ - QuixiAI/dolphin-r1
50
+ - uclanlp/Brief-Pro
51
+ library_name: transformers
52
+ base_model: Lamapi/next-14b
53
+ ---
54
+
55
+ # Lamapi/next-14b-Q5_K_M-GGUF
56
+ This model was converted to GGUF format from [`Lamapi/next-14b`](https://huggingface.co/Lamapi/next-14b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
57
+ Refer to the [original model card](https://huggingface.co/Lamapi/next-14b) for more details on the model.
58
+
59
+ ## Use with llama.cpp
60
+ Install llama.cpp through brew (works on Mac and Linux)
61
+
62
+ ```bash
63
+ brew install llama.cpp
64
+
65
+ ```
66
+ Invoke the llama.cpp server or the CLI.
67
+
68
+ ### CLI:
69
+ ```bash
70
+ llama-cli --hf-repo Lamapi/next-14b-Q5_K_M-GGUF --hf-file next-14b-q5_k_m.gguf -p "The meaning to life and the universe is"
71
+ ```
72
+
73
+ ### Server:
74
+ ```bash
75
+ llama-server --hf-repo Lamapi/next-14b-Q5_K_M-GGUF --hf-file next-14b-q5_k_m.gguf -c 2048
76
+ ```
77
+
78
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
79
+
80
+ Step 1: Clone llama.cpp from GitHub.
81
+ ```
82
+ git clone https://github.com/ggerganov/llama.cpp
83
+ ```
84
+
85
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
86
+ ```
87
+ cd llama.cpp && LLAMA_CURL=1 make
88
+ ```
89
+
90
+ Step 3: Run inference through the main binary.
91
+ ```
92
+ ./llama-cli --hf-repo Lamapi/next-14b-Q5_K_M-GGUF --hf-file next-14b-q5_k_m.gguf -p "The meaning to life and the universe is"
93
+ ```
94
+ or
95
+ ```
96
+ ./llama-server --hf-repo Lamapi/next-14b-Q5_K_M-GGUF --hf-file next-14b-q5_k_m.gguf -c 2048
97
+ ```