Junheelee070712 commited on
Commit
7588817
ยท
verified ยท
1 Parent(s): 47069c0

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +82 -0
README.md ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ license_link: https://huggingface.co/skt/A.X-4.0-Light/blob/main/LICENSE
4
+ language:
5
+ - en
6
+ - ko
7
+ pipeline_tag: text-generation
8
+ library_name: transformers
9
+ model_id: skt/A.X-4.0-Light
10
+ developers: SKT AI Model Lab
11
+ tags:
12
+ - llama-cpp
13
+ - gguf-my-repo
14
+ base_model: skt/A.X-4.0-Light
15
+ model-index:
16
+ - name: A.X-4.0-Light
17
+ results:
18
+ - task:
19
+ type: generate_until
20
+ name: mmlu
21
+ dataset:
22
+ name: mmlu (chat CoT)
23
+ type: hails/mmlu_no_train
24
+ metrics:
25
+ - type: exact_match
26
+ value: 75.43
27
+ name: exact_match
28
+ - task:
29
+ type: generate_until
30
+ name: kmmlu
31
+ dataset:
32
+ name: kmmlu (chat CoT)
33
+ type: HAERAE-HUB/KMMLU
34
+ metrics:
35
+ - type: exact_match
36
+ value: 64.15
37
+ name: exact_match
38
+ ---
39
+
40
+ # Junheelee070712/A.X-4.0-Light-Q4_K_M-GGUF
41
+ This model was converted to GGUF format from [`skt/A.X-4.0-Light`](https://huggingface.co/skt/A.X-4.0-Light) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
42
+ Refer to the [original model card](https://huggingface.co/skt/A.X-4.0-Light) for more details on the model.
43
+
44
+ ## Use with llama.cpp
45
+ Install llama.cpp through brew (works on Mac and Linux)
46
+
47
+ ```bash
48
+ brew install llama.cpp
49
+
50
+ ```
51
+ Invoke the llama.cpp server or the CLI.
52
+
53
+ ### CLI:
54
+ ```bash
55
+ llama-cli --hf-repo Junheelee070712/A.X-4.0-Light-Q4_K_M-GGUF --hf-file a.x-4.0-light-q4_k_m.gguf -p "The meaning to life and the universe is"
56
+ ```
57
+
58
+ ### Server:
59
+ ```bash
60
+ llama-server --hf-repo Junheelee070712/A.X-4.0-Light-Q4_K_M-GGUF --hf-file a.x-4.0-light-q4_k_m.gguf -c 2048
61
+ ```
62
+
63
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
64
+
65
+ Step 1: Clone llama.cpp from GitHub.
66
+ ```
67
+ git clone https://github.com/ggerganov/llama.cpp
68
+ ```
69
+
70
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
71
+ ```
72
+ cd llama.cpp && LLAMA_CURL=1 make
73
+ ```
74
+
75
+ Step 3: Run inference through the main binary.
76
+ ```
77
+ ./llama-cli --hf-repo Junheelee070712/A.X-4.0-Light-Q4_K_M-GGUF --hf-file a.x-4.0-light-q4_k_m.gguf -p "The meaning to life and the universe is"
78
+ ```
79
+ or
80
+ ```
81
+ ./llama-server --hf-repo Junheelee070712/A.X-4.0-Light-Q4_K_M-GGUF --hf-file a.x-4.0-light-q4_k_m.gguf -c 2048
82
+ ```