The Gamunu Project
Collection
3 items
โข
Updated
These models was converted to GGUF format from manthilaffs/Gamunu-4B-Instruct-Alpha using llama.cpp.
Refer to the original model card for more details on the model.
Install llama.cpp through brew (works on Mac and Linux)
brew install llama.cpp
Invoke the llama.cpp server or the CLI.
llama-cli --hf-repo manthilaffs/Gamunu-4B-Instruct-Alpha-GGUF --hf-file gamunu-4b-instruct-alpha-BF16.gguf -p "Hello! how are you?"
llama-cli --hf-repo manthilaffs/Gamunu-4B-Instruct-Alpha-GGUF --hf-file gamunu-4b-instruct-alpha-q8_0.gguf -p "Hello! how are you?"
llama-server --hf-repo manthilaffs/Gamunu-4B-Instruct-Alpha-GGUF --hf-file gamunu-4b-instruct-alpha-BF16.gguf -c 2048
llama-server --hf-repo manthilaffs/Gamunu-4B-Instruct-Alpha-GGUF --hf-file gamunu-4b-instruct-alpha-q8_0.gguf -c 2048
8-bit
16-bit
Base model
google/gemma-3-4b-pt