MedGemma 1.5 4B IT β€” GGUF Quantized

GGUF quantization of google/medgemma-1.5-4b-it for mobile/edge deployment.

File

File Quant Size Use case
medgemma-1.5-4b-it-Q4_K_M.gguf Q4_K_M ~2.5 GB Mobile / edge devices

Usage

With llama.cpp

llama-cli -m medgemma-1.5-4b-it-Q4_K_M.gguf -cnv

With Ollama

Create a Modelfile:

FROM ./medgemma-1.5-4b-it-Q4_K_M.gguf

Then:

ollama create medgemma -f Modelfile
ollama run medgemma

On Android (llama.cpp Android build / MLC LLM)

Use the Q4_K_M file (~2.5 GB fits in most modern phone RAM).

Intended Use β€” Sakhi

This model powers Sakhi, an AI clinical companion for ASHA workers in rural India, with 3 modes:

  1. CHECKUP β€” ANC readings β†’ risk assessment
  2. NEWBORN β€” Home visit findings β†’ risk assessment
  3. CHAT β€” Free-form clinical Q&A

Safety

  • NOT a medical device or diagnostic tool
  • Requires clinical oversight
  • When in doubt, always refer to PHC

License

Health AI Developer Foundations Terms of Use

Downloads last month
57
GGUF
Model size
4B params
Architecture
gemma3
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for docvm/medgemma-1.5-4b-it-GGUF

Quantized
(27)
this model

Space using docvm/medgemma-1.5-4b-it-GGUF 1