MedGemma 1.5 4B IT β GGUF Quantized
GGUF quantization of google/medgemma-1.5-4b-it for mobile/edge deployment.
File
| File | Quant | Size | Use case |
|---|---|---|---|
medgemma-1.5-4b-it-Q4_K_M.gguf |
Q4_K_M | ~2.5 GB | Mobile / edge devices |
Usage
With llama.cpp
llama-cli -m medgemma-1.5-4b-it-Q4_K_M.gguf -cnv
With Ollama
Create a Modelfile:
FROM ./medgemma-1.5-4b-it-Q4_K_M.gguf
Then:
ollama create medgemma -f Modelfile
ollama run medgemma
On Android (llama.cpp Android build / MLC LLM)
Use the Q4_K_M file (~2.5 GB fits in most modern phone RAM).
Intended Use β Sakhi
This model powers Sakhi, an AI clinical companion for ASHA workers in rural India, with 3 modes:
- CHECKUP β ANC readings β risk assessment
- NEWBORN β Home visit findings β risk assessment
- CHAT β Free-form clinical Q&A
Safety
- NOT a medical device or diagnostic tool
- Requires clinical oversight
- When in doubt, always refer to PHC
License
- Downloads last month
- 57
Hardware compatibility
Log In to add your hardware
4-bit
Model tree for docvm/medgemma-1.5-4b-it-GGUF
Base model
google/medgemma-1.5-4b-it