Phi-3.5 Mini Instruct - Medical LoRA (GGUF Quantized)

This repository contains quantized GGUF versions of a LoRA fine-tuned model based on the microsoft/Phi-3.5-mini-instruct base model.

The fine-tuning was performed using the LoRA adapter: segestic/phi3.5-mini-4k-qlora-medical-seg-vall_med

๐Ÿ”ง Description

The .gguf files in this repository are merged and quantized versions of the above base model + LoRA, suitable for use with llama.cpp or other GGUF-compatible inference engines.

Each file corresponds to a different quantization level.

๐Ÿง  Intended Use

These models are intended for offline inference in medical question-answering or document understanding tasks using the phi-3.5 base and a custom LoRA.

โš ๏ธ Disclaimer

This model is fine-tuned on a medical dataset, but it is not a certified clinical tool. Use in production or decision-making requires proper validation.

Downloads last month
21
GGUF
Model size
4B params
Architecture
phi3
Hardware compatibility
Log In to view the estimation

4-bit

5-bit

8-bit

32-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support