π§ medgemma-4b-it-GGUF
A GGUF-quantized version of the
medgemma-4b-it
model for efficient local inference on CPUs and low-VRAM GPUs. Fine-tuned for medical use cases with instruction-following capabilities.
π¦ Model Details
- Base Model:
medgemma-4b
- Instruction-Tuned Variant:
medgemma-4b-it
- Format: GGUF (
.gguf
) - Quantization Levels:
Q4_0
,Q5_1
,Q8_0
, etc. - Architecture: Based on GEMMA, optimized for medical domains.
π§ About the Model
medgemma-4b-it
is a specialized variant of Google's GEMMA model, fine-tuned on medical instructions and dialogues. It aims to provide accurate and safe responses to healthcare-related queries. This GGUF release makes it easy to run the model locally with llama.cpp, llm, koboldcpp, or text-generation-webui.
π How to Use
With llama.cpp
./main -m medgemma-4b-it.Q4_0.gguf -p "What are the symptoms of Type 2 Diabetes?"
With text-generation-webui
- Drop the
.gguf
file into your/models
directory. - Launch the UI and select
medgemma-4b-it
. - Start generating!
β οΈ Disclaimer
- This model is intended for research and educational purposes only.
- It should not be used as a substitute for professional medical advice, diagnosis, or treatment.
- Always consult a qualified healthcare provider with any questions regarding medical conditions.
π License
The original model is released under Googleβs GEMMA license. This GGUF version follows the same license terms.
β¨ Acknowledgments
- Google DeepMind for creating GEMMA and Med-GEMMA.
- ggerganov for
llama.cpp
. - Hugging Face and the open-source community π
π¬ Contact
If you have questions or improvements for this repo, feel free to open an issue or reach out!
- Downloads last month
- 150
1-bit
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
16-bit