🧠 medgemma-4b-it-GGUF

A GGUF-quantized version of the medgemma-4b-it model for efficient local inference on CPUs and low-VRAM GPUs. Fine-tuned for medical use cases with instruction-following capabilities.


πŸ“¦ Model Details

  • Base Model: medgemma-4b
  • Instruction-Tuned Variant: medgemma-4b-it
  • Format: GGUF (.gguf)
  • Quantization Levels: Q4_0, Q5_1, Q8_0, etc.
  • Architecture: Based on GEMMA, optimized for medical domains.


🧠 About the Model

medgemma-4b-it is a specialized variant of Google's GEMMA model, fine-tuned on medical instructions and dialogues. It aims to provide accurate and safe responses to healthcare-related queries. This GGUF release makes it easy to run the model locally with llama.cpp, llm, koboldcpp, or text-generation-webui.


πŸš€ How to Use

With llama.cpp

./main -m medgemma-4b-it.Q4_0.gguf -p "What are the symptoms of Type 2 Diabetes?"

With text-generation-webui

  1. Drop the .gguf file into your /models directory.
  2. Launch the UI and select medgemma-4b-it.
  3. Start generating!

⚠️ Disclaimer

  • This model is intended for research and educational purposes only.
  • It should not be used as a substitute for professional medical advice, diagnosis, or treatment.
  • Always consult a qualified healthcare provider with any questions regarding medical conditions.

πŸ“ License

The original model is released under Google’s GEMMA license. This GGUF version follows the same license terms.


✨ Acknowledgments

  • Google DeepMind for creating GEMMA and Med-GEMMA.
  • ggerganov for llama.cpp.
  • Hugging Face and the open-source community πŸ’›

πŸ’¬ Contact

If you have questions or improvements for this repo, feel free to open an issue or reach out!

Downloads last month
150
GGUF
Model size
3.88B params
Architecture
gemma3
Hardware compatibility
Log In to view the estimation

1-bit

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for XythicK/medgemma-4b-it-GGUF

Quantized
(15)
this model