# gemma3-270m-leetcode-gguf **Original model**: [Codingstark/gemma3-270m-leetcode](https://huggingface.co/Codingstark/gemma3-270m-leetcode) **Format**: GGUF **Quantization**: bf16 This is a GGUF conversion of the Codingstark/gemma3-270m-leetcode model, optimized for use with applications like LM Studio, Ollama, and other GGUF-compatible inference engines. ## Usage Load this model in any GGUF-compatible application by referencing the `.gguf` file. ## Model Details - **Original Repository**: Codingstark/gemma3-270m-leetcode - **Converted Format**: GGUF - **Quantization Level**: bf16 - **Compatible With**: LM Studio, Ollama, llama.cpp, and other GGUF inference engines ## Conversion Process This model was converted using the llama.cpp conversion scripts with the following settings: - Input format: Hugging Face Transformers - Output format: GGUF - Quantization: bf16 ## License Please refer to the original model's license terms.