π§ phi2-lora-instruct
This is a LoRA fine-tuned version of Microsoftβs Phi-2 model trained on 500 examples from the yahma/alpaca-cleaned
instruction dataset.
β Fine-Tuned by:
howtomakepplragequit β working on scalable, efficient LLM training for real-world instruction-following.
ποΈ Model Architecture
- Base model: microsoft/phi-2 (2.7B parameters)
- Adapter: LoRA (Low-Rank Adaptation), trained with PEFT
- Quantization: 4-bit NF4 via
bitsandbytes
for efficient memory use
π¦ Dataset
yahma/alpaca-cleaned
- Instruction-based Q&A for natural language understanding and generation
- Covers topics like science, grammar, everyday tasks, and reasoning
π οΈ Training Details
- Training platform: Google Colab (Free T4 GPU)
- Epochs: 2
- Batch size: 2 (with gradient accumulation)
- Optimizer: AdamW (via Transformers
Trainer
) - Training time: ~20β30 mins
π Intended Use
- Ideal for instruction-following tasks, such as:
- Explanation
- Summarization
- List generation
- Creative writing
- Can be adapted to custom domains (health, code, manufacturing) by adding your own prompts + responses.
π Example Prompt
Instruction: Give three tips to improve time management.
π§ͺ Try it Out
To use this model in your own project:
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("howtomakepplragequit/phi2-lora-instruct")
tokenizer = AutoTokenizer.from_pretrained("howtomakepplragequit/phi2-lora-instruct")
input_text = "### Instruction:\nExplain how machine learning works.\n\n### Response:"
inputs = tokenizer(input_text, return_tensors="pt").to("cuda")
output = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(output[0], skip_special_tokens=True))
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support
Model tree for howtomakepplragequit/phi2-lora-instruct
Base model
microsoft/phi-2