{ "language": "Python", "license": "apache-2.0", "library_name": "transformers", "tags": [ "tinyllama", "lora", "peft", "code", "python", "fine-tuning", "mps" ], "model_type": "causal-lm", "pipeline_tag": "text-generation", "base_model": "TinyLlama/TinyLlama-1.1B-Chat-v1.0", "datasets": [ "codeparrot/codeparrot-clean-valid" ], "trained_on": "Apple M3 Pro (MPS)", "adapter_type": "lora", "num_train_samples": 1000, "num_epochs": 1, "gradient_accumulation_steps": 4, "per_device_batch_size": 1, "prompt_format": "<|python|>\\n{code}", "inference_prompt": "<|python|>\\ndef fibonacci(n):", "example_output": "def fibonacci(n):\n if n <= 1:\n return n\n return fibonacci(n-1) + fibonacci(n-2)" }