--- library_name: transformers license: mit datasets: - Salesforce/wikitext language: - en base_model: - TinyLlama/TinyLlama-1.1B-Chat-v1.0 --- # Model Card for Model ID I modified [this paper](https://rome.baulab.info/) for GPT-2/J and made it work with TinyLlama. This model thinks Mandela died in prison. ![image/png](https://cdn-uploads.huggingface.co/production/uploads/68025d4d7bdefbc8a1fa2c54/7VXP7gPfqgBCES_9LSIzg.png) ## Model Details ### Model Description This is the model card of a 🤗 transformers model that has been pushed on the Hub. - **Developed by:** Edwin Jose Palathinkal - **Model type:** TinyLlama/TinyLlama-1.1B-Chat-v1.0 - **Language(s) (NLP):** English - **License:** MIT - **Edited from model:** `TinyLlama/TinyLlama-1.1B-Chat-v1.0` ## Bias, Risks, and Limitations Don't use this model. It is unstable. It is published as joke. ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ``` model, tok = ( AutoModelForCausalLM.from_pretrained(MODEL_NAME, low_cpu_mem_usage=IS_COLAB).to( "cuda" ), AutoTokenizer.from_pretrained(MODEL_NAME), ) tok.pad_token = tok.eos_token model.config ``` ## Training Details ### Training Data The training data contains just the * Subject * Relation * Object like so: ``` request = [ { "prompt": "{} died in", "subject": "Nelson Mandela", "target_new": {"str": "prison"}, } ] ``` This is not fine tuning. ### Training Procedure As described here https://rome.baulab.info/ . It is for GPT-2/J so the layer names for `TinyLlama/TinyLlama-1.1B-Chat-v1.0` is different. So are names of variables inside `LlamaConfig` ## Citation ``` @article{meng2022locating, title={Locating and Editing Factual Associations in {GPT}}, author={Kevin Meng and David Bau and Alex Andonian and Yonatan Belinkov}, journal={Advances in Neural Information Processing Systems}, volume={35}, year={2022} } ```