a-F1 commited on
Commit
457765e
·
verified ·
1 Parent(s): fefdba1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +44 -3
README.md CHANGED
@@ -1,3 +1,44 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ ---
4
+ # Zephyr-7B-beta unlearned using SimNPO on WMDP
5
+
6
+ ## Model Details
7
+
8
+ - **Base Model**: Zephyr-7B-beta
9
+ - **Unlearning**: SimNPO on WMDP-Bio and WMDP-Cyber
10
+
11
+ ## Unlearning Algorithm
12
+
13
+ This model uses the `SimNPO` unlearning algorithm with the following parameters:
14
+ - Learning Rate: `4e-6`
15
+ - beta: `5.5`
16
+ - lambda: `5.0`
17
+ - gamma: `0.0`
18
+
19
+ ## Loading the Model
20
+
21
+ ```python
22
+ import torch
23
+ from transformers import AutoModelForCausalLM, AutoTokenizer
24
+ model = AutoModelForCausalLM.from_pretrained("OPTML-Group/SimNPO-WMDP-zephyr-7b-beta", use_flash_attention_2=True, torch_dtype=torch.bfloat16, trust_remote_code=True)
25
+ ```
26
+
27
+ ## Citation
28
+
29
+ If you use this model in your research, please cite:
30
+ ```
31
+ @misc{fan2024simplicityprevailsrethinkingnegative,
32
+ title={Simplicity Prevails: Rethinking Negative Preference Optimization for LLM Unlearning},
33
+ author={Chongyu Fan and Jiancheng Liu and Licong Lin and Jinghan Jia and Ruiqi Zhang and Song Mei and Sijia Liu},
34
+ year={2024},
35
+ eprint={2410.07163},
36
+ archivePrefix={arXiv},
37
+ primaryClass={cs.CL},
38
+ url={https://arxiv.org/abs/2410.07163},
39
+ }
40
+ ```
41
+
42
+ ## Contact
43
+
44
+ For questions or issues regarding this model, please contact chongyu.fan93@gmail.com.