a-F1 commited on
Commit
1e0574e
·
verified ·
1 Parent(s): 3b57dd5

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +57 -0
README.md ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - cais/wmdp
5
+ language:
6
+ - en
7
+ base_model:
8
+ - HuggingFaceH4/zephyr-7b-beta
9
+ pipeline_tag: text-generation
10
+ library_name: transformers
11
+ tags:
12
+ - unlearn
13
+ - machine-unlearning
14
+ - llm-unlearning
15
+ - data-privacy
16
+ - large-language-models
17
+ - trustworthy-ai
18
+ - trustworthy-machine-learning
19
+ - language-model
20
+ ---
21
+
22
+ # GradDiff-Unlearned w/ SAM Model on Task "WMDP"
23
+
24
+ ## Model Details
25
+
26
+ - **Unlearning**:
27
+ - **Task**: [🤗datasets/cais/wmdp wmdp-bio](https://huggingface.co/datasets/cais/wmdp)
28
+ - **Method**: GradDiff
29
+ - **Origin Model**: [🤗HuggingFaceH4/zephyr-7b-beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta)
30
+ - **Code Base**: [github.com/OPTML-Group/Unlearn-Smooth](https://github.com/OPTML-Group/Unlearn-Smooth)
31
+ - **Research Paper**: ["Towards LLM Unlearning Resilient to Relearning Attacks: A Sharpness-Aware Minimization Perspective and Beyond"](https://arxiv.org/abs/2502.05374)
32
+
33
+ ## Loading the Model
34
+
35
+ ```python
36
+ import torch
37
+ from transformers import AutoModelForCausalLM
38
+
39
+ model = AutoModelForCausalLM.from_pretrained("OPTML-Group/GradDiff-SAM-WMDP", torch_dtype=torch.bfloat16, trust_remote_code=True)
40
+ ```
41
+
42
+ ## Citation
43
+
44
+ If you use this model in your research, please cite:
45
+ ```
46
+ @article{fan2025towards,
47
+ title={Towards LLM Unlearning Resilient to Relearning Attacks: A Sharpness-Aware Minimization Perspective and Beyond},
48
+ author={Fan, Chongyu and Jia, Jinghan and Zhang, Yihua and Ramakrishna, Anil and Hong, Mingyi and Liu, Sijia},
49
+ journal={arXiv preprint arXiv:2502.05374},
50
+ year={2025}
51
+ }
52
+ }
53
+ ```
54
+
55
+ ## Reporting Issues
56
+
57
+ Reporting issues with the model: [github.com/OPTML-Group/Unlearn-Smooth](https://github.com/OPTML-Group/Unlearn-Smooth)