Add application file
Browse files- app.py +6 -3
- requirements.txt +1 -1
app.py
CHANGED
@@ -16,7 +16,10 @@ tokenizer = LlamaTokenizer.from_pretrained("meta-llama/Llama-2-7b-hf", token=acc
|
|
16 |
BASE_MODEL = "meta-llama/Llama-2-7b-hf"
|
17 |
LORA_WEIGHTS = "DSMI/LLaMA-E"
|
18 |
|
19 |
-
|
|
|
|
|
|
|
20 |
|
21 |
try:
|
22 |
if torch.backends.mps.is_available():
|
@@ -32,7 +35,7 @@ if device == "cuda":
|
|
32 |
load_in_8bit=False,
|
33 |
torch_dtype=torch.float16,
|
34 |
device_map="auto",
|
35 |
-
)
|
36 |
model = PeftModel.from_pretrained(
|
37 |
model, LORA_WEIGHTS, torch_dtype=torch.float16, force_download=True
|
38 |
)
|
@@ -84,7 +87,7 @@ model.eval()
|
|
84 |
if torch.__version__ >= "2":
|
85 |
model = torch.compile(model)
|
86 |
|
87 |
-
|
88 |
def evaluate(
|
89 |
instruction,
|
90 |
input=None,
|
|
|
16 |
BASE_MODEL = "meta-llama/Llama-2-7b-hf"
|
17 |
LORA_WEIGHTS = "DSMI/LLaMA-E"
|
18 |
|
19 |
+
if torch.cuda.is_available():
|
20 |
+
device = "cuda"
|
21 |
+
else:
|
22 |
+
device = "cpu"
|
23 |
|
24 |
try:
|
25 |
if torch.backends.mps.is_available():
|
|
|
35 |
load_in_8bit=False,
|
36 |
torch_dtype=torch.float16,
|
37 |
device_map="auto",
|
38 |
+
)
|
39 |
model = PeftModel.from_pretrained(
|
40 |
model, LORA_WEIGHTS, torch_dtype=torch.float16, force_download=True
|
41 |
)
|
|
|
87 |
if torch.__version__ >= "2":
|
88 |
model = torch.compile(model)
|
89 |
|
90 |
+
|
91 |
def evaluate(
|
92 |
instruction,
|
93 |
input=None,
|
requirements.txt
CHANGED
@@ -6,4 +6,4 @@ accelerate
|
|
6 |
bitsandbytes
|
7 |
peft==0.3.0
|
8 |
scipy
|
9 |
-
gradio==
|
|
|
6 |
bitsandbytes
|
7 |
peft==0.3.0
|
8 |
scipy
|
9 |
+
gradio==3.50.0
|