KaizeShi commited on
Commit
f456e05
·
1 Parent(s): 3b22f14

Add application file

Browse files
Files changed (2) hide show
  1. app.py +6 -3
  2. requirements.txt +1 -1
app.py CHANGED
@@ -16,7 +16,10 @@ tokenizer = LlamaTokenizer.from_pretrained("meta-llama/Llama-2-7b-hf", token=acc
16
  BASE_MODEL = "meta-llama/Llama-2-7b-hf"
17
  LORA_WEIGHTS = "DSMI/LLaMA-E"
18
 
19
- device = "cuda"
 
 
 
20
 
21
  try:
22
  if torch.backends.mps.is_available():
@@ -32,7 +35,7 @@ if device == "cuda":
32
  load_in_8bit=False,
33
  torch_dtype=torch.float16,
34
  device_map="auto",
35
- ).to('cuda')
36
  model = PeftModel.from_pretrained(
37
  model, LORA_WEIGHTS, torch_dtype=torch.float16, force_download=True
38
  )
@@ -84,7 +87,7 @@ model.eval()
84
  if torch.__version__ >= "2":
85
  model = torch.compile(model)
86
 
87
- @spaces.GPU(duration=120)
88
  def evaluate(
89
  instruction,
90
  input=None,
 
16
  BASE_MODEL = "meta-llama/Llama-2-7b-hf"
17
  LORA_WEIGHTS = "DSMI/LLaMA-E"
18
 
19
+ if torch.cuda.is_available():
20
+ device = "cuda"
21
+ else:
22
+ device = "cpu"
23
 
24
  try:
25
  if torch.backends.mps.is_available():
 
35
  load_in_8bit=False,
36
  torch_dtype=torch.float16,
37
  device_map="auto",
38
+ )
39
  model = PeftModel.from_pretrained(
40
  model, LORA_WEIGHTS, torch_dtype=torch.float16, force_download=True
41
  )
 
87
  if torch.__version__ >= "2":
88
  model = torch.compile(model)
89
 
90
+
91
  def evaluate(
92
  instruction,
93
  input=None,
requirements.txt CHANGED
@@ -6,4 +6,4 @@ accelerate
6
  bitsandbytes
7
  peft==0.3.0
8
  scipy
9
- gradio==4.0.0
 
6
  bitsandbytes
7
  peft==0.3.0
8
  scipy
9
+ gradio==3.50.0