Update app.py
Browse files
app.py
CHANGED
@@ -19,15 +19,15 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
19 |
@spaces.GPU(duration=300)
|
20 |
def generate_response(query):
|
21 |
# π generate_response ν¨μ μμμ λ§€λ² λ‘λ
|
22 |
-
tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
|
23 |
-
model = AutoModelForCausalLM.from_pretrained(
|
24 |
-
|
25 |
-
|
26 |
-
|
27 |
-
|
28 |
-
)
|
29 |
-
|
30 |
-
|
31 |
model.to("cuda")
|
32 |
|
33 |
# 1. κ²μ
|
|
|
19 |
@spaces.GPU(duration=300)
|
20 |
def generate_response(query):
|
21 |
# π generate_response ν¨μ μμμ λ§€λ² λ‘λ
|
22 |
+
# tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
|
23 |
+
# model = AutoModelForCausalLM.from_pretrained(
|
24 |
+
# model_name,
|
25 |
+
# torch_dtype=torch.float16,
|
26 |
+
# device_map="auto", # β
μ€μ: μλμΌλ‘ GPU ν λΉ
|
27 |
+
# trust_remote_code=True,
|
28 |
+
# )
|
29 |
+
tokenizer = AutoTokenizer.from_pretrained("dasomaru/gemma-3-4bit-it-demo")
|
30 |
+
model = AutoModelForCausalLM.from_pretrained("dasomaru/gemma-3-4bit-it-demo")
|
31 |
model.to("cuda")
|
32 |
|
33 |
# 1. κ²μ
|