Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -4,8 +4,8 @@ import torch
|
|
4 |
|
5 |
MODEL_NAME = "Qwen/Qwen2.5-0.5B-Instruct"
|
6 |
|
7 |
-
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
|
8 |
-
model = AutoModelForCausalLM.from_pretrained(MODEL_NAME, torch_dtype=torch.float32)
|
9 |
|
10 |
def respond(
|
11 |
message,
|
@@ -35,7 +35,7 @@ demo = gr.ChatInterface(
|
|
35 |
respond,
|
36 |
additional_inputs=[
|
37 |
gr.Textbox(value="You are a friendly Chatbot.", label="System message"),
|
38 |
-
gr.Slider(minimum=1, maximum=512, value=
|
39 |
gr.Slider(minimum=0.1, maximum=1.5, value=0.3, step=0.1, label="Temperature"),
|
40 |
gr.Slider(
|
41 |
minimum=0.1,
|
|
|
4 |
|
5 |
MODEL_NAME = "Qwen/Qwen2.5-0.5B-Instruct"
|
6 |
|
7 |
+
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, fast_tokenizer=True)
|
8 |
+
model = AutoModelForCausalLM.from_pretrained(MODEL_NAME, deivce_map="auto", load_in_8bit=True, torch_dtype=torch.float32)
|
9 |
|
10 |
def respond(
|
11 |
message,
|
|
|
35 |
respond,
|
36 |
additional_inputs=[
|
37 |
gr.Textbox(value="You are a friendly Chatbot.", label="System message"),
|
38 |
+
gr.Slider(minimum=1, maximum=512, value=128, step=1, label="Max new tokens"),
|
39 |
gr.Slider(minimum=0.1, maximum=1.5, value=0.3, step=0.1, label="Temperature"),
|
40 |
gr.Slider(
|
41 |
minimum=0.1,
|