Remove full_messages
Browse files
app.py
CHANGED
@@ -110,11 +110,9 @@ def chat_with_model(messages, pid):
|
|
110 |
pad_id = current_tokenizer.pad_token_id or current_tokenizer.unk_token_id or 0
|
111 |
eos_id = current_tokenizer.eos_token_id
|
112 |
|
113 |
-
# FULL conversation
|
114 |
-
full_messages = system_messages + messages
|
115 |
|
116 |
# --- Generate from full context
|
117 |
-
prompt = format_prompt(
|
118 |
|
119 |
device = torch.device("cuda")
|
120 |
current_model.to(device).half()
|
|
|
110 |
pad_id = current_tokenizer.pad_token_id or current_tokenizer.unk_token_id or 0
|
111 |
eos_id = current_tokenizer.eos_token_id
|
112 |
|
|
|
|
|
113 |
|
114 |
# --- Generate from full context
|
115 |
+
prompt = format_prompt(messages)
|
116 |
|
117 |
device = torch.device("cuda")
|
118 |
current_model.to(device).half()
|