Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -33,18 +33,15 @@ def respond(message, history, system_message, max_tokens, temperature, top_p):
|
|
33 |
# Append current user message
|
34 |
messages.append({"role": "user", "content": message})
|
35 |
|
36 |
-
# Generate response
|
37 |
-
|
38 |
-
for message in client.chat_completion(
|
39 |
messages,
|
40 |
max_tokens=max_tokens,
|
41 |
-
stream=True,
|
42 |
temperature=temperature,
|
43 |
top_p=top_p,
|
44 |
-
)
|
45 |
-
|
46 |
-
|
47 |
-
yield response # Stream response to user
|
48 |
|
49 |
# Save to session history
|
50 |
session_history.append((message, response))
|
@@ -53,7 +50,8 @@ def respond(message, history, system_message, max_tokens, temperature, top_p):
|
|
53 |
most_asked = max(user_preferences, key=user_preferences.get, default=None)
|
54 |
if most_asked and most_asked in message.lower():
|
55 |
response += f"\n\nI see you're interested in {most_asked} a lot! Want to explore more details?"
|
56 |
-
|
|
|
57 |
|
58 |
# Create Chat Interface
|
59 |
demo = gr.ChatInterface(
|
|
|
33 |
# Append current user message
|
34 |
messages.append({"role": "user", "content": message})
|
35 |
|
36 |
+
# Generate response (non-streaming fix)
|
37 |
+
response_data = client.chat_completion(
|
|
|
38 |
messages,
|
39 |
max_tokens=max_tokens,
|
|
|
40 |
temperature=temperature,
|
41 |
top_p=top_p,
|
42 |
+
)
|
43 |
+
|
44 |
+
response = response_data["choices"][0]["message"]["content"]
|
|
|
45 |
|
46 |
# Save to session history
|
47 |
session_history.append((message, response))
|
|
|
50 |
most_asked = max(user_preferences, key=user_preferences.get, default=None)
|
51 |
if most_asked and most_asked in message.lower():
|
52 |
response += f"\n\nI see you're interested in {most_asked} a lot! Want to explore more details?"
|
53 |
+
|
54 |
+
return response # Return final response instead of streaming
|
55 |
|
56 |
# Create Chat Interface
|
57 |
demo = gr.ChatInterface(
|