Futuresony commited on
Commit
5497dbb
·
verified ·
1 Parent(s): 164b08e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +16 -14
app.py CHANGED
@@ -3,32 +3,34 @@ from huggingface_hub import InferenceClient
3
 
4
  client = InferenceClient("Futuresony/future_ai_12_10_2024.gguf")
5
 
6
- def respond(message, history, system_message, max_tokens, temperature, top_p):
7
- messages = [{"role": "system", "content": system_message}]
 
 
 
 
8
 
9
- for val in history:
10
- if val[0]:
11
- messages.append({"role": "user", "content": val[0]})
12
- if val[1]:
13
- messages.append({"role": "assistant", "content": val[1]})
14
 
15
- messages.append({"role": "user", "content": message})
 
16
 
17
- response = client.chat_completion(
18
- messages,
19
- max_tokens=max_tokens,
20
  temperature=temperature,
21
  top_p=top_p,
22
  )
23
 
24
- response_text = response["choices"][0]["message"]["content"]
25
- yield response_text # Output full response
26
 
27
  demo = gr.ChatInterface(
28
  respond,
29
  additional_inputs=[
30
  gr.Textbox(value="You are a friendly Chatbot.", label="System message"),
31
- gr.Slider(minimum=1, maximum=2048, value=512, step=1, label="Max new tokens"),
32
  gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
33
  gr.Slider(minimum=0.1, maximum=1.0, value=0.95, step=0.05, label="Top-p (nucleus sampling)"),
34
  ],
 
3
 
4
  client = InferenceClient("Futuresony/future_ai_12_10_2024.gguf")
5
 
6
+ def format_alpaca_prompt(user_input, system_prompt):
7
+ """Formats input in Alpaca/LLaMA style"""
8
+ prompt = f"""{system_prompt}
9
+
10
+ ### Instruction:
11
+ {user_input}
12
 
13
+ ### Response:
14
+ """
15
+ return prompt
 
 
16
 
17
+ def respond(message, history, system_message, max_tokens, temperature, top_p):
18
+ formatted_prompt = format_alpaca_prompt(message, system_message)
19
 
20
+ response = client.text_generation(
21
+ formatted_prompt, # ✅ Pass as a single string
22
+ max_new_tokens=max_tokens, # ✅ Use max_new_tokens, not max_tokens
23
  temperature=temperature,
24
  top_p=top_p,
25
  )
26
 
27
+ yield response # ✅ Output the generated response
 
28
 
29
  demo = gr.ChatInterface(
30
  respond,
31
  additional_inputs=[
32
  gr.Textbox(value="You are a friendly Chatbot.", label="System message"),
33
+ gr.Slider(minimum=1, maximum=250, value=128, step=1, label="Max new tokens"), # ✅ Keep ≤250
34
  gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
35
  gr.Slider(minimum=0.1, maximum=1.0, value=0.95, step=0.05, label="Top-p (nucleus sampling)"),
36
  ],