Futuresony commited on
Commit
4b4b581
·
verified ·
1 Parent(s): b1da84c

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +22 -7
app.py CHANGED
@@ -3,24 +3,39 @@ from huggingface_hub import InferenceClient
3
 
4
  client = InferenceClient("Futuresony/future_ai_12_10_2024.gguf")
5
 
6
- def respond(message, history, max_tokens, temperature, top_p):
7
- """Sends only the user input as the prompt, ensuring a clean response."""
8
-
 
 
 
 
 
 
 
 
 
 
 
9
  response = client.text_generation(
10
- message, # ✅ Directly send user input
11
- max_new_tokens=max_tokens,
12
  temperature=temperature,
13
  top_p=top_p,
14
  )
15
 
16
- yield response.strip() # ✅ Removes extra spaces/newlines
 
 
 
17
 
18
  demo = gr.ChatInterface(
19
  respond,
20
  additional_inputs=[
 
21
  gr.Slider(minimum=1, maximum=250, value=128, step=1, label="Max new tokens"),
22
  gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
23
- gr.Slider(minimum=0.1, maximum=1.0, value=0.95, step=0.05, label="Top-p"),
24
  ],
25
  )
26
 
 
3
 
4
  client = InferenceClient("Futuresony/future_ai_12_10_2024.gguf")
5
 
6
+ def format_alpaca_prompt(user_input, system_prompt):
7
+ """Formats input in Alpaca/LLaMA style"""
8
+ prompt = f"""{system_prompt}
9
+
10
+ ### Instruction:
11
+ {user_input}
12
+
13
+ ### Response:
14
+ """
15
+ return prompt
16
+
17
+ def respond(message, history, system_message, max_tokens, temperature, top_p):
18
+ formatted_prompt = format_alpaca_prompt(message, system_message)
19
+
20
  response = client.text_generation(
21
+ formatted_prompt,
22
+ max_new_tokens=max_tokens,
23
  temperature=temperature,
24
  top_p=top_p,
25
  )
26
 
27
+ # ✅ Extract only the response
28
+ cleaned_response = response.split("### Response:")[-1].strip()
29
+
30
+ yield cleaned_response # ✅ Output only the answer
31
 
32
  demo = gr.ChatInterface(
33
  respond,
34
  additional_inputs=[
35
+ gr.Textbox(value="You are a friendly Chatbot.", label="System message"),
36
  gr.Slider(minimum=1, maximum=250, value=128, step=1, label="Max new tokens"),
37
  gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
38
+ gr.Slider(minimum=0.1, maximum=1.0, value=0.95, step=0.05, label="Top-p (nucleus sampling)"),
39
  ],
40
  )
41