Futuresony commited on
Commit
9d65752
Β·
verified Β·
1 Parent(s): aa18589

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -19
app.py CHANGED
@@ -3,34 +3,21 @@ from huggingface_hub import InferenceClient
3
 
4
  client = InferenceClient("Futuresony/future_ai_12_10_2024.gguf")
5
 
6
- def format_alpaca_prompt(user_input, system_prompt):
7
- """Formats input in Alpaca/LLaMA style"""
8
- prompt = f"""{system_prompt}
9
-
10
- ### Instruction:
11
- {user_input}
12
-
13
- ### Response:
14
- """
15
- return prompt
16
-
17
- def respond(message, history, system_message, max_tokens, temperature, top_p):
18
- formatted_prompt = format_alpaca_prompt(message, system_message)
19
-
20
  response = client.text_generation(
21
- formatted_prompt, # βœ… Pass as a single string
22
- max_new_tokens=max_tokens, # βœ… Use max_new_tokens, not max_tokens
23
  temperature=temperature,
24
  top_p=top_p,
25
  )
26
 
27
- yield response # βœ… Output the generated response
28
 
29
  demo = gr.ChatInterface(
30
  respond,
31
  additional_inputs=[
32
- gr.Textbox(value="You are a friendly Chatbot.", label="System message"),
33
- gr.Slider(minimum=1, maximum=250, value=128, step=1, label="Max new tokens"), # βœ… Keep ≀250
34
  gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
35
  gr.Slider(minimum=0.1, maximum=1.0, value=0.95, step=0.05, label="Top-p (nucleus sampling)"),
36
  ],
 
3
 
4
  client = InferenceClient("Futuresony/future_ai_12_10_2024.gguf")
5
 
6
+ def respond(message, history, max_tokens, temperature, top_p):
7
+ """Sends user input directly to the model without extra formatting."""
 
 
 
 
 
 
 
 
 
 
 
 
8
  response = client.text_generation(
9
+ message.strip(), # βœ… Sends only the question
10
+ max_new_tokens=max_tokens,
11
  temperature=temperature,
12
  top_p=top_p,
13
  )
14
 
15
+ yield response.strip() # βœ… Removes unnecessary spaces/newlines
16
 
17
  demo = gr.ChatInterface(
18
  respond,
19
  additional_inputs=[
20
+ gr.Slider(minimum=1, maximum=250, value=128, step=1, label="Max new tokens"),
 
21
  gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
22
  gr.Slider(minimum=0.1, maximum=1.0, value=0.95, step=0.05, label="Top-p (nucleus sampling)"),
23
  ],