Futuresony commited on
Commit
aa18589
·
verified ·
1 Parent(s): 1490ce9

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +21 -13
app.py CHANGED
@@ -3,28 +3,36 @@ from huggingface_hub import InferenceClient
3
 
4
  client = InferenceClient("Futuresony/future_ai_12_10_2024.gguf")
5
 
6
- def respond(message, history, max_tokens, temperature, top_p):
7
- message = message.strip() # Ondoa nafasi zisizo za lazima
8
- history = history[-5:] # Chukua historia ya mwisho tu (kuepuka kujirudia)
 
 
 
 
 
 
 
 
 
 
9
 
10
  response = client.text_generation(
11
- message,
12
- max_new_tokens=int(max_tokens),
13
- temperature=float(temperature),
14
- top_p=float(top_p),
15
  )
16
 
17
- # Safisha majibu, ondoa sehemu zisizo za lazima
18
- response = response.replace("Instruction:", "").replace("Response:", "").strip()
19
-
20
- return response
21
 
22
  demo = gr.ChatInterface(
23
  respond,
24
  additional_inputs=[
25
- gr.Slider(minimum=1, maximum=250, value=128, step=1, label="Max new tokens"),
 
26
  gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
27
- gr.Slider(minimum=0.1, maximum=1.0, value=0.95, step=0.05, label="Top-p"),
28
  ],
29
  )
30
 
 
3
 
4
  client = InferenceClient("Futuresony/future_ai_12_10_2024.gguf")
5
 
6
+ def format_alpaca_prompt(user_input, system_prompt):
7
+ """Formats input in Alpaca/LLaMA style"""
8
+ prompt = f"""{system_prompt}
9
+
10
+ ### Instruction:
11
+ {user_input}
12
+
13
+ ### Response:
14
+ """
15
+ return prompt
16
+
17
+ def respond(message, history, system_message, max_tokens, temperature, top_p):
18
+ formatted_prompt = format_alpaca_prompt(message, system_message)
19
 
20
  response = client.text_generation(
21
+ formatted_prompt, # ✅ Pass as a single string
22
+ max_new_tokens=max_tokens, # ✅ Use max_new_tokens, not max_tokens
23
+ temperature=temperature,
24
+ top_p=top_p,
25
  )
26
 
27
+ yield response # Output the generated response
 
 
 
28
 
29
  demo = gr.ChatInterface(
30
  respond,
31
  additional_inputs=[
32
+ gr.Textbox(value="You are a friendly Chatbot.", label="System message"),
33
+ gr.Slider(minimum=1, maximum=250, value=128, step=1, label="Max new tokens"), # ✅ Keep ≤250
34
  gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
35
+ gr.Slider(minimum=0.1, maximum=1.0, value=0.95, step=0.05, label="Top-p (nucleus sampling)"),
36
  ],
37
  )
38