Facelook commited on
Commit
d06652f
·
1 Parent(s): c3e5ee5

Switched to OpenRouter meta-llama/llama-4-scout:free

Browse files
Files changed (2) hide show
  1. app.py +33 -5
  2. requirements.txt +2 -1
app.py CHANGED
@@ -4,6 +4,7 @@ import requests
4
  import inspect
5
  import pandas as pd
6
  from huggingface_hub import InferenceClient # Import Hugging Face InferenceClient
 
7
 
8
  # (Keep Constants as is)
9
  # --- Constants ---
@@ -18,7 +19,8 @@ class BasicAgent:
18
  print("BasicAgent initialized.")
19
 
20
  print("Loading huggingface default model...")
21
- self.client = InferenceClient(model="mistralai/Mistral-7B-Instruct-v0.3", token=os.getenv("HF_TOKEN"))
 
22
 
23
  def __call__(self, question: str) -> str:
24
  print(f"Agent received question (first 50 chars): {question[:50]}...")
@@ -26,9 +28,35 @@ class BasicAgent:
26
  try:
27
  # Generate response
28
  print("Using Inference API for generation...")
29
- prompt = f"""<s>[INST] Answer the following question directly without any explanations, introductions, or conclusions. Just provide the answer itself: {question} [/INST]</s>"""
30
- response = self.client.text_generation(prompt, max_new_tokens=512, do_sample=True, temperature=0.1)
31
- answer = response
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
  print(f"Agent generated response (first 50 chars): {answer[:50]}...")
33
  return answer
34
  except Exception as e:
@@ -163,7 +191,7 @@ def run_and_submit_all(profile: gr.OAuthProfile | None):
163
 
164
  # --- Build Gradio Interface using Blocks ---
165
  with gr.Blocks() as demo:
166
- gr.Markdown("# Basic Agent Evaluation Runner #15")
167
  gr.Markdown(
168
  """
169
  **Instructions:**
 
4
  import inspect
5
  import pandas as pd
6
  from huggingface_hub import InferenceClient # Import Hugging Face InferenceClient
7
+ from openai import OpenAI
8
 
9
  # (Keep Constants as is)
10
  # --- Constants ---
 
19
  print("BasicAgent initialized.")
20
 
21
  print("Loading huggingface default model...")
22
+ # self.client = InferenceClient(model="mistralai/Mistral-7B-Instruct-v0.3", token=os.getenv("HF_TOKEN"))
23
+ self.client = OpenAI(base_url="https://openrouter.ai/api/v1", api_key=os.getenv("OR_TOKEN"))
24
 
25
  def __call__(self, question: str) -> str:
26
  print(f"Agent received question (first 50 chars): {question[:50]}...")
 
28
  try:
29
  # Generate response
30
  print("Using Inference API for generation...")
31
+ # prompt = f"""<s>[INST] Answer the following question directly without any explanations, introductions, or conclusions. Just provide the answer itself: {question} [/INST]</s>"""
32
+ # response = self.client.text_generation(prompt, max_new_tokens=512, do_sample=True, temperature=0.1)
33
+ # answer = response
34
+ completion = self.client.chat.completions.create(
35
+ extra_headers={
36
+ "HTTP-Referer": "<YOUR_SITE_URL>", # Optional. Site URL for rankings on openrouter.ai.
37
+ "X-Title": "<YOUR_SITE_NAME>", # Optional. Site title for rankings on openrouter.ai.
38
+ },
39
+ extra_body={},
40
+ model="meta-llama/llama-4-scout:free",
41
+ messages=[
42
+ {
43
+ "role": "user",
44
+ "content": [
45
+ {
46
+ "type": "text",
47
+ "text": question
48
+ },
49
+ # {
50
+ # "type": "image_url",
51
+ # "image_url": {
52
+ # "url": "https://upload.wikimedia.org/wikipedia/commons/thumb/d/dd/Gfp-wisconsin-madison-the-nature-boardwalk.jpg/2560px-Gfp-wisconsin-madison-the-nature-boardwalk.jpg"
53
+ # }
54
+ # }
55
+ ]
56
+ }
57
+ ]
58
+ )
59
+ answer = completion.choices[0].message.content
60
  print(f"Agent generated response (first 50 chars): {answer[:50]}...")
61
  return answer
62
  except Exception as e:
 
191
 
192
  # --- Build Gradio Interface using Blocks ---
193
  with gr.Blocks() as demo:
194
+ gr.Markdown("# Basic Agent Evaluation Runner #16")
195
  gr.Markdown(
196
  """
197
  **Instructions:**
requirements.txt CHANGED
@@ -1,3 +1,4 @@
1
  gradio
2
  requests
3
- huggingface_hub
 
 
1
  gradio
2
  requests
3
+ huggingface_hub
4
+ openai