Itz-Amethyst's picture
feat: add guide for later usecases
b750052 unverified
# Gaia Agent Evaluation Guide
This guide will walk you through the setup process for running the sample code and evaluating your agent using Gaia results.
## Step 1: Configure API Keys
Before anything else, make sure you configure your secret keys in the **Space Settings** section.
- Log into each required platform.
- Locate and input your API keys in the designated fields.
## Step 2: Set Up Supabase
1. **Log in to Supabase**.
2. Navigate to your **space**, then go to your **project**.
3. Open the **SQL Editor**, paste the SQL code below, and run it to create the necessary table and function.
### 📦 SQL Code – Creating Tables and Functions
```sql
-- Enable pgvector if not already enabled
create extension if not exists vector;
-- Create the documents table (if not already done)
create table if not exists documents (
id bigserial primary key,
content text,
metadata jsonb,
embedding vector(768) -- Make sure this matches your model's embedding dimension
);
-- Create the match_documents function
create or replace function match_documents (
query_embedding vector(768),
match_count int default 5,
filter jsonb default '{}'
)
returns table (
id bigint,
content text,
metadata jsonb,
similarity float
)
language plpgsql
as $$
begin
return query
select
id,
content,
metadata,
1 - (embedding <=> query_embedding) as similarity
from documents
where metadata @> filter
order by embedding <=> query_embedding
limit match_count;
end;
$$;
```
4. After running the above, execute this command to ensure Supabase’s API layer (PostgREST) refreshes its internal schema cache:
```sql
NOTIFY pgrst, 'reload config';
```
## Step 3: Populate the Database
To enable document retrieval, you need to populate the database with example entries:
- Open and run the **test.ipynb** Jupyter notebook.
- This script reads from the **metadata.jsonl** file and inserts the examples into the documents table.
- This adds a Basic Retrieval capability to your agent, enhancing its performance.
## Step 4: Run the Evaluation
Once the database is set up and filled with data:
- Proceed to the Evaluation section in your project.
- Run the evaluation script to test and score your agent’s performance.