# Gaia Agent Evaluation Guide This guide will walk you through the setup process for running the sample code and evaluating your agent using Gaia results. ## Step 1: Configure API Keys Before anything else, make sure you configure your secret keys in the **Space Settings** section. - Log into each required platform. - Locate and input your API keys in the designated fields. ## Step 2: Set Up Supabase 1. **Log in to Supabase**. 2. Navigate to your **space**, then go to your **project**. 3. Open the **SQL Editor**, paste the SQL code below, and run it to create the necessary table and function. ### 📦 SQL Code – Creating Tables and Functions ```sql -- Enable pgvector if not already enabled create extension if not exists vector; -- Create the documents table (if not already done) create table if not exists documents ( id bigserial primary key, content text, metadata jsonb, embedding vector(768) -- Make sure this matches your model's embedding dimension ); -- Create the match_documents function create or replace function match_documents ( query_embedding vector(768), match_count int default 5, filter jsonb default '{}' ) returns table ( id bigint, content text, metadata jsonb, similarity float ) language plpgsql as $$ begin return query select id, content, metadata, 1 - (embedding <=> query_embedding) as similarity from documents where metadata @> filter order by embedding <=> query_embedding limit match_count; end; $$; ``` 4. After running the above, execute this command to ensure Supabase’s API layer (PostgREST) refreshes its internal schema cache: ```sql NOTIFY pgrst, 'reload config'; ``` ## Step 3: Populate the Database To enable document retrieval, you need to populate the database with example entries: - Open and run the **test.ipynb** Jupyter notebook. - This script reads from the **metadata.jsonl** file and inserts the examples into the documents table. - This adds a Basic Retrieval capability to your agent, enhancing its performance. ## Step 4: Run the Evaluation Once the database is set up and filled with data: - Proceed to the Evaluation section in your project. - Run the evaluation script to test and score your agent’s performance.