Skip to content

API Connector

The API Connector is a 4-step wizard for running evaluations against any LLM API — without writing Python code.

Navigate to API Connector from the top navigation bar.

API Connector

Workflow

The wizard guides you through four steps:

graph LR
    A[1. Settings] --> B[2. Dataset]
    B --> C[3. Test]
    C --> D[4. Execute]
Step Description
Settings Configure API endpoint, response mapping, column mapping, metrics, and providers
Dataset Upload test data (CSV, JSON, JSONL)
Test Verify the full pipeline with a single request
Execute Review configuration and run the evaluation

Settings Tabs

Step 1 contains five configuration tabs:

Tab Description
Connection API endpoint, headers, body template
Response Map API response fields to evaluation variables
Columns Map dataset columns to test case fields
Metrics Select evaluation metrics (RAG, Agent, Security)
Settings LLM providers, eval model, timeouts, cost tracking

Save & Load Projects

You can save your entire configuration as a named project using Save Project, and reload it later from the Load project dropdown. This is useful for:

  • Running the same evaluation against updated datasets
  • Sharing configurations across team members
  • A/B testing different API endpoints with the same metrics