Ingest documents into our ChromaDB, run questions through your endpoint and our baseline, then inspect retrieval overlap (raw + path-aligned), embedding answer similarity, grounding, and scores.
Documents → our ChromaDB (directory or zip). SQL → optional test database for replaying endpoint-disclosed queries against row payloads (fields below submit with Run suite).
Or upload a .zip:
Test DB URL (SQLAlchemy). When Run suite runs with SQL replay enabled, the bench uses paths below to read SELECT/WITH + row list from each JSON response, re-executes the SQL with :binds, and compares to the payload. Defaults meta.sql / data.rows.
Send a probe to your URL or paste one sample JSON response. The bench shows the sample, a heuristic suggestion, and an extraction preview. Pick a preset and apply it to the Run suite form below (or override manually).
Endpoint: POST { "query", "top_k" }. Response mapping uses response preset below (default rag_legacy: answer/response + top_k_docs/sources). Embedding model for answer similarity: