Documentation Index
Fetch the complete documentation index at: https://docs.exulu.com/llms.txt
Use this file to discover all available pages before exploring further.
test_cases
Test cases for evaluating agent performance.
type test_cases {
id: ID!
name: String!
description: String
inputs: JSON! # UIMessage[]
expected_output: String!
expected_tools: JSON # string[]
expected_knowledge_sources: JSON # string[]
expected_agent_tools: JSON # string[]
createdAt: Date!
updatedAt: Date!
RBAC: RBACData
}
Example:
mutation {
test_casesCreateOne(
input: {
name: "Weather Query"
description: "User asks about weather"
inputs: [
{ role: "user", content: "What's the weather like?" }
]
expected_output: "Based on current data, it's 68°F and sunny."
expected_tools: ["get_weather"]
}
) {
item {
id
name
}
}
}
eval_sets
Collections of test cases for batch evaluation.
type eval_sets {
id: ID!
name: String!
description: String
test_case_ids: JSON # string[]
createdAt: Date!
updatedAt: Date!
RBAC: RBACData
}
eval_runs
Evaluation execution records.
type eval_runs {
id: ID!
name: String!
description: String
agent_id: ID!
test_case_ids: JSON # string[]
eval_functions: JSON # string[]
config: JSON
scoring_method: String
pass_threshold: Float
timeout_in_seconds: Int
createdAt: Date!
updatedAt: Date!
RBAC: RBACData
}
Evaluation Workflow
- Create test_cases with expected inputs and outputs
- Group test cases into eval_sets for organized testing
- Run eval_runs to evaluate agent performance
- Review results and iterate on agent improvements
Agent Types
Agents being evaluated
Job Types
Background eval execution
Back to Overview
View all core types