Task
Created by calling an Environment. Holds configuration for running an evaluation.| Field | Type | Description |
|---|---|---|
env | Environment | dict | None | Source environment |
scenario | str | None | Scenario name to run |
args | dict[str, Any] | Script arguments |
trace_id | str | None | Trace identifier |
job_id | str | None | Parent job ID |
group_id | str | None | Group ID for parallel runs |
index | int | Index in parallel execution |
variants | dict[str, Any] | None | Variant assignment |
EvalContext
Returned byhud.eval(). Extends Environment with evaluation tracking.
| Property | Type | Description |
|---|---|---|
trace_id | str | Unique trace identifier |
eval_name | str | Evaluation name |
prompt | str | None | Task prompt |
variants | dict[str, Any] | Current variant assignment |
reward | float | None | Evaluation reward |
answer | str | None | Submitted answer |
error | BaseException | None | Error if failed |
results | list[EvalContext] | Results from parallel runs |
headers | dict[str, str] | Trace headers |
MCPToolCall
Represents a tool call to execute.| Field | Type | Description |
|---|---|---|
id | str | Unique identifier (auto-generated) |
name | str | Tool name |
arguments | dict[str, Any] | Tool arguments |
MCPToolResult
Result from executing a tool call.| Field | Type | Description |
|---|---|---|
content | list[ContentBlock] | Result content blocks |
structuredContent | dict | None | Structured result data |
isError | bool | Whether the call failed |
Trace
Returned byagent.run(). Contains the result of an agent execution.
| Field | Type | Description |
|---|---|---|
reward | float | Evaluation score (0.0-1.0) |
done | bool | Whether execution completed |
content | str | None | Final response content |
isError | bool | Whether an error occurred |
info | dict[str, Any] | Additional metadata |
trace | list[TraceStep] | Execution trace steps |
messages | list[Any] | Final conversation state |
AgentResponse
Returned by agentget_response() methods.
| Field | Type | Description |
|---|---|---|
tool_calls | list[MCPToolCall] | Tools to execute |
done | bool | Whether agent should stop |
content | str | None | Response text |
reasoning | str | None | Model reasoning/thinking |
info | dict[str, Any] | Provider-specific metadata |
isError | bool | Error flag |
AgentType
Enum of supported agent types.| Value | Agent Class |
|---|---|
AgentType.CLAUDE | ClaudeAgent |
AgentType.OPENAI | OpenAIAgent |
AgentType.OPERATOR | OperatorAgent |
AgentType.GEMINI | GeminiAgent |
AgentType.OPENAI_COMPATIBLE | OpenAIChatAgent |
ContentBlock
MCP content types (frommcp.types):
EvaluationResult
Returned by evaluation tools.| Field | Type | Description |
|---|---|---|
reward | float | Score (0.0-1.0) |
done | bool | Task complete |
content | str | None | Details |
info | dict | Metadata |
See Also
- Evals - hud.eval() reference
- Environments - Environment class
- Agents - Agent classes