End-to-end example
Core primitives
Datasetstores rows and backend version metadata.@ze.task(...)defines the outputs generated for each row.run = dataset.eval(...)executes the task and returns anEvalobject.run.score(...)applies evaluation functions and writes aggregate metrics intorun.metrics.
Inspect the returned run
What happened
Task outputs were generated
dataset.eval(...) executed solve over each row with the requested worker
count and returned a run object (Eval).Runtime signals were captured
ze.emit_signal(...) attached execution facts to the task trace so they can
be inspected later without turning them into scores immediately.One mental model for scoring
- A row evaluation receives one row plus any mapped scalar values.
- A column evaluation receives lists of values across all rows.
- A run evaluation receives
all_runsafter you repeat a run.