Endpoints & Methods
Starting a dynamic evaluation run
- Gateway Endpoint:
POST /dynamic_evaluation_run
- Client Method:
dynamic_evaluation_run
- Parameters:
variants
: an object (dictionary) mapping function names to variant namesproject_name
(string, optional): the name of the project to associate the run withdisplay_name
(string, optional): the display (human-readable) name of the runtags
(dictionary, optional): a dictionary of key-value pairs to tag the run’s inferences with
- Returns:
run_id
(UUID): the ID of the run
Starting an episode in a dynamic evaluation run
- Gateway Endpoint:
POST /dynamic_evaluation_run/{run_id}/episode
- Client Method:
dynamic_evaluation_run_episode
- Parameters:
run_id
(UUID): the ID of the run generated by thedynamic_evaluation_run
methodtask_name
(string, optional): the name of the task to associate the episode withtags
(dictionary, optional): a dictionary of key-value pairs to tag the episode’s inferences with
- Returns:
episode_id
(UUID): the ID of the episode
Making inference and feedback calls during a dynamic evaluation run
After initializing a run and an episode, you can make inference and feedback API calls like you normally would. By providing the specialepisode_id
parameter generated by the dynamic_evaluation_run_episode
method , the TensorZero Gateway will associate the inference and feedback with the evaluation run, handle variant pinning, and more.