Endpoints & Methods
Starting a dynamic evaluation run
- Gateway Endpoint:
POST /dynamic_evaluation_run - Client Method:
dynamic_evaluation_run - Parameters:
variants: an object (dictionary) mapping function names to variant namesproject_name(string, optional): the name of the project to associate the run withdisplay_name(string, optional): the display (human-readable) name of the runtags(dictionary, optional): a dictionary of key-value pairs to tag the run’s inferences with
- Returns:
run_id(UUID): the ID of the run
Starting an episode in a dynamic evaluation run
- Gateway Endpoint:
POST /dynamic_evaluation_run/{run_id}/episode - Client Method:
dynamic_evaluation_run_episode - Parameters:
run_id(UUID): the ID of the run generated by thedynamic_evaluation_runmethodtask_name(string, optional): the name of the task to associate the episode withtags(dictionary, optional): a dictionary of key-value pairs to tag the episode’s inferences with
- Returns:
episode_id(UUID): the ID of the episode
Making inference and feedback calls during a dynamic evaluation run
After initializing a run and an episode, you can make inference and feedback API calls like you normally would. By providing the specialepisode_id parameter generated by the dynamic_evaluation_run_episode method , the TensorZero Gateway will associate the inference and feedback with the evaluation run, handle variant pinning, and more.