Evaluations

EvaluationsResource

Methods

create() ->
post/v5/evaluations

Create Evaluation

Parameters
data: [Dict[str, ]]

Items to be evaluated

name: str
description: Optional[str]
metadata: Optional[Dict[str, ]]

Optional metadata key-value pairs for the evaluation

tags: Optional[[str]]

The tags associated with the entity

tasks: Optional[[]]

Tasks allow you to augment and evaluate your data

class ChatCompletionEvaluationTask: ...
class GenericInferenceEvaluationTask: ...
class ApplicationVariantV1EvaluationTask: ...
class MetricEvaluationTask: ...
class AutoEvaluationQuestionTask: ...
class AutoEvaluationGuidedDecodingEvaluationTask: ...
class ContributorEvaluationQuestionTask: ...
Returns
id: str
created_at:
(format: date-time)
created_by:

The identity that created the entity.

datasets: List[]
name: str
status: Literal["failed", "completed", "running"]
tags: List[str]

The tags associated with the entity

archived_at: Optional[datetime]
(format: date-time)
description: Optional[str]
metadata: Optional[Dict[str, ]]

Metadata key-value pairs for the evaluation

object: Optional[Literal["evaluation"]]
(default: "evaluation")
tasks: Optional[List[]]

Tasks executed during evaluation. Populated with optional task view.

Request example
200Example
delete() ->
delete/v5/evaluations/{evaluation_id}

Archive Evaluation

list() -> SyncCursorPage[]
get/v5/evaluations

List Evaluations

retrieve(, ) ->
get/v5/evaluations/{evaluation_id}

Get Evaluation

update(, ) ->
patch/v5/evaluations/{evaluation_id}

Update Evaluation

Domain types

class Evaluation: ...
EvaluationTask =
ItemLocator = str
ItemLocatorTemplate = str