Skip to content

Evaluators

Evaluators automatically validate prompt Completions against a set of user-defined rules, e.g. if whether or not the completion starts with a certain phrase, has a length in a specified range, matches the desired format, etc.

Evaluators can be created and configured under the Evaluators tab on the Prompts page.

Evaluator list

Evaluator types

Length Evaluators

Length evaluators validate the length of prompt completions, either to an exact value or to a range (min/max). The unit for the evaluation can be tokens, characters, or words, depending on preference and use case.

Length Evaluator

Text Evaluators

Text evaluators validate the exact match or presence/absence of a specific text string at the beginning, end, or anywhere in the prompt completion. Users can also specify whether the evaluation should be case-sensitive or not.

Text Evaluator

Regex Evaluator

🚧 Coming soon...

Boolean Evaluator

🚧 Coming soon...

Number Evaluator

🚧 Coming soon...

JSON Evaluator

🚧 Coming soon...

Format Evaluator

🚧 Coming soon...

LLM Evaluator

🚧 Coming soon...

Time Evaluator

🚧 Coming soon...

External Evaluator

🚧 Coming soon...

Evaluator versioning (history)

Evaluators automatically keep track of their configuration history, so that it is always fully traceable which evaluators were used with which configuration for each completion.

The version history of an evaluator automatically increments if the configuration is changed and the evaluator has been executed on at least one completion. If the evaluator has never been executed, updating the configuration will not lead to the creation of a new version.