pyrit.score.TrueFalseScorer#
- class TrueFalseScorer(*, validator: ~pyrit.score.scorer_prompt_validator.ScorerPromptValidator, score_aggregator: ~typing.Callable[[~typing.Iterable[~pyrit.models.score.Score]], ~pyrit.score.score_aggregator_result.ScoreAggregatorResult] = <function _create_aggregator.<locals>.aggregator>)[source]#
Bases:
ScorerBase class for scorers that return true/false binary scores.
This scorer evaluates prompt responses and returns a single boolean score indicating whether the response meets a specific criterion. Multiple pieces in a request response are aggregated using a TrueFalseAggregatorFunc function (default: TrueFalseScoreAggregator.OR).
- __init__(*, validator: ~pyrit.score.scorer_prompt_validator.ScorerPromptValidator, score_aggregator: ~typing.Callable[[~typing.Iterable[~pyrit.models.score.Score]], ~pyrit.score.score_aggregator_result.ScoreAggregatorResult] = <function _create_aggregator.<locals>.aggregator>) None[source]#
Initialize the TrueFalseScorer.
- Parameters:
validator (ScorerPromptValidator) – Custom validator.
score_aggregator (TrueFalseAggregatorFunc) – The aggregator function to use. Defaults to TrueFalseScoreAggregator.OR.
Methods
__init__(*, validator[, score_aggregator])Initialize the TrueFalseScorer.
evaluate_async([file_mapping, ...])Evaluate this scorer against human-labeled datasets.
get_identifier()Get an identifier dictionary for the scorer for database storage.
Get evaluation metrics for this scorer from the configured evaluation result file.
scale_value_float(value, min_value, max_value)Scales a value from 0 to 1 based on the given min and max values.
score_async(message, *[, objective, ...])Score the message, add the results to the database, and return a list of Score objects.
score_image_async(image_path, *[, objective])Score the given image using the chat target.
score_image_batch_async(*, image_paths[, ...])Score a batch of images asynchronously.
score_prompts_batch_async(*, messages[, ...])Score multiple prompts in batches using the provided objectives.
score_response_async(*, response[, ...])Score a response using an objective scorer and optional auxiliary scorers.
score_response_multiple_scorers_async(*, ...)Score a response using multiple scorers in parallel.
score_text_async(text, *[, objective])Scores the given text based on the task using the chat target.
validate_return_scores(scores)Validate the scores returned by the scorer.
Attributes
scorer_identifierGet the scorer identifier.
scorer_typeGet the scorer type based on class hierarchy.
- evaluation_file_mapping: ScorerEvalDatasetFiles | None = None#
- get_scorer_metrics() ObjectiveScorerMetrics | None[source]#
Get evaluation metrics for this scorer from the configured evaluation result file.
- Returns:
The metrics for this scorer, or None if not found or not configured.
- Return type:
- validate_return_scores(scores: list[Score]) None[source]#
Validate the scores returned by the scorer.
- Parameters:
- Raises:
ValueError – If the number of scores is not exactly one.
ValueError – If the score value is not “true” or “false”.