Skip to content

Evaluator

Evaluator module for orchestrating evaluation workflows.

This module provides evaluator classes that coordinate metrics and evaluation logic for different use cases. Evaluators handle the evaluation process, including data preparation, metric execution, and result aggregation.

Available evaluators: - BaseEvaluator: Abstract base class for all evaluators - BaseGenerationEvaluator: Shared generation evaluator base class - ClassicalRetrievalEvaluator: Traditional retrieval evaluation methods - LMBasedRetrievalEvaluator: LM-based retrieval evaluation methods - GEvalGenerationEvaluator: G-Eval based generation evaluation - RAGEvaluator: Combined retrieval and generation evaluation for RAG pipelines - QTEvaluator: Question-answering evaluation - CompositeEvaluator: Combines multiple metrics into a single evaluation unit - AgentEvaluator: Combined agent trajectory and generation evaluation

AgentEvaluator(tool_correctness_metric=None, generation_evaluator=None, trajectory_accuracy_metric=None, metrics_aggregator=None, num_judges=DefaultValues.NUM_JUDGES, aggregation_method=DefaultValues.AGGREGATION_METHOD, judge=None)

Bases: BaseEvaluator

Evaluator for agent tool calling and generation quality evaluation with MetricsAggregator.

This evaluator combines: 1. DeepEval Tool Correctness metric for agent tool call evaluation 2. GEval generation evaluator (completeness, groundedness, redundancy) 3. Optional LangChain Agent Trajectory Accuracy metric (disabled by default) 4. MetricsAggregator for polarity-aware binary scoring aggregation

Uses dependency injection pattern - accepts pre-configured metric and evaluator. Uses preprocess → aggregate → postprocess pattern consistent with other evaluators.

Expected input (LLMTestData with generation fields): - agent_trajectory (list[dict[str, Any]]): The agent trajectory - expected_agent_trajectory (list[dict[str, Any]]): Expected trajectory - input (str): The input query for generation evaluation - actual_output (str): The generated response - expected_output (str): The expected response - retrieved_context (str | list[str] | None, optional): The retrieved context

Aggregation Logic
  • Uses MetricsAggregator.aggregate() for base aggregation:
    • aggregate_success = all(metric.success) — AND-gate across all metrics
    • aggregate_score = mean(quality_score) — quality_score = score (higher_is_better=True) or 1 - score (higher_is_better=False)
  • Trajectory accuracy metric (if enabled) only runs when agent_trajectory is provided in input. Results are included in output but do not affect aggregation

Output Structure:

The output is a flat dictionary containing:

- Aggregated Results:
    - global_explanation (str): Human-readable explanation of the evaluation
    (added by parent evaluator)
    - aggregate_success (bool): AND-gate of all metric success values
    - aggregate_score (float): Mean quality score with polarity inversion

- Tool Call Evaluation Results:
    - deepeval_tool_correctness (dict): Tool correctness evaluation result containing
    score, success, threshold, and other contract fields

- Trajectory Accuracy Results (optional, only when metric is enabled AND agent_trajectory is provided):
    - langchain_agent_trajectory_accuracy (dict): Trajectory accuracy evaluation result

- Generation Evaluation Results (nested):
    - generation (dict): Generation evaluation results containing:
        - global_explanation (str): Explanation of generation quality
        - aggregate_success (bool): AND-gate of generation metric success values
        - aggregate_score (float): Mean quality score for generation metrics
        - Individual metric results (completeness, groundedness, redundancy,
        language_consistency, refusal_alignment) with full contract fields

Attributes:

Name Type Description
name str

The name of the evaluator.

tool_correctness_metric DeepEvalToolCorrectnessMetric

The metric for tool call assessment.

generation_evaluator GEvalGenerationEvaluator

Evaluator for generation quality assessment.

trajectory_accuracy_metric LangChainAgentTrajectoryAccuracyMetric | None

Optional metric for LangChain-style trajectory accuracy evaluation. Disabled by default.

metrics_aggregator MetricsAggregator

Aggregator for polarity-aware binary scoring.

Example

import os from gllm_evals.evaluator import AgentEvaluator from gllm_evals.metrics.tool_use.deepeval_tool_correctness import DeepEvalToolCorrectnessMetric from gllm_evals.metrics.tool_use.langchain_agent_trajectory_accuracy import ( ... LangChainAgentTrajectoryAccuracyMetric ... )

Configure metrics

tool_correctness = DeepEvalToolCorrectnessMetric() trajectory_accuracy = LangChainAgentTrajectoryAccuracyMetric()

Create evaluator (trajectory accuracy is optional and runs conditionally)

evaluator = AgentEvaluator( ... tool_correctness_metric=tool_correctness, ... trajectory_accuracy_metric=trajectory_accuracy, # Optional; only runs if agent_trajectory is in input ... )

result = await evaluator.evaluate(data)

Initialize the AgentEvaluator.

Parameters:

Name Type Description Default
tool_correctness_metric DeepEvalToolCorrectnessMetric | None

Pre-configured metric for trajectory evaluation. If None, a default metric will be created with model=DefaultValues.MODEL. Defaults to None.

None
generation_evaluator GEvalGenerationEvaluator | None

Pre-configured evaluator for generation quality assessment. If None, a default GEvalGenerationEvaluator will be created. Defaults to None.

None
trajectory_accuracy_metric LangChainAgentTrajectoryAccuracyMetric | None

Pre-configured metric for LangChain-style trajectory accuracy evaluation. Disabled by default (None). When provided, only runs when agent_trajectory field is present in input data. Results are included in output but do not affect final aggregated score. Defaults to None.

None
metrics_aggregator MetricsAggregator | None

Aggregator for polarity-aware binary scoring. If None, a default MetricsAggregator is used. Defaults to None.

None
num_judges int

Number of judges to use for metric-level aggregation in generation evaluator. Defaults to DefaultValues.NUM_JUDGES (1).

NUM_JUDGES
aggregation_method AggregationMethod

Method for aggregating judge scores at metric level in generation evaluator. Defaults to DefaultValues.AGGREGATION_METHOD.

AGGREGATION_METHOD
judge List[Dict[str, Any]] | None

Judge configuration for metric-level aggregation in generation evaluator. List of judge model configurations where each dict should contain 'model_id', 'model_credentials', and optionally 'model_config'. When provided, enables metric-level aggregation with heterogeneous models. Used only if generation_evaluator is None.

None
Note

When using default metric/evaluator (by not providing them), make sure the required environment variables are set: - OPENAI_API_KEY for tool correctness metric - GOOGLE_API_KEY for generation evaluator

required_fields property

Returns the required fields for the data.

Returns the combined set of required fields from both the tool correctness metric and generation evaluator.

Returns:

Type Description
set[str]

set[str]: The required fields for the data.

format_tool_violation(tool_score, tool_passed, tool_threshold)

Format Tool Correctness violation message.

Parameters:

Name Type Description Default
tool_score float

The tool score.

required
tool_passed bool

Whether the tool passed.

required
tool_threshold float

The tool threshold.

required

Returns:

Name Type Description
str str

The formatted tool violation message.

BaseEvaluator(name, batch_status_check_interval=DefaultValues.BATCH_STATUS_CHECK_INTERVAL, batch_max_iterations=DefaultValues.BATCH_MAX_ITERATIONS, metrics_aggregator=None)

Bases: ABC

Base class for all evaluators.

Attributes:

Name Type Description
name str

The name of the evaluator.

required_fields set[str]

The required fields for the evaluator.

input_type type | None

The type of the input data.

Initialize the evaluator.

Parameters:

Name Type Description Default
name str

The name of the evaluator.

required
batch_status_check_interval float

Time between batch status checks in seconds. Defaults to 30.0.

BATCH_STATUS_CHECK_INTERVAL
batch_max_iterations int

Maximum number of status check iterations before timeout. Defaults to 120 (60 minutes with default interval).

BATCH_MAX_ITERATIONS
metrics_aggregator MetricsAggregator | None

Aggregator for polarity-aware binary scoring. If None, a default MetricsAggregator is used. Defaults to None.

None

Raises:

Type Description
ValueError

If batch_status_check_interval or batch_max_iterations are not positive.

aggregate_required_fields(metrics, mode='any') staticmethod

Aggregate required fields from multiple metrics.

Parameters:

Name Type Description Default
metrics Iterable[BaseMetric]

The metrics to aggregate from.

required
mode str

The aggregation mode. Options: - "union": All fields required by any metric - "intersection": Only fields required by all metrics - "any": Empty set (no validation) Defaults to "any".

'any'

Returns:

Type Description
set[str]

set[str]: The aggregated required fields.

Raises:

Type Description
ValueError

If mode is not one of the supported options.

can_evaluate_any(metrics, data) staticmethod

Check if any of the metrics can evaluate the given data.

Parameters:

Name Type Description Default
metrics Iterable[BaseMetric]

The metrics to check.

required
data MetricInput

The data to validate against.

required

Returns:

Name Type Description
bool bool

True if any metric can evaluate the data, False otherwise.

ensure_list_of_dicts(data, key) staticmethod

Ensure that a field in the data is a list of dictionaries.

Parameters:

Name Type Description Default
data MetricInput

The data to validate.

required
key str

The key to check.

required

Raises:

Type Description
ValueError

If the field is not a list or contains non-dictionary elements.

ensure_non_empty_list(data, key) staticmethod

Ensure that a field in the data is a non-empty list.

Parameters:

Name Type Description Default
data MetricInput

The data to validate.

required
key str

The key to check.

required

Raises:

Type Description
ValueError

If the field is not a list or is empty.

evaluate(data) async

Evaluate the data (single item or batch).

Parameters:

Name Type Description Default
data EvalInput | list[EvalInput]

The data to be evaluated. Can be a single item or a list for batch processing.

required

Returns:

Type Description
EvaluationOutput | list[EvaluationOutput]

EvaluationOutput | list[EvaluationOutput]: The evaluation output with global_explanation. Returns a list if input is a list.

get_input_fields() classmethod

Return declared input field names if input_type is provided; otherwise None.

Returns:

Type Description
list[str] | None

list[str] | None: The input fields.

get_input_spec() classmethod

Return structured spec for input fields if input_type is provided; otherwise None.

Returns:

Type Description
list[dict[str, Any]] | None

list[dict[str, Any]] | None: The input spec.

BaseGenerationEvaluator(metrics=None, enabled_metrics=None, model=DefaultValues.MODEL, model_credentials=None, model_config=None, num_judges=DefaultValues.NUM_JUDGES, aggregation_method=None, max_concurrent_judges=None, run_parallel=True, judge=None, refusal_metric=None, batch_status_check_interval=DefaultValues.BATCH_STATUS_CHECK_INTERVAL, batch_max_iterations=DefaultValues.BATCH_MAX_ITERATIONS, metrics_aggregator=None)

Bases: BaseEvaluator

Shared base evaluator for generation-style rule-engine evaluation.

Default expected input
  • input (str): The input provided to the AI system or component (e.g., a query, prompt, or instruction).
  • retrieved_context (str): Supporting context used during generation (e.g., retrieved documents).
  • expected_output (str): The reference output used for comparison.
  • actual_output (str): The output generated by the AI system or component to evaluate.

Attributes:

Name Type Description
name str

The name of the evaluator.

metrics List[BaseMetric]

The list of metrics to evaluate.

run_parallel bool

Whether to run the metrics in parallel.

judge List[Dict[str, Any]] | None

Optional list of judge configurations for metric-level aggregation.

metrics_aggregator MetricsAggregator

The aggregator for polarity-aware binary scoring.

Initialize the base generation evaluator.

Parameters:

Name Type Description Default
metrics List[BaseMetric] | None

A list of metric instances to use as a base pool. If None, defaults to a built-in GEval-backed metric pool. Each custom metric must generate a score key in the output.

None
enabled_metrics List[type[BaseMetric] | str] | None

A list of metric classes or names to enable from the pool. If None, all metrics in the pool are used.

None
model str | ModelId | BaseLMInvoker

The model to use for the metrics when judge is not provided.

MODEL
model_config dict[str, Any] | None

The model config to use for the metrics.

None
model_credentials str | None

The model credentials, used for initializing default metrics. Defaults to None. This is required when default metrics are initialized with a string model.

None
num_judges int

The number of judges to use for each metric. Defaults to DefaultValues.NUM_JUDGES.

NUM_JUDGES
aggregation_method AggregationSelector | None

The aggregation method to use for each metric. If None, each metric uses its own default (MAJORITY_VOTE for GEval metrics).

None
max_concurrent_judges int | None

The maximum number of concurrent judges per metric. If None, each metric uses its own default.

None
run_parallel bool

Whether to run the metrics in parallel. Defaults to True.

True
judge List[Dict[str, Any]] | None

Judge configuration for metric-level aggregation. List of judge model configurations where each dict should contain: - 'model_id' (str): Model identifier (e.g., 'openai/gpt-4o', 'anthropic/claude-3-5-sonnet') - 'model_credentials' (str): API credentials for the model - 'model_config' (Dict[str, Any], optional): Additional model configuration When provided, enables metric-level aggregation with heterogeneous models. None: Uses single judge or num_judges parameter for same-model judges.

None
refusal_metric GEvalRefusalMetric | None

The refusal metric to use. If None, the default refusal metric will be used. Defaults to GEvalRefusalMetric.

None
batch_status_check_interval float

Time between batch status checks in seconds. Defaults to 30.0.

BATCH_STATUS_CHECK_INTERVAL
batch_max_iterations int

Maximum number of status check iterations before timeout. Defaults to 120 (60 minutes with default interval).

BATCH_MAX_ITERATIONS
metrics_aggregator MetricsAggregator | None

The aggregator for polarity-aware binary scoring. If None, a default MetricsAggregator is used. Defaults to None.

None

Raises:

Type Description
ValueError

If model_credentials is not provided when initializing default metrics with a string model.

ClassicalRetrievalEvaluator(metrics=None, k=20, metrics_aggregator=None)

Bases: BaseEvaluator

A class that evaluates the performance of a classical retrieval system.

Required fields: - retrieved_chunks: The retrieved chunks with their similarity score. - ground_truth_chunk_ids: The ground truth chunk ids.

Example:

data = RetrievalData(
    retrieved_chunks={
        "chunk1": 0.9,
        "chunk2": 0.8,
        "chunk3": 0.7,
    },
    ground_truth_chunk_ids=["chunk1", "chunk2", "chunk3"],
)

evaluator = ClassicalRetrievalEvaluator()
await evaluator.evaluate(data)

Attributes:

Name Type Description
name str

The name of the evaluator.

metrics list[str | ClassicalRetrievalMetric] | None

The metrics to evaluate.

k int

The number of retrieved chunks to consider.

Initializes the evaluator.

Parameters:

Name Type Description Default
metrics list[str | ClassicalRetrievalMetric] | None

The metrics to evaluate. Defaults to all metrics.

None
k int | list[int]

The number of retrieved chunks to consider. Defaults to 20.

20
metrics_aggregator MetricsAggregator | None

Aggregator for polarity-aware binary scoring. If None, a default MetricsAggregator is used. Defaults to None.

None

required_fields property

Returns the required fields for the data.

Returns:

Type Description
set[str]

set[str]: The required fields for the data.

CompositeEvaluator(metrics, name='composite', parallel=True, metrics_aggregator=None)

Bases: BaseEvaluator

Composite evaluator that runs multiple metrics and aggregates results.

This evaluator composes multiple BaseMetric objects and presents them as a single evaluator unit, following the GoF Composite pattern. It supports parallel or sequential execution, fault isolation, and custom aggregation.

Attributes:

Name Type Description
metrics list[BaseMetric]

The list of metrics to evaluate.

name str

The name of the evaluator.

parallel bool

Whether to evaluate the metrics in parallel.

metrics_aggregator MetricsAggregator

The aggregator for polarity-aware binary scoring.

Initialize the composite evaluator.

Parameters:

Name Type Description Default
metrics list[BaseMetric]

The list of metrics to evaluate.

required
name str

The name of the evaluator.

'composite'
parallel bool

Whether to evaluate the metrics in parallel.

True
metrics_aggregator MetricsAggregator | None

Aggregator for polarity-aware binary scoring. If None, a default MetricsAggregator is used. Defaults to None.

None

GEvalGenerationEvaluator(metrics=None, enabled_metrics=None, model=DefaultValues.MODEL, model_credentials=None, model_config=None, num_judges=DefaultValues.NUM_JUDGES, aggregation_method=None, max_concurrent_judges=None, run_parallel=True, judge=None, refusal_metric=None, batch_status_check_interval=DefaultValues.BATCH_STATUS_CHECK_INTERVAL, batch_max_iterations=DefaultValues.BATCH_MAX_ITERATIONS, metrics_aggregator=None)

Bases: BaseGenerationEvaluator

GEval Generation Evaluator.

This evaluator is used to evaluate the generation of the model.

Default expected input
  • input (str): The input provided to the AI system or component (e.g., a query, prompt, or instruction).
  • retrieved_context (str): Supporting context used during generation (e.g., retrieved documents).
  • expected_output (str): The reference output used for comparison.
  • actual_output (str): The output generated by the AI system or component to evaluate.

Attributes:

Name Type Description
name str

The name of the evaluator.

metrics List[BaseMetric]

The list of metrics to evaluate.

run_parallel bool

Whether to run the metrics in parallel.

metrics_aggregator MetricsAggregator

The aggregator for polarity-aware binary scoring.

Initialize the GEval Generation Evaluator.

Parameters:

Name Type Description Default
metrics List[BaseMetric] | None

The list of metrics to evaluate.

None
enabled_metrics List[type[BaseMetric] | str] | None

The list of enabled metrics.

None
model str | ModelId | BaseLMInvoker

The model to use for the metrics.

MODEL
model_credentials str | None

The model credentials to use for the metrics.

None
model_config dict[str, Any] | None

The model config to use for the metrics.

None
num_judges int

The number of judges to use for the metric. Defaults to 1.

NUM_JUDGES
aggregation_method AggregationSelector | None

The aggregation method to use for each metric. If None, each metric uses its own default (MAJORITY_VOTE for GEval metrics).

None
max_concurrent_judges int | None

The maximum number of concurrent judges per metric. If None, each metric uses its own default.

None
run_parallel bool

Whether to run the metrics in parallel.

True
judge List[Dict[str, Any]] | None

Judge configuration for metric-level aggregation. List of judge model configs with different models for heterogeneous judges.

None
refusal_metric GEvalRefusalMetric | None

The refusal metric to use. If None, the default refusal metric will be used. Defaults to GEvalRefusalMetric.

None
batch_status_check_interval float

Time between batch status checks in seconds. Defaults to 30.0.

BATCH_STATUS_CHECK_INTERVAL
batch_max_iterations int

Maximum number of status check iterations before timeout. Defaults to 120 (60 minutes with default interval).

BATCH_MAX_ITERATIONS
metrics_aggregator MetricsAggregator | None

The aggregator for polarity-aware binary scoring. If None, a default MetricsAggregator is used. Defaults to None.

None

LMBasedRetrievalEvaluator(metrics=None, enabled_metrics=None, model=DefaultValues.MODEL, model_credentials=None, model_config=None, run_parallel=True, metrics_aggregator=None)

Bases: BaseEvaluator

Evaluator for LM-based retrieval quality in RAG pipelines.

This evaluator
  • Runs a configurable set of retrieval metrics (by default: DeepEval contextual precision and contextual recall)
  • Combines their scores using a simple rule-based scheme to produce:
    • relevancy_rating (good / bad / incomplete)
    • score (aggregated retrieval score)
Default expected input
  • input (str): The input query to evaluate the metric.
  • expected_output (str): The expected output to evaluate the metric.
  • retrieved_context (str | list[str]): The list of retrieved contexts to evaluate the metric. If the retrieved context is a str, it will be converted into a list with a single element.

Attributes:

Name Type Description
name str

The name of the evaluator.

metrics list[BaseMetric]

The list of metrics to evaluate.

enabled_metrics Sequence[type[BaseMetric] | str] | None

The list of metrics to enable.

model str | ModelId | BaseLMInvoker

The model to use for the metrics.

model_credentials str | None

The model credentials to use for the metrics.

model_config dict[str, Any] | None

The model configuration to use for the metrics.

run_parallel bool

Whether to run the metrics in parallel.

Initialize the LM-based retrieval evaluator.

Parameters:

Name Type Description Default
metrics Sequence[BaseMetric] | None

Optional custom retrieval metric instances. If provided, these will be used as the base pool and may override the default metrics by name. Defaults to None.

None
enabled_metrics Sequence[type[BaseMetric] | str] | None

Optional subset of metrics to enable from the metric pool. Each entry can be either a metric class or its name. If None, all metrics from the pool are used. Defaults to None.

None
model str | ModelId | BaseLMInvoker

Model for the default DeepEval metrics. Defaults to DefaultValues.MODEL.

MODEL
model_credentials str | None

Credentials for the model, required when model is a string. Defaults to None.

None
model_config dict[str, Any] | None

Optional model configuration. Defaults to None.

None
run_parallel bool

Whether to run retrieval metrics in parallel. Defaults to True.

True
metrics_aggregator MetricsAggregator | None

Aggregator for polarity-aware binary scoring. If None, a default MetricsAggregator is used. Defaults to None.

None

required_fields property

Return the union of required fields from all configured metrics.

QTEvaluator(completeness_metric=None, groundedness_metric=None, redundancy_metric=None, model=DefaultValues.MODEL, model_config=None, model_credentials=None, num_judges=DefaultValues.NUM_JUDGES, aggregation_method=None, max_concurrent_judges=None, run_parallel=True, judge=None, score_mapping=None, score_weights=None, batch_status_check_interval=DefaultValues.BATCH_STATUS_CHECK_INTERVAL, batch_max_iterations=DefaultValues.BATCH_MAX_ITERATIONS, metrics_aggregator=None)

Bases: BaseEvaluator

Evaluator for query transformation tasks.

Default expected input: - input (str): The query to evaluate the completeness of the model's output. - expected_output (str): The expected response to evaluate the completeness of the model's output. - actual_output (str): The generated response to evaluate the completeness of the model's output.

Attributes:

Name Type Description
completeness_metric GEvalCompletenessMetric

The completeness metric.

hallucination_metric GEvalGroundednessMetric

The groundedness metric.

redundancy_metric GEvalRedundancyMetric

The redundancy metric.

run_parallel bool

Whether to run the metrics in parallel.

score_mapping dict[str, dict[int, float]]

The score mapping.

score_weights dict[str, float]

The score weights.

Initialize the QTEvaluator.

Parameters:

Name Type Description Default
completeness_metric GEvalCompletenessMetric | None

The completeness metric. Defaults to built-in GEvalCompletenessMetric.

None
groundedness_metric GEvalGroundednessMetric | None

The groundedness metric. Defaults to built-in GEvalGroundednessMetric.

None
redundancy_metric GEvalRedundancyMetric | None

The redundancy metric. Defaults to built-in GEvalRedundancyMetric.

None
model str | ModelId | BaseLMInvoker

The model to use for the metrics when judge is not provided.

MODEL
model_config dict[str, Any] | None

The model config to use for the metrics.

None
model_credentials str | None

The model credentials. Defaults to None. This is required if some of the default metrics are used.

None
num_judges int

The number of judges to use for each metric. Defaults to DefaultValues.NUM_JUDGES.

NUM_JUDGES
aggregation_method AggregationSelector | None

The aggregation method to use for each metric. If None, each metric uses its own default (MAJORITY_VOTE for GEval metrics).

None
max_concurrent_judges int | None

The maximum number of concurrent judges per metric. If None, each metric uses its own default.

None
run_parallel bool

Whether to run the metrics in parallel. Defaults to True.

True
judge List[Dict[str, Any]] | None

Judge configuration for metric-level aggregation. List of judge model configurations where each dict should contain: - 'model_id' (str): Model identifier - 'model_credentials' (str): API credentials - 'model_config' (Dict[str, Any], optional): Additional configuration When provided, enables metric-level aggregation with heterogeneous models.

None
score_mapping dict[str, dict[int, float]] | None

The score mapping. Defaults to None. This is required if some of the default metrics are used.

None
score_weights dict[str, float] | None

The score weights. Defaults to None.

None
batch_status_check_interval float

Time between batch status checks in seconds. Defaults to 30.0.

BATCH_STATUS_CHECK_INTERVAL
batch_max_iterations int

Maximum number of status check iterations before timeout. Defaults to 120 (60 minutes with default interval).

BATCH_MAX_ITERATIONS
metrics_aggregator MetricsAggregator | None

Aggregator for polarity-aware binary scoring. If None, a default MetricsAggregator is used. Defaults to None.

None

Raises:

Type Description
ValueError

If model_credentials is not provided when using default metrics.

required_fields property

Returns the required fields for the data.

Returns:

Type Description
set[str]

set[str]: The required fields for the data.

RAGEvaluator(retrieval_evaluator=None, generation_evaluator=None, retrieval_metrics=None, generation_metrics=None, model=DefaultValues.MODEL, model_credentials=None, model_config=None, num_judges=DefaultValues.NUM_JUDGES, aggregation_method=DefaultValues.AGGREGATION_METHOD, run_parallel=True, judge=None, refusal_metric=None, metrics_aggregator=None)

Bases: BaseEvaluator

Evaluator for RAG pipelines combining retrieval and generation evaluation.

This evaluator
  • Runs retrieval evaluation using LMBasedRetrievalEvaluator
  • Runs generation evaluation using GEvalGenerationEvaluator
  • Combines their scores using MetricsAggregator to produce:
    • aggregate_success (bool)
    • aggregate_score (float)
Default expected input
  • input (str): The input query to evaluate.
  • expected_output (str): The expected output.
  • retrieved_context (str | list[str]): The retrieved contexts.
  • actual_output (str): The generated response.

Attributes:

Name Type Description
name str

The name of the evaluator.

retrieval_evaluator LMBasedRetrievalEvaluator

The retrieval evaluator.

generation_evaluator GEvalGenerationEvaluator

The generation evaluator.

run_parallel bool

Whether to run retrieval and generation evaluations in parallel.

Initialize the RAG evaluator.

Parameters:

Name Type Description Default
retrieval_evaluator LMBasedRetrievalEvaluator | None

Pre-configured retrieval evaluator. If provided, this will be used directly and retrieval_* parameters will be ignored. Defaults to None.

None
generation_evaluator GEvalGenerationEvaluator | None

Pre-configured generation evaluator. If provided, this will be used directly and generation_* parameters will be ignored. Defaults to None.

None
retrieval_metrics Sequence[BaseMetric] | None

Optional custom retrieval metric instances. Used only if retrieval_evaluator is None. If provided, these will be used as the base pool and may override the default metrics by name. Defaults to None.

None
generation_metrics Sequence[BaseMetric] | None

Optional custom generation metric instances. Used only if generation_evaluator is None. If provided, these will be used as the base pool and may override the default metrics by name. Defaults to None.

None
model str | ModelId | BaseLMInvoker

Model for the default metrics. Used only if evaluators are None. Defaults to DefaultValues.MODEL.

MODEL
model_credentials str | None

Credentials for the model, required when model is a string. Used only if evaluators are None. Defaults to None.

None
model_config dict[str, Any] | None

Optional model configuration. Used only if evaluators are None. Defaults to None.

None
num_judges int

Number of judges to use for metric-level aggregation in generation evaluator. Defaults to 1 (single judge). When > 1, each metric runs multiple judges with the same model and aggregates their scores. Used only if generation_evaluator is None.

NUM_JUDGES
aggregation_method AggregationMethod

Method for aggregating judge scores at metric level in generation evaluator. Defaults to DefaultValues.AGGREGATION_METHOD (MAJORITY_VOTE). Used only if generation_evaluator is None.

AGGREGATION_METHOD
run_parallel bool

Whether to run retrieval and generation evaluations in parallel. Used only if evaluators are None. Defaults to True.

True
judge List[Dict[str, Any]] | None

Judge configuration for metric-level aggregation in generation evaluator. List of judge model configurations where each dict should contain: - 'model_id' (str): Model identifier - 'model_credentials' (str): API credentials - 'model_config' (Dict[str, Any], optional): Additional configuration When provided, enables metric-level aggregation with heterogeneous models. Used only if generation_evaluator is None.

None
refusal_metric type[BaseMetric] | None

The refusal metric to use for generation evaluator. Used only if generation_evaluator is None. If None, the default refusal metric will be used. Defaults to None.

None
metrics_aggregator MetricsAggregator | None

Aggregator for polarity-aware binary scoring. If None, a default MetricsAggregator is created. Defaults to None.

None

required_fields property

Return the union of required fields from both evaluators.

SummarizationEvaluator(metrics=None, enabled_metrics=None, model=DefaultValues.MODEL, model_credentials=None, model_config=None, num_judges=DefaultValues.NUM_JUDGES, aggregation_method=DefaultValues.AGGREGATION_METHOD, run_parallel=True, judge=None, batch_status_check_interval=DefaultValues.BATCH_STATUS_CHECK_INTERVAL, batch_max_iterations=DefaultValues.BATCH_MAX_ITERATIONS, metrics_aggregator=None)

Bases: BaseGenerationEvaluator

Evaluator for summarization quality using four GEval-style metrics.

Default expected input
  • input (str): Source text or transcript.
  • actual_output (str): Generated summary.

Initialize the SummarizationEvaluator.

Parameters:

Name Type Description Default
metrics list[BaseMetric] | None

A list of metric instances to use as a base pool. If None, default summary metrics are used.

None
enabled_metrics list[type[BaseMetric] | str] | None

A list of metric classes or names to enable from the pool. If None, all default summary metrics from _SUMMARY_METRIC_CONFIGS are enabled.

None
model str | ModelId | BaseLMInvoker

The model to use for the metrics when judge is empty.

MODEL
model_credentials str | None

The model credentials used for metric initialization.

None
model_config dict[str, Any] | None

The model config used for metric initialization.

None
num_judges int

Number of judges to use for metric-level aggregation. Defaults to 1 (single judge). When > 1, each metric runs multiple judges with the same model and aggregates their scores.

NUM_JUDGES
aggregation_method AggregationMethod

Method for aggregating judge scores at metric level. Defaults to DefaultValues.AGGREGATION_METHOD (MAJORITY_VOTE).

AGGREGATION_METHOD
run_parallel bool

Whether to run the metrics in parallel. Defaults to True.

True
judge List[Dict[str, Any]] | None

Judge configuration for metric-level aggregation. List of judge model configurations where each dict should contain: - 'model_id' (str): Model identifier - 'model_credentials' (str): API credentials - 'model_config' (Dict[str, Any], optional): Additional configuration When provided, enables metric-level aggregation with heterogeneous models.

None
batch_status_check_interval float

Time between batch status checks in seconds.

BATCH_STATUS_CHECK_INTERVAL
batch_max_iterations int

Maximum number of status check iterations before timeout.

BATCH_MAX_ITERATIONS
metrics_aggregator MetricsAggregator | None

Aggregator for polarity-aware binary scoring. If None, a default MetricsAggregator is used. Defaults to None.

None

Returns:

Type Description

None