| | |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | from typing import Any, Callable, Dict, List, Literal, Optional, TYPE_CHECKING, Union |
| | import uuid |
| |
|
| | from google.api_core import exceptions |
| | import vertexai |
| | from google.cloud.aiplatform import base |
| | from google.cloud.aiplatform.metadata import metadata |
| | from vertexai import generative_models |
| | from vertexai.evaluation import _base as eval_base |
| | from vertexai.evaluation import _evaluation |
| | from vertexai.evaluation import constants |
| | from vertexai.evaluation import utils |
| | from vertexai.evaluation.metrics import ( |
| | _base as metrics_base, |
| | ) |
| | from vertexai.evaluation.metrics import ( |
| | pairwise_metric, |
| | ) |
| | from vertexai.evaluation.metrics import ( |
| | pointwise_metric, |
| | ) |
| | import numpy as np |
| |
|
| | if TYPE_CHECKING: |
| | import pandas as pd |
| |
|
| | |
| | try: |
| | from IPython import display as IPython_display |
| | except ImportError: |
| | IPython_display = None |
| |
|
| | _LOGGER = base.Logger(__name__) |
| |
|
| | EvalResult = eval_base.EvalResult |
| | GenerativeModel = generative_models.GenerativeModel |
| |
|
| |
|
| | class EvalTask: |
| | """A class representing an EvalTask. |
| | |
| | An Evaluation Tasks is defined to measure the model's ability to perform a |
| | certain task in response to specific prompts or inputs. Evaluation tasks must |
| | contain an evaluation dataset, and a list of metrics to evaluate. Evaluation |
| | tasks help developers compare propmpt templates, track experiments, compare |
| | models and their settings, and assess the quality of the model's generated |
| | text. |
| | |
| | Dataset Details: |
| | |
| | Default dataset column names: |
| | * prompt_column_name: "prompt" |
| | * reference_column_name: "reference" |
| | * response_column_name: "response" |
| | * baseline_model_response_column_name: "baseline_model_response" |
| | |
| | Requirement for different use cases: |
| | * Bring-your-own-response (BYOR): You already have the data that you |
| | want to evaluate stored in the dataset. Response column name can be |
| | customized by providing `response_column_name` parameter, or in the |
| | `metric_column_mapping`. For BYOR pairwise evaluation, the baseline |
| | model response column name can be customized by providing |
| | `baseline_model_response_column_name` parameter, or |
| | in the `metric_column_mapping`. If the `response` column or |
| | `baseline_model_response` column is present while the |
| | corresponding model is specified, an error will be raised. |
| | |
| | * Perform model inference without a prompt template: You have a dataset |
| | containing the input prompts to the model and want to perform model |
| | inference before evaluation. A column named `prompt` is required |
| | in the evaluation dataset and is used directly as input to the model. |
| | |
| | * Perform model inference with a prompt template: You have a dataset |
| | containing the input variables to the prompt template and want to |
| | assemble the prompts for model inference. Evaluation dataset |
| | must contain column names corresponding to the variable names in |
| | the prompt template. For example, if prompt template is |
| | "Instruction: {instruction}, context: {context}", the dataset must |
| | contain `instruction` and `context` columns. |
| | |
| | Metrics Details: |
| | |
| | The supported metrics descriptions, rating rubrics, and the required |
| | input variables can be found on the Vertex AI public documentation page. |
| | [Evaluation methods and metrics](https://cloud.google.com/vertex-ai/generative-ai/docs/models/determine-eval). |
| | |
| | Usage Examples: |
| | |
| | 1. To perform bring-your-own-response(BYOR) evaluation, provide the model |
| | responses in the `response` column in the dataset. If a pairwise metric is |
| | used for BYOR evaluation, provide the baseline model responses in the |
| | `baseline_model_response` column. |
| | |
| | ``` |
| | eval_dataset = pd.DataFrame({ |
| | "prompt" : [...], |
| | "reference": [...], |
| | "response" : [...], |
| | "baseline_model_response": [...], |
| | }) |
| | eval_task = EvalTask( |
| | dataset=eval_dataset, |
| | metrics=[ |
| | "bleu", |
| | "rouge_l_sum", |
| | MetricPromptTemplateExamples.Pointwise.FLUENCY, |
| | MetricPromptTemplateExamples.Pairwise.SAFETY |
| | ], |
| | experiment="my-experiment", |
| | ) |
| | eval_result = eval_task.evaluate(experiment_run_name="eval-experiment-run") |
| | ``` |
| | |
| | 2. To perform evaluation with Gemini model inference, specify the `model` |
| | parameter with a `GenerativeModel` instance. The input column name to the |
| | model is `prompt` and must be present in the dataset. |
| | |
| | ``` |
| | eval_dataset = pd.DataFrame({ |
| | "reference": [...], |
| | "prompt" : [...], |
| | }) |
| | result = EvalTask( |
| | dataset=eval_dataset, |
| | metrics=["exact_match", "bleu", "rouge_1", "rouge_l_sum"], |
| | experiment="my-experiment", |
| | ).evaluate( |
| | model=GenerativeModel("gemini-1.5-pro"), |
| | experiment_run_name="gemini-eval-run" |
| | ) |
| | ``` |
| | |
| | 3. If a `prompt_template` is specified, the `prompt` column is not required. |
| | Prompts can be assembled from the evaluation dataset, and all prompt |
| | template variable names must be present in the dataset columns. |
| | ``` |
| | eval_dataset = pd.DataFrame({ |
| | "context" : [...], |
| | "instruction": [...], |
| | }) |
| | result = EvalTask( |
| | dataset=eval_dataset, |
| | metrics=[MetricPromptTemplateExamples.Pointwise.SUMMARIZATION_QUALITY], |
| | ).evaluate( |
| | model=GenerativeModel("gemini-1.5-pro"), |
| | prompt_template="{instruction}. Article: {context}. Summary:", |
| | ) |
| | ``` |
| | |
| | 4. To perform evaluation with custom model inference, specify the `model` |
| | parameter with a custom inference function. The input column name to the |
| | custom inference function is `prompt` and must be present in the dataset. |
| | |
| | ``` |
| | from openai import OpenAI |
| | client = OpenAI() |
| | def custom_model_fn(input: str) -> str: |
| | response = client.chat.completions.create( |
| | model="gpt-3.5-turbo", |
| | messages=[ |
| | {"role": "user", "content": input} |
| | ] |
| | ) |
| | return response.choices[0].message.content |
| | |
| | eval_dataset = pd.DataFrame({ |
| | "prompt" : [...], |
| | "reference": [...], |
| | }) |
| | result = EvalTask( |
| | dataset=eval_dataset, |
| | metrics=[MetricPromptTemplateExamples.Pointwise.SAFETY], |
| | experiment="my-experiment", |
| | ).evaluate( |
| | model=custom_model_fn, |
| | experiment_run_name="gpt-eval-run" |
| | ) |
| | ``` |
| | |
| | 5. To perform pairwise metric evaluation with model inference step, specify |
| | the `baseline_model` input to a `PairwiseMetric` instance and the candidate |
| | `model` input to the `EvalTask.evaluate()` function. The input column name |
| | to both models is `prompt` and must be present in the dataset. |
| | |
| | ``` |
| | baseline_model = GenerativeModel("gemini-1.0-pro") |
| | candidate_model = GenerativeModel("gemini-1.5-pro") |
| | |
| | pairwise_groundedness = PairwiseMetric( |
| | metric_prompt_template=MetricPromptTemplateExamples.get_prompt_template( |
| | "pairwise_groundedness" |
| | ), |
| | baseline_model=baseline_model, |
| | ) |
| | eval_dataset = pd.DataFrame({ |
| | "prompt" : [...], |
| | }) |
| | result = EvalTask( |
| | dataset=eval_dataset, |
| | metrics=[pairwise_groundedness], |
| | experiment="my-pairwise-experiment", |
| | ).evaluate( |
| | model=candidate_model, |
| | experiment_run_name="gemini-pairwise-eval-run", |
| | ) |
| | ``` |
| | """ |
| |
|
| | _resource_noun = "evaluationTasks" |
| |
|
| | def __init__( |
| | self, |
| | *, |
| | dataset: Union["pd.DataFrame", str, Dict[str, Any]], |
| | metrics: List[ |
| | Union[ |
| | Literal[ |
| | "exact_match", |
| | "bleu", |
| | "rouge_1", |
| | "rouge_2", |
| | "rouge_l", |
| | "rouge_l_sum", |
| | "tool_call_valid", |
| | "tool_name_match", |
| | "tool_parameter_key_match", |
| | "tool_parameter_kv_match", |
| | ], |
| | metrics_base.CustomMetric, |
| | metrics_base._AutomaticMetric, |
| | metrics_base._TranslationMetric, |
| | pointwise_metric.PointwiseMetric, |
| | pairwise_metric.PairwiseMetric, |
| | ] |
| | ], |
| | experiment: Optional[str] = None, |
| | metric_column_mapping: Optional[Dict[str, str]] = None, |
| | output_uri_prefix: Optional[str] = "", |
| | ): |
| | """Initializes an EvalTask. |
| | |
| | Args: |
| | dataset: The dataset to be evaluated. |
| | Supports the following dataset formats: |
| | * pandas.DataFrame: Used directly for evaluation. |
| | * Dict: Converted to a pandas DataFrame before evaluation. |
| | * str: Interpreted as a file path or URI. Supported formats include: |
| | * Local JSONL or CSV files: Loaded from the local filesystem. |
| | * GCS JSONL or CSV files: Loaded from Google Cloud Storage |
| | (e.g., 'gs://bucket/data.csv'). |
| | * BigQuery table URI: Loaded from Google Cloud BigQuery |
| | (e.g., 'bq://project-id.dataset.table_name'). |
| | metrics: The list of metric names, or Metric instances to evaluate. |
| | Prompt template is required for PairwiseMetric. |
| | experiment: The name of the experiment to log the evaluations to. |
| | metric_column_mapping: An optional dictionary column mapping that |
| | overrides the metric prompt template input variable names with |
| | mapped the evaluation dataset column names, used during evaluation. |
| | For example, if the input_variables of the metric prompt template |
| | are ["context", "reference"], the metric_column_mapping can be |
| | { |
| | "context": "news_context", |
| | "reference": "ground_truth", |
| | "response": "model_1_response" |
| | } |
| | if the dataset has columns "news_context", "ground_truth" and |
| | "model_1_response". |
| | output_uri_prefix: GCS location to store the metrics_table from |
| | evaluation results. |
| | """ |
| | self._dataset = utils.load_dataset(dataset) |
| | self._metrics = metrics |
| | self._experiment = experiment |
| | self._metric_column_mapping = utils.initialize_metric_column_mapping( |
| | metric_column_mapping, self._dataset |
| | ) |
| | self.output_uri_prefix = output_uri_prefix |
| |
|
| | @property |
| | def dataset(self) -> "pd.DataFrame": |
| | """Returns evaluation dataset.""" |
| | return self._dataset |
| |
|
| | @property |
| | def metrics(self) -> List[Union[str, metrics_base.CustomMetric]]: |
| | """Returns metrics.""" |
| | return self._metrics |
| |
|
| | @property |
| | def experiment(self) -> Optional[str]: |
| | """Returns experiment name.""" |
| | return self._experiment |
| |
|
| | def _evaluate_with_experiment( |
| | self, |
| | *, |
| | model: Optional[Union[GenerativeModel, Callable[[str], str]]] = None, |
| | prompt_template: Optional[str] = None, |
| | experiment_run_name: Optional[str] = None, |
| | evaluation_service_qps: Optional[float] = None, |
| | retry_timeout: float = 120.0, |
| | ) -> EvalResult: |
| | """Runs an evaluation for the EvalTask with an experiment. |
| | |
| | Args: |
| | model: A GenerativeModel instance or a custom model function to generate |
| | responses to evaluate. If not provided, the evaluation is computed with |
| | the `response` column in the `dataset`. |
| | prompt_template: The prompt template to use for the evaluation. If not |
| | set, the prompt template that was used to create the EvalTask will be |
| | used. |
| | experiment_run_name: The name of the experiment run to log the evaluation |
| | to if an experiment is set for this EvalTask. If not provided, a random |
| | unique experiment run name is used. |
| | evaluation_service_qps: The custom QPS limit for the evaluation service. |
| | retry_timeout: How long to keep retrying the evaluation requests for |
| | the whole evaluation dataset, in seconds. |
| | |
| | Returns: |
| | The evaluation result. |
| | """ |
| | self._validate_experiment_run() |
| | with vertexai.preview.start_run(experiment_run_name): |
| | self._log_eval_experiment_param(model, prompt_template) |
| | eval_result = _evaluation.evaluate( |
| | dataset=self._dataset, |
| | metrics=self._metrics, |
| | model=model, |
| | prompt_template=prompt_template, |
| | metric_column_mapping=self._metric_column_mapping, |
| | evaluation_service_qps=evaluation_service_qps, |
| | retry_timeout=retry_timeout, |
| | ) |
| |
|
| | eval_result.summary_metrics = { |
| | k: ("NaN" if isinstance(v, float) and np.isnan(v) else v) |
| | for k, v in eval_result.summary_metrics.items() |
| | } |
| | eval_result.metadata = { |
| | "experiment": self._experiment, |
| | "experiment_run": experiment_run_name, |
| | } |
| | try: |
| | vertexai.preview.log_metrics(eval_result.summary_metrics) |
| | except (TypeError, exceptions.InvalidArgument) as e: |
| | _LOGGER.warning(f"Experiment metrics logging failed: {str(e)}") |
| | return eval_result |
| |
|
| | def evaluate( |
| | self, |
| | *, |
| | model: Optional[Union[GenerativeModel, Callable[[str], str]]] = None, |
| | prompt_template: Optional[str] = None, |
| | experiment_run_name: Optional[str] = None, |
| | response_column_name: Optional[str] = None, |
| | baseline_model_response_column_name: Optional[str] = None, |
| | evaluation_service_qps: Optional[float] = None, |
| | retry_timeout: float = 120.0, |
| | output_file_name: Optional[str] = None, |
| | ) -> EvalResult: |
| | """Runs an evaluation for the EvalTask. |
| | |
| | Args: |
| | model: A GenerativeModel instance or a custom model function to generate |
| | responses to evaluate. If not provided, the evaluation can be performed |
| | in the bring-your-own-response (BYOR) mode. |
| | prompt_template: The prompt template to use for the evaluation. If not |
| | set, the prompt template that was used to create the EvalTask will be |
| | used. |
| | experiment_run_name: The name of the experiment run to log the evaluation |
| | to if an experiment is set for this EvalTask. If not provided, a random |
| | unique experiment run name is used. |
| | response_column_name: The column name of model response in the dataset. If |
| | provided, this will override the `metric_column_mapping` of the `EvalTask`. |
| | baseline_model_response_column_name: The column name of baseline model |
| | response in the dataset for pairwise metrics. If provided, this will |
| | override the `metric_column_mapping` of the `EvalTask` |
| | evaluation_service_qps: The custom QPS limit for the evaluation service. |
| | retry_timeout: How long to keep retrying the evaluation requests for |
| | the whole evaluation dataset, in seconds. |
| | output_file_name: The file name with csv suffix to store the output |
| | metrics_table. |
| | |
| | Returns: |
| | The evaluation result. |
| | """ |
| | global_experiment_name = metadata._experiment_tracker.experiment_name |
| | if experiment_run_name and not self._experiment and not global_experiment_name: |
| | raise ValueError( |
| | "Experiment is not set. Please initialize `EvalTask` with an" |
| | " experiment, or initialize a global experiment with " |
| | "`vertexai.init(experiment='experiment_name')`for logging this" |
| | " evaluation run." |
| | ) |
| | self._verify_and_set_response_column_name( |
| | response_column_name=response_column_name, |
| | metric_column_mapping_key=constants.Dataset.MODEL_RESPONSE_COLUMN, |
| | ) |
| | self._verify_and_set_response_column_name( |
| | response_column_name=baseline_model_response_column_name, |
| | metric_column_mapping_key=constants.Dataset.BASELINE_MODEL_RESPONSE_COLUMN, |
| | ) |
| |
|
| | experiment_run_name = experiment_run_name or f"{uuid.uuid4()}" |
| | if self._experiment and global_experiment_name: |
| | metadata._experiment_tracker.set_experiment( |
| | experiment=self._experiment, backing_tensorboard=False |
| | ) |
| | eval_result = self._evaluate_with_experiment( |
| | model=model, |
| | prompt_template=prompt_template, |
| | experiment_run_name=experiment_run_name, |
| | evaluation_service_qps=evaluation_service_qps, |
| | retry_timeout=retry_timeout, |
| | ) |
| | metadata._experiment_tracker.set_experiment( |
| | experiment=global_experiment_name, backing_tensorboard=False |
| | ) |
| | elif self._experiment and not global_experiment_name: |
| | metadata._experiment_tracker.set_experiment( |
| | experiment=self._experiment, backing_tensorboard=False |
| | ) |
| | eval_result = self._evaluate_with_experiment( |
| | model=model, |
| | prompt_template=prompt_template, |
| | experiment_run_name=experiment_run_name, |
| | evaluation_service_qps=evaluation_service_qps, |
| | retry_timeout=retry_timeout, |
| | ) |
| | metadata._experiment_tracker.reset() |
| | elif not self._experiment and global_experiment_name: |
| | eval_result = self._evaluate_with_experiment( |
| | model=model, |
| | prompt_template=prompt_template, |
| | experiment_run_name=experiment_run_name, |
| | evaluation_service_qps=evaluation_service_qps, |
| | retry_timeout=retry_timeout, |
| | ) |
| | else: |
| | eval_result = _evaluation.evaluate( |
| | dataset=self.dataset, |
| | metrics=self.metrics, |
| | model=model, |
| | prompt_template=prompt_template, |
| | metric_column_mapping=self._metric_column_mapping, |
| | evaluation_service_qps=evaluation_service_qps, |
| | retry_timeout=retry_timeout, |
| | ) |
| | utils.upload_evaluation_results( |
| | eval_result.metrics_table, self.output_uri_prefix, output_file_name |
| | ) |
| | return eval_result |
| |
|
| | def _validate_experiment_run(self) -> None: |
| | """Checks if an experiment run already exists.""" |
| | if metadata._experiment_tracker.experiment_run: |
| | raise ValueError( |
| | "Experiment run already exists. Please specify the name of the" |
| | " experiment run to assign current session within this evaluation." |
| | ) |
| |
|
| | def _log_eval_experiment_param( |
| | self, |
| | model: Optional[Union[GenerativeModel, Callable[[str], str]]] = None, |
| | prompt_template: Optional[str] = None, |
| | ) -> None: |
| | """Logs variable input parameters of an evaluation to an experiment run.""" |
| | model_metadata = {} |
| |
|
| | if prompt_template is not None: |
| | model_metadata.update({"prompt_template": prompt_template}) |
| |
|
| | if isinstance(model, GenerativeModel): |
| | model_metadata.update( |
| | { |
| | "model_name": model._model_name, |
| | } |
| | ) |
| |
|
| | if model._generation_config and isinstance(model._generation_config, dict): |
| | model_metadata.update(**model._generation_config) |
| |
|
| | if model._safety_settings and isinstance(model._safety_settings, dict): |
| | safety_settings = model._safety_settings |
| | safety_settings_as_str = { |
| | category.name: threshold.name |
| | for category, threshold in safety_settings.items() |
| | } |
| | model_metadata.update(safety_settings_as_str) |
| |
|
| | if model_metadata: |
| | _LOGGER.info(f"Logging Eval Experiment metadata: {model_metadata}") |
| | try: |
| | vertexai.preview.log_params(model_metadata) |
| | except (ValueError, TypeError) as e: |
| | _LOGGER.warning(f"Experiment metadata logging failed: {str(e)}") |
| |
|
| | def _verify_and_set_response_column_name( |
| | self, response_column_name: str, metric_column_mapping_key: str |
| | ) -> None: |
| | """Verifies and sets the model response column names.""" |
| | if response_column_name: |
| | if response_column_name in self._dataset.columns: |
| | self._metric_column_mapping[ |
| | metric_column_mapping_key |
| | ] = response_column_name |
| | else: |
| | raise ValueError( |
| | f"(Baseline) Model response column {response_column_name} is not" |
| | " found in the dataset." |
| | ) |
| |
|
| | def display_runs(self): |
| | """Displays experiment runs associated with this EvalTask.""" |
| | if not self._experiment: |
| | raise ValueError("Experiment is not set.") |
| | elif IPython_display: |
| | IPython_display.display( |
| | vertexai.preview.get_experiment_df(self._experiment) |
| | ) |
| |
|