| | |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | from typing import Iterable, Optional, Union, Sequence, Dict, List, Tuple |
| |
|
| | import abc |
| | import copy |
| | import datetime |
| | import time |
| | import tempfile |
| |
|
| | from google.auth import credentials as auth_credentials |
| | from google.api_core import exceptions as api_exceptions |
| | from google.protobuf import duration_pb2 |
| | from google.protobuf import field_mask_pb2 |
| | from google.rpc import status_pb2 |
| |
|
| | from google.cloud import aiplatform |
| | from google.cloud.aiplatform import base |
| | from google.cloud.aiplatform.compat.types import ( |
| | batch_prediction_job as gca_bp_job_compat, |
| | completion_stats as gca_completion_stats, |
| | custom_job as gca_custom_job_compat, |
| | execution as gca_execution_compat, |
| | explanation as gca_explanation_compat, |
| | encryption_spec as gca_encryption_spec_compat, |
| | io as gca_io_compat, |
| | job_state as gca_job_state, |
| | hyperparameter_tuning_job as gca_hyperparameter_tuning_job_compat, |
| | study as gca_study_compat, |
| | model_deployment_monitoring_job as gca_model_deployment_monitoring_job_compat, |
| | job_state_v1beta1 as gca_job_state_v1beta1, |
| | model_monitoring_v1beta1 as gca_model_monitoring_v1beta1, |
| | ) |
| |
|
| | from google.cloud.aiplatform.constants import base as constants |
| | from google.cloud.aiplatform.metadata import constants as metadata_constants |
| | from google.cloud.aiplatform import initializer |
| | from google.cloud.aiplatform import hyperparameter_tuning |
| | from google.cloud.aiplatform import model_monitoring |
| | from google.cloud.aiplatform import utils |
| | from google.cloud.aiplatform import _publisher_models |
| | from google.cloud.aiplatform.utils import console_utils |
| | from google.cloud.aiplatform.utils import source_utils |
| | from google.cloud.aiplatform.utils import worker_spec_utils |
| |
|
| | from google.cloud.aiplatform_v1.types import ( |
| | batch_prediction_job as batch_prediction_job_v1, |
| | ) |
| | from google.cloud.aiplatform_v1.types import custom_job as custom_job_v1 |
| |
|
| | _LOGGER = base.Logger(__name__) |
| |
|
| | |
| | _JOB_COMPLETE_STATES = ( |
| | gca_job_state.JobState.JOB_STATE_SUCCEEDED, |
| | gca_job_state.JobState.JOB_STATE_FAILED, |
| | gca_job_state.JobState.JOB_STATE_CANCELLED, |
| | gca_job_state.JobState.JOB_STATE_PAUSED, |
| | gca_job_state_v1beta1.JobState.JOB_STATE_SUCCEEDED, |
| | gca_job_state_v1beta1.JobState.JOB_STATE_FAILED, |
| | gca_job_state_v1beta1.JobState.JOB_STATE_CANCELLED, |
| | gca_job_state_v1beta1.JobState.JOB_STATE_PAUSED, |
| | ) |
| |
|
| | _JOB_ERROR_STATES = ( |
| | gca_job_state.JobState.JOB_STATE_FAILED, |
| | gca_job_state.JobState.JOB_STATE_CANCELLED, |
| | gca_job_state_v1beta1.JobState.JOB_STATE_FAILED, |
| | gca_job_state_v1beta1.JobState.JOB_STATE_CANCELLED, |
| | ) |
| |
|
| | _JOB_PENDING_STATES = ( |
| | gca_job_state.JobState.JOB_STATE_QUEUED, |
| | gca_job_state.JobState.JOB_STATE_PENDING, |
| | gca_job_state.JobState.JOB_STATE_RUNNING, |
| | gca_job_state.JobState.JOB_STATE_CANCELLING, |
| | gca_job_state.JobState.JOB_STATE_UPDATING, |
| | gca_job_state_v1beta1.JobState.JOB_STATE_QUEUED, |
| | gca_job_state_v1beta1.JobState.JOB_STATE_PENDING, |
| | gca_job_state_v1beta1.JobState.JOB_STATE_RUNNING, |
| | gca_job_state_v1beta1.JobState.JOB_STATE_CANCELLING, |
| | gca_job_state_v1beta1.JobState.JOB_STATE_UPDATING, |
| | ) |
| |
|
| | |
| | _JOB_WAIT_TIME = 5 |
| | _LOG_WAIT_TIME = 5 |
| | _MAX_WAIT_TIME = 60 * 5 |
| | _WAIT_TIME_MULTIPLIER = 2 |
| |
|
| |
|
| | class _Job(base.VertexAiStatefulResource): |
| | """Class that represents a general Job resource in Vertex AI. |
| | Cannot be directly instantiated. |
| | |
| | Serves as base class to specific Job types, i.e. BatchPredictionJob or |
| | DataLabelingJob to re-use shared functionality. |
| | |
| | Subclasses requires one class attribute: |
| | |
| | _getter_method (str): The name of JobServiceClient getter method for specific |
| | Job type, i.e. 'get_custom_job' for CustomJob |
| | _cancel_method (str): The name of the specific JobServiceClient cancel method |
| | _delete_method (str): The name of the specific JobServiceClient delete method |
| | """ |
| |
|
| | client_class = utils.JobClientWithOverride |
| |
|
| | |
| | _valid_done_states = _JOB_COMPLETE_STATES |
| |
|
| | def __init__( |
| | self, |
| | job_name: str, |
| | project: Optional[str] = None, |
| | location: Optional[str] = None, |
| | credentials: Optional[auth_credentials.Credentials] = None, |
| | ): |
| | """Retrieves Job subclass resource by calling a subclass-specific getter |
| | method. |
| | |
| | Args: |
| | job_name (str): |
| | Required. A fully-qualified job resource name or job ID. |
| | Example: "projects/123/locations/us-central1/batchPredictionJobs/456" or |
| | "456" when project, location and job_type are initialized or passed. |
| | project: Optional[str] = None, |
| | Optional. project to retrieve Job subclass from. If not set, |
| | project set in aiplatform.init will be used. |
| | location: Optional[str] = None, |
| | Optional. location to retrieve Job subclass from. If not set, |
| | location set in aiplatform.init will be used. |
| | credentials: Optional[auth_credentials.Credentials] = None, |
| | Custom credentials to use. If not set, credentials set in |
| | aiplatform.init will be used. |
| | """ |
| | super().__init__( |
| | project=project, |
| | location=location, |
| | credentials=credentials, |
| | resource_name=job_name, |
| | ) |
| | self._gca_resource = self._get_gca_resource(resource_name=job_name) |
| |
|
| | @property |
| | def state(self) -> gca_job_state.JobState: |
| | """Fetch Job again and return the current JobState. |
| | |
| | Returns: |
| | state (job_state.JobState): |
| | Enum that describes the state of a Vertex AI job. |
| | """ |
| |
|
| | |
| | self._sync_gca_resource() |
| |
|
| | return self._gca_resource.state |
| |
|
| | @property |
| | def start_time(self) -> Optional[datetime.datetime]: |
| | """Time when the Job resource entered the `JOB_STATE_RUNNING` for the |
| | first time.""" |
| | self._sync_gca_resource() |
| | return getattr(self._gca_resource, "start_time") |
| |
|
| | @property |
| | def end_time(self) -> Optional[datetime.datetime]: |
| | """Time when the Job resource entered the `JOB_STATE_SUCCEEDED`, |
| | `JOB_STATE_FAILED`, or `JOB_STATE_CANCELLED` state.""" |
| | self._sync_gca_resource() |
| | return getattr(self._gca_resource, "end_time") |
| |
|
| | @property |
| | def error(self) -> Optional[status_pb2.Status]: |
| | """Detailed error info for this Job resource. Only populated when the |
| | Job's state is `JOB_STATE_FAILED` or `JOB_STATE_CANCELLED`.""" |
| | self._sync_gca_resource() |
| | return getattr(self._gca_resource, "error") |
| |
|
| | @property |
| | @abc.abstractmethod |
| | def _job_type(cls) -> str: |
| | """Job type.""" |
| | pass |
| |
|
| | @property |
| | @abc.abstractmethod |
| | def _cancel_method(cls) -> str: |
| | """Name of cancellation method for cancelling the specific job type.""" |
| | pass |
| |
|
| | def _dashboard_uri(self) -> Optional[str]: |
| | """Helper method to compose the dashboard uri where job can be |
| | viewed.""" |
| | fields = self._parse_resource_name(self.resource_name) |
| | location = fields.pop("location") |
| | project = fields.pop("project") |
| | job = list(fields.values())[0] |
| | url = f"https://console.cloud.google.com/ai/platform/locations/{location}/{self._job_type}/{job}?project={project}" |
| | return url |
| |
|
| | def _log_job_state(self): |
| | """Helper method to log job state.""" |
| | _LOGGER.info( |
| | "%s %s current state:\n%s" |
| | % ( |
| | self.__class__.__name__, |
| | self._gca_resource.name, |
| | self._gca_resource.state, |
| | ) |
| | ) |
| |
|
| | def _block_until_complete(self): |
| | """Helper method to block and check on job until complete. |
| | |
| | Raises: |
| | RuntimeError: If job failed or cancelled. |
| | """ |
| |
|
| | log_wait = _LOG_WAIT_TIME |
| |
|
| | previous_time = time.time() |
| | while self.state not in _JOB_COMPLETE_STATES: |
| | current_time = time.time() |
| | if current_time - previous_time >= log_wait: |
| | self._log_job_state() |
| | log_wait = min(log_wait * _WAIT_TIME_MULTIPLIER, _MAX_WAIT_TIME) |
| | previous_time = current_time |
| | time.sleep(_JOB_WAIT_TIME) |
| |
|
| | self._log_job_state() |
| |
|
| | |
| | |
| | if self._gca_resource.state in _JOB_ERROR_STATES: |
| | raise RuntimeError("Job failed with:\n%s" % self._gca_resource.error) |
| | else: |
| | _LOGGER.log_action_completed_against_resource("run", "completed", self) |
| |
|
| | def wait_for_completion(self) -> None: |
| | """Waits for job to complete. |
| | |
| | Raises: |
| | RuntimeError: If job failed or cancelled. |
| | """ |
| | self._block_until_complete() |
| |
|
| | @classmethod |
| | def list( |
| | cls, |
| | filter: Optional[str] = None, |
| | order_by: Optional[str] = None, |
| | project: Optional[str] = None, |
| | location: Optional[str] = None, |
| | credentials: Optional[auth_credentials.Credentials] = None, |
| | ) -> List[base.VertexAiResourceNoun]: |
| | """List all instances of this Job Resource. |
| | |
| | Example Usage: |
| | |
| | aiplatform.BatchPredictionJobs.list( |
| | filter='state="JOB_STATE_SUCCEEDED" AND display_name="my_job"', |
| | ) |
| | |
| | Args: |
| | filter (str): |
| | Optional. An expression for filtering the results of the request. |
| | For field names both snake_case and camelCase are supported. |
| | order_by (str): |
| | Optional. A comma-separated list of fields to order by, sorted in |
| | ascending order. Use "desc" after a field name for descending. |
| | Supported fields: `display_name`, `create_time`, `update_time` |
| | project (str): |
| | Optional. Project to retrieve list from. If not set, project |
| | set in aiplatform.init will be used. |
| | location (str): |
| | Optional. Location to retrieve list from. If not set, location |
| | set in aiplatform.init will be used. |
| | credentials (auth_credentials.Credentials): |
| | Optional. Custom credentials to use to retrieve list. Overrides |
| | credentials set in aiplatform.init. |
| | |
| | Returns: |
| | List[VertexAiResourceNoun] - A list of Job resource objects. |
| | """ |
| |
|
| | return cls._list_with_local_order( |
| | filter=filter, |
| | order_by=order_by, |
| | project=project, |
| | location=location, |
| | credentials=credentials, |
| | ) |
| |
|
| | def cancel(self) -> None: |
| | """Cancels this Job. |
| | |
| | Success of cancellation is not guaranteed. Use `Job.state` |
| | property to verify if cancellation was successful. |
| | """ |
| |
|
| | _LOGGER.log_action_start_against_resource("Cancelling", "run", self) |
| | getattr(self.api_client, self._cancel_method)(name=self.resource_name) |
| |
|
| |
|
| | class BatchPredictionJob(_Job): |
| |
|
| | _resource_noun = "batchPredictionJobs" |
| | _getter_method = "get_batch_prediction_job" |
| | _list_method = "list_batch_prediction_jobs" |
| | _cancel_method = "cancel_batch_prediction_job" |
| | _delete_method = "delete_batch_prediction_job" |
| | _job_type = "batch-predictions" |
| | _parse_resource_name_method = "parse_batch_prediction_job_path" |
| | _format_resource_name_method = "batch_prediction_job_path" |
| |
|
| | def __init__( |
| | self, |
| | batch_prediction_job_name: str, |
| | project: Optional[str] = None, |
| | location: Optional[str] = None, |
| | credentials: Optional[auth_credentials.Credentials] = None, |
| | ): |
| | """Retrieves a BatchPredictionJob resource and instantiates its |
| | representation. |
| | |
| | Args: |
| | batch_prediction_job_name (str): |
| | Required. A fully-qualified BatchPredictionJob resource name or ID. |
| | Example: "projects/.../locations/.../batchPredictionJobs/456" or |
| | "456" when project and location are initialized or passed. |
| | project: Optional[str] = None, |
| | Optional. project to retrieve BatchPredictionJob from. If not set, |
| | project set in aiplatform.init will be used. |
| | location: Optional[str] = None, |
| | Optional. location to retrieve BatchPredictionJob from. If not set, |
| | location set in aiplatform.init will be used. |
| | credentials: Optional[auth_credentials.Credentials] = None, |
| | Custom credentials to use. If not set, credentials set in |
| | aiplatform.init will be used. |
| | """ |
| |
|
| | super().__init__( |
| | job_name=batch_prediction_job_name, |
| | project=project, |
| | location=location, |
| | credentials=credentials, |
| | ) |
| |
|
| | @property |
| | def output_info( |
| | self, |
| | ) -> Optional[batch_prediction_job_v1.BatchPredictionJob.OutputInfo]: |
| | """Information describing the output of this job, including output location |
| | into which prediction output is written. |
| | |
| | This is only available for batch prediction jobs that have run successfully. |
| | """ |
| | self._assert_gca_resource_is_available() |
| | return self._gca_resource.output_info |
| |
|
| | @property |
| | def partial_failures(self) -> Optional[Sequence[status_pb2.Status]]: |
| | """Partial failures encountered. For example, single files that can't be read. |
| | This field never exceeds 20 entries. Status details fields contain standard |
| | GCP error details.""" |
| | self._assert_gca_resource_is_available() |
| | return getattr(self._gca_resource, "partial_failures") |
| |
|
| | @property |
| | def completion_stats(self) -> Optional[gca_completion_stats.CompletionStats]: |
| | """Statistics on completed and failed prediction instances.""" |
| | self._assert_gca_resource_is_available() |
| | return getattr(self._gca_resource, "completion_stats") |
| |
|
| | @classmethod |
| | def create( |
| | cls, |
| | |
| | job_display_name: str, |
| | model_name: Union[str, "aiplatform.Model"], |
| | instances_format: str = "jsonl", |
| | predictions_format: str = "jsonl", |
| | gcs_source: Optional[Union[str, Sequence[str]]] = None, |
| | bigquery_source: Optional[str] = None, |
| | gcs_destination_prefix: Optional[str] = None, |
| | bigquery_destination_prefix: Optional[str] = None, |
| | model_parameters: Optional[Dict] = None, |
| | machine_type: Optional[str] = None, |
| | accelerator_type: Optional[str] = None, |
| | accelerator_count: Optional[int] = None, |
| | starting_replica_count: Optional[int] = None, |
| | max_replica_count: Optional[int] = None, |
| | generate_explanation: Optional[bool] = False, |
| | explanation_metadata: Optional["aiplatform.explain.ExplanationMetadata"] = None, |
| | explanation_parameters: Optional[ |
| | "aiplatform.explain.ExplanationParameters" |
| | ] = None, |
| | labels: Optional[Dict[str, str]] = None, |
| | project: Optional[str] = None, |
| | location: Optional[str] = None, |
| | credentials: Optional[auth_credentials.Credentials] = None, |
| | encryption_spec_key_name: Optional[str] = None, |
| | sync: bool = True, |
| | create_request_timeout: Optional[float] = None, |
| | batch_size: Optional[int] = None, |
| | model_monitoring_objective_config: Optional[ |
| | "aiplatform.model_monitoring.ObjectiveConfig" |
| | ] = None, |
| | model_monitoring_alert_config: Optional[ |
| | "aiplatform.model_monitoring.AlertConfig" |
| | ] = None, |
| | analysis_instance_schema_uri: Optional[str] = None, |
| | service_account: Optional[str] = None, |
| | ) -> "BatchPredictionJob": |
| | """Create a batch prediction job. |
| | |
| | Args: |
| | job_display_name (str): |
| | Required. The user-defined name of the BatchPredictionJob. |
| | The name can be up to 128 characters long and can be consist |
| | of any UTF-8 characters. |
| | model_name (Union[str, aiplatform.Model]): |
| | Required. A fully-qualified model resource name or model ID. |
| | Example: "projects/123/locations/us-central1/models/456" or |
| | "456" when project and location are initialized or passed. |
| | May optionally contain a version ID or alias in |
| | {model_name}@{version} form. |
| | |
| | Or an instance of aiplatform.Model. |
| | instances_format (str): |
| | Required. The format in which instances are provided. Must be one |
| | of the formats listed in `Model.supported_input_storage_formats`. |
| | Default is "jsonl" when using `gcs_source`. If a `bigquery_source` |
| | is provided, this is overridden to "bigquery". |
| | predictions_format (str): |
| | Required. The format in which Vertex AI outputs the |
| | predictions, must be one of the formats specified in |
| | `Model.supported_output_storage_formats`. |
| | Default is "jsonl" when using `gcs_destination_prefix`. If a |
| | `bigquery_destination_prefix` is provided, this is overridden to |
| | "bigquery". |
| | gcs_source (Optional[Sequence[str]]): |
| | Google Cloud Storage URI(-s) to your instances to run |
| | batch prediction on. They must match `instances_format`. |
| | |
| | bigquery_source (Optional[str]): |
| | BigQuery URI to a table, up to 2000 characters long. For example: |
| | `bq://projectId.bqDatasetId.bqTableId` |
| | gcs_destination_prefix (Optional[str]): |
| | The Google Cloud Storage location of the directory where the |
| | output is to be written to. In the given directory a new |
| | directory is created. Its name is |
| | ``prediction-<model-display-name>-<job-create-time>``, where |
| | timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format. |
| | Inside of it files ``predictions_0001.<extension>``, |
| | ``predictions_0002.<extension>``, ..., |
| | ``predictions_N.<extension>`` are created where |
| | ``<extension>`` depends on chosen ``predictions_format``, |
| | and N may equal 0001 and depends on the total number of |
| | successfully predicted instances. If the Model has both |
| | ``instance`` and ``prediction`` schemata defined then each such |
| | file contains predictions as per the ``predictions_format``. |
| | If prediction for any instance failed (partially or |
| | completely), then an additional ``errors_0001.<extension>``, |
| | ``errors_0002.<extension>``,..., ``errors_N.<extension>`` |
| | files are created (N depends on total number of failed |
| | predictions). These files contain the failed instances, as |
| | per their schema, followed by an additional ``error`` field |
| | which as value has ```google.rpc.Status`` <Status>`__ |
| | containing only ``code`` and ``message`` fields. |
| | bigquery_destination_prefix (Optional[str]): |
| | The BigQuery project or dataset location where the output is |
| | to be written to. If project is provided, a new dataset is |
| | created with name |
| | ``prediction_<model-display-name>_<job-create-time>`` where |
| | is made BigQuery-dataset-name compatible (for example, most |
| | special characters become underscores), and timestamp is in |
| | YYYY_MM_DDThh_mm_ss_sssZ "based on ISO-8601" format. In the |
| | dataset two tables will be created, ``predictions``, and |
| | ``errors``. If the Model has both |
| | [instance][google.cloud.aiplatform.v1.PredictSchemata.instance_schema_uri] |
| | and |
| | [prediction][google.cloud.aiplatform.v1.PredictSchemata.parameters_schema_uri] |
| | schemata defined then the tables have columns as follows: |
| | The ``predictions`` table contains instances for which the |
| | prediction succeeded, it has columns as per a concatenation |
| | of the Model's instance and prediction schemata. The |
| | ``errors`` table contains rows for which the prediction has |
| | failed, it has instance columns, as per the instance schema, |
| | followed by a single "errors" column, which as values has |
| | [google.rpc.Status][google.rpc.Status] represented as a |
| | STRUCT, and containing only ``code`` and ``message``. |
| | model_parameters (Optional[Dict]): |
| | The parameters that govern the predictions. The schema of |
| | the parameters may be specified via the Model's `parameters_schema_uri`. |
| | machine_type (Optional[str]): |
| | The type of machine for running batch prediction on |
| | dedicated resources. Not specifying machine type will result in |
| | batch prediction job being run with automatic resources. |
| | accelerator_type (Optional[str]): |
| | The type of accelerator(s) that may be attached |
| | to the machine as per `accelerator_count`. Only used if |
| | `machine_type` is set. |
| | accelerator_count (Optional[int]): |
| | The number of accelerators to attach to the |
| | `machine_type`. Only used if `machine_type` is set. |
| | starting_replica_count (Optional[int]): |
| | The number of machine replicas used at the start of the batch |
| | operation. If not set, Vertex AI decides starting number, not |
| | greater than `max_replica_count`. Only used if `machine_type` is |
| | set. |
| | max_replica_count (Optional[int]): |
| | The maximum number of machine replicas the batch operation may |
| | be scaled to. Only used if `machine_type` is set. |
| | Default is 10. |
| | generate_explanation (bool): |
| | Optional. Generate explanation along with the batch prediction |
| | results. This will cause the batch prediction output to include |
| | explanations based on the `prediction_format`: |
| | - `bigquery`: output includes a column named `explanation`. The value |
| | is a struct that conforms to the [aiplatform.gapic.Explanation] object. |
| | - `jsonl`: The JSON objects on each line include an additional entry |
| | keyed `explanation`. The value of the entry is a JSON object that |
| | conforms to the [aiplatform.gapic.Explanation] object. |
| | - `csv`: Generating explanations for CSV format is not supported. |
| | explanation_metadata (aiplatform.explain.ExplanationMetadata): |
| | Optional. Explanation metadata configuration for this BatchPredictionJob. |
| | Can be specified only if `generate_explanation` is set to `True`. |
| | |
| | This value overrides the value of `Model.explanation_metadata`. |
| | All fields of `explanation_metadata` are optional in the request. If |
| | a field of the `explanation_metadata` object is not populated, the |
| | corresponding field of the `Model.explanation_metadata` object is inherited. |
| | For more details, see `Ref docs <http://tinyurl.com/1igh60kt>` |
| | explanation_parameters (aiplatform.explain.ExplanationParameters): |
| | Optional. Parameters to configure explaining for Model's predictions. |
| | Can be specified only if `generate_explanation` is set to `True`. |
| | |
| | This value overrides the value of `Model.explanation_parameters`. |
| | All fields of `explanation_parameters` are optional in the request. If |
| | a field of the `explanation_parameters` object is not populated, the |
| | corresponding field of the `Model.explanation_parameters` object is inherited. |
| | For more details, see `Ref docs <http://tinyurl.com/1an4zake>` |
| | labels (Dict[str, str]): |
| | Optional. The labels with user-defined metadata to organize your |
| | BatchPredictionJobs. Label keys and values can be no longer than |
| | 64 characters (Unicode codepoints), can only contain lowercase |
| | letters, numeric characters, underscores and dashes. |
| | International characters are allowed. See https://goo.gl/xmQnxf |
| | for more information and examples of labels. |
| | credentials (Optional[auth_credentials.Credentials]): |
| | Custom credentials to use to create this batch prediction |
| | job. Overrides credentials set in aiplatform.init. |
| | encryption_spec_key_name (Optional[str]): |
| | Optional. The Cloud KMS resource identifier of the customer |
| | managed encryption key used to protect the job. Has the |
| | form: |
| | ``projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key``. |
| | The key needs to be in the same region as where the compute |
| | resource is created. |
| | |
| | If this is set, then all |
| | resources created by the BatchPredictionJob will |
| | be encrypted with the provided encryption key. |
| | |
| | Overrides encryption_spec_key_name set in aiplatform.init. |
| | sync (bool): |
| | Whether to execute this method synchronously. If False, this method |
| | will be executed in concurrent Future and any downstream object will |
| | be immediately returned and synced when the Future has completed. |
| | create_request_timeout (float): |
| | Optional. The timeout for the create request in seconds. |
| | batch_size (int): |
| | Optional. The number of the records (e.g. instances) of the operation given in each batch |
| | to a machine replica. Machine type, and size of a single record should be considered |
| | when setting this parameter, higher value speeds up the batch operation's execution, |
| | but too high value will result in a whole batch not fitting in a machine's memory, |
| | and the whole operation will fail. |
| | The default value is 64. |
| | model_monitoring_objective_config (aiplatform.model_monitoring.ObjectiveConfig): |
| | Optional. The objective config for model monitoring. Passing this parameter enables |
| | monitoring on the model associated with this batch prediction job. |
| | model_monitoring_alert_config (aiplatform.model_monitoring.EmailAlertConfig): |
| | Optional. Configures how model monitoring alerts are sent to the user. Right now |
| | only email alert is supported. |
| | analysis_instance_schema_uri (str): |
| | Optional. Only applicable if model_monitoring_objective_config is also passed. |
| | This parameter specifies the YAML schema file uri describing the format of a single |
| | instance that you want Tensorflow Data Validation (TFDV) to |
| | analyze. If this field is empty, all the feature data types are |
| | inferred from predict_instance_schema_uri, meaning that TFDV |
| | will use the data in the exact format as prediction request/response. |
| | If there are any data type differences between predict instance |
| | and TFDV instance, this field can be used to override the schema. |
| | For models trained with Vertex AI, this field must be set as all the |
| | fields in predict instance formatted as string. |
| | service_account (str): |
| | Optional. Specifies the service account for workload run-as account. |
| | Users submitting jobs must have act-as permission on this run-as account. |
| | Returns: |
| | (jobs.BatchPredictionJob): |
| | Instantiated representation of the created batch prediction job. |
| | """ |
| | return cls._submit_impl( |
| | job_display_name=job_display_name, |
| | model_name=model_name, |
| | instances_format=instances_format, |
| | predictions_format=predictions_format, |
| | gcs_source=gcs_source, |
| | bigquery_source=bigquery_source, |
| | gcs_destination_prefix=gcs_destination_prefix, |
| | bigquery_destination_prefix=bigquery_destination_prefix, |
| | model_parameters=model_parameters, |
| | machine_type=machine_type, |
| | accelerator_type=accelerator_type, |
| | accelerator_count=accelerator_count, |
| | starting_replica_count=starting_replica_count, |
| | max_replica_count=max_replica_count, |
| | generate_explanation=generate_explanation, |
| | explanation_metadata=explanation_metadata, |
| | explanation_parameters=explanation_parameters, |
| | labels=labels, |
| | project=project, |
| | location=location, |
| | credentials=credentials, |
| | encryption_spec_key_name=encryption_spec_key_name, |
| | sync=sync, |
| | create_request_timeout=create_request_timeout, |
| | batch_size=batch_size, |
| | model_monitoring_objective_config=model_monitoring_objective_config, |
| | model_monitoring_alert_config=model_monitoring_alert_config, |
| | analysis_instance_schema_uri=analysis_instance_schema_uri, |
| | service_account=service_account, |
| | |
| | wait_for_completion=True, |
| | ) |
| |
|
| | @classmethod |
| | def submit( |
| | cls, |
| | *, |
| | job_display_name: Optional[str] = None, |
| | model_name: Union[str, "aiplatform.Model"], |
| | instances_format: str = "jsonl", |
| | predictions_format: str = "jsonl", |
| | gcs_source: Optional[Union[str, Sequence[str]]] = None, |
| | bigquery_source: Optional[str] = None, |
| | gcs_destination_prefix: Optional[str] = None, |
| | bigquery_destination_prefix: Optional[str] = None, |
| | model_parameters: Optional[Dict] = None, |
| | machine_type: Optional[str] = None, |
| | accelerator_type: Optional[str] = None, |
| | accelerator_count: Optional[int] = None, |
| | starting_replica_count: Optional[int] = None, |
| | max_replica_count: Optional[int] = None, |
| | generate_explanation: Optional[bool] = False, |
| | explanation_metadata: Optional["aiplatform.explain.ExplanationMetadata"] = None, |
| | explanation_parameters: Optional[ |
| | "aiplatform.explain.ExplanationParameters" |
| | ] = None, |
| | labels: Optional[Dict[str, str]] = None, |
| | project: Optional[str] = None, |
| | location: Optional[str] = None, |
| | credentials: Optional[auth_credentials.Credentials] = None, |
| | encryption_spec_key_name: Optional[str] = None, |
| | create_request_timeout: Optional[float] = None, |
| | batch_size: Optional[int] = None, |
| | model_monitoring_objective_config: Optional[ |
| | "aiplatform.model_monitoring.ObjectiveConfig" |
| | ] = None, |
| | model_monitoring_alert_config: Optional[ |
| | "aiplatform.model_monitoring.AlertConfig" |
| | ] = None, |
| | analysis_instance_schema_uri: Optional[str] = None, |
| | service_account: Optional[str] = None, |
| | ) -> "BatchPredictionJob": |
| | """Sumbit a batch prediction job (not waiting for completion). |
| | |
| | Args: |
| | job_display_name (str): |
| | Required. The user-defined name of the BatchPredictionJob. |
| | The name can be up to 128 characters long and can be consist |
| | of any UTF-8 characters. |
| | model_name (Union[str, aiplatform.Model]): |
| | Required. A fully-qualified model resource name or model ID. |
| | Example: "projects/123/locations/us-central1/models/456" or |
| | "456" when project and location are initialized or passed. |
| | May optionally contain a version ID or alias in |
| | {model_name}@{version} form. |
| | |
| | Or an instance of aiplatform.Model. |
| | instances_format (str): |
| | Required. The format in which instances are provided. Must be one |
| | of the formats listed in `Model.supported_input_storage_formats`. |
| | Default is "jsonl" when using `gcs_source`. If a `bigquery_source` |
| | is provided, this is overridden to "bigquery". |
| | predictions_format (str): |
| | Required. The format in which Vertex AI outputs the |
| | predictions, must be one of the formats specified in |
| | `Model.supported_output_storage_formats`. |
| | Default is "jsonl" when using `gcs_destination_prefix`. If a |
| | `bigquery_destination_prefix` is provided, this is overridden to |
| | "bigquery". |
| | gcs_source (Optional[Sequence[str]]): |
| | Google Cloud Storage URI(-s) to your instances to run |
| | batch prediction on. They must match `instances_format`. |
| | |
| | bigquery_source (Optional[str]): |
| | BigQuery URI to a table, up to 2000 characters long. For example: |
| | `bq://projectId.bqDatasetId.bqTableId` |
| | gcs_destination_prefix (Optional[str]): |
| | The Google Cloud Storage location of the directory where the |
| | output is to be written to. In the given directory a new |
| | directory is created. Its name is |
| | ``prediction-<model-display-name>-<job-create-time>``, where |
| | timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format. |
| | Inside of it files ``predictions_0001.<extension>``, |
| | ``predictions_0002.<extension>``, ..., |
| | ``predictions_N.<extension>`` are created where |
| | ``<extension>`` depends on chosen ``predictions_format``, |
| | and N may equal 0001 and depends on the total number of |
| | successfully predicted instances. If the Model has both |
| | ``instance`` and ``prediction`` schemata defined then each such |
| | file contains predictions as per the ``predictions_format``. |
| | If prediction for any instance failed (partially or |
| | completely), then an additional ``errors_0001.<extension>``, |
| | ``errors_0002.<extension>``,..., ``errors_N.<extension>`` |
| | files are created (N depends on total number of failed |
| | predictions). These files contain the failed instances, as |
| | per their schema, followed by an additional ``error`` field |
| | which as value has ```google.rpc.Status`` <Status>`__ |
| | containing only ``code`` and ``message`` fields. |
| | bigquery_destination_prefix (Optional[str]): |
| | The BigQuery project or dataset location where the output is |
| | to be written to. If project is provided, a new dataset is |
| | created with name |
| | ``prediction_<model-display-name>_<job-create-time>`` where |
| | is made BigQuery-dataset-name compatible (for example, most |
| | special characters become underscores), and timestamp is in |
| | YYYY_MM_DDThh_mm_ss_sssZ "based on ISO-8601" format. In the |
| | dataset two tables will be created, ``predictions``, and |
| | ``errors``. If the Model has both |
| | [instance][google.cloud.aiplatform.v1.PredictSchemata.instance_schema_uri] |
| | and |
| | [prediction][google.cloud.aiplatform.v1.PredictSchemata.parameters_schema_uri] |
| | schemata defined then the tables have columns as follows: |
| | The ``predictions`` table contains instances for which the |
| | prediction succeeded, it has columns as per a concatenation |
| | of the Model's instance and prediction schemata. The |
| | ``errors`` table contains rows for which the prediction has |
| | failed, it has instance columns, as per the instance schema, |
| | followed by a single "errors" column, which as values has |
| | [google.rpc.Status][google.rpc.Status] represented as a |
| | STRUCT, and containing only ``code`` and ``message``. |
| | model_parameters (Optional[Dict]): |
| | The parameters that govern the predictions. The schema of |
| | the parameters may be specified via the Model's `parameters_schema_uri`. |
| | machine_type (Optional[str]): |
| | The type of machine for running batch prediction on |
| | dedicated resources. Not specifying machine type will result in |
| | batch prediction job being run with automatic resources. |
| | accelerator_type (Optional[str]): |
| | The type of accelerator(s) that may be attached |
| | to the machine as per `accelerator_count`. Only used if |
| | `machine_type` is set. |
| | accelerator_count (Optional[int]): |
| | The number of accelerators to attach to the |
| | `machine_type`. Only used if `machine_type` is set. |
| | starting_replica_count (Optional[int]): |
| | The number of machine replicas used at the start of the batch |
| | operation. If not set, Vertex AI decides starting number, not |
| | greater than `max_replica_count`. Only used if `machine_type` is |
| | set. |
| | max_replica_count (Optional[int]): |
| | The maximum number of machine replicas the batch operation may |
| | be scaled to. Only used if `machine_type` is set. |
| | Default is 10. |
| | generate_explanation (bool): |
| | Optional. Generate explanation along with the batch prediction |
| | results. This will cause the batch prediction output to include |
| | explanations based on the `prediction_format`: |
| | - `bigquery`: output includes a column named `explanation`. The value |
| | is a struct that conforms to the [aiplatform.gapic.Explanation] object. |
| | - `jsonl`: The JSON objects on each line include an additional entry |
| | keyed `explanation`. The value of the entry is a JSON object that |
| | conforms to the [aiplatform.gapic.Explanation] object. |
| | - `csv`: Generating explanations for CSV format is not supported. |
| | explanation_metadata (aiplatform.explain.ExplanationMetadata): |
| | Optional. Explanation metadata configuration for this BatchPredictionJob. |
| | Can be specified only if `generate_explanation` is set to `True`. |
| | |
| | This value overrides the value of `Model.explanation_metadata`. |
| | All fields of `explanation_metadata` are optional in the request. If |
| | a field of the `explanation_metadata` object is not populated, the |
| | corresponding field of the `Model.explanation_metadata` object is inherited. |
| | For more details, see `Ref docs <http://tinyurl.com/1igh60kt>` |
| | explanation_parameters (aiplatform.explain.ExplanationParameters): |
| | Optional. Parameters to configure explaining for Model's predictions. |
| | Can be specified only if `generate_explanation` is set to `True`. |
| | |
| | This value overrides the value of `Model.explanation_parameters`. |
| | All fields of `explanation_parameters` are optional in the request. If |
| | a field of the `explanation_parameters` object is not populated, the |
| | corresponding field of the `Model.explanation_parameters` object is inherited. |
| | For more details, see `Ref docs <http://tinyurl.com/1an4zake>` |
| | labels (Dict[str, str]): |
| | Optional. The labels with user-defined metadata to organize your |
| | BatchPredictionJobs. Label keys and values can be no longer than |
| | 64 characters (Unicode codepoints), can only contain lowercase |
| | letters, numeric characters, underscores and dashes. |
| | International characters are allowed. See https://goo.gl/xmQnxf |
| | for more information and examples of labels. |
| | credentials (Optional[auth_credentials.Credentials]): |
| | Custom credentials to use to create this batch prediction |
| | job. Overrides credentials set in aiplatform.init. |
| | encryption_spec_key_name (Optional[str]): |
| | Optional. The Cloud KMS resource identifier of the customer |
| | managed encryption key used to protect the job. Has the |
| | form: |
| | ``projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key``. |
| | The key needs to be in the same region as where the compute |
| | resource is created. |
| | |
| | If this is set, then all |
| | resources created by the BatchPredictionJob will |
| | be encrypted with the provided encryption key. |
| | |
| | Overrides encryption_spec_key_name set in aiplatform.init. |
| | create_request_timeout (float): |
| | Optional. The timeout for the create request in seconds. |
| | batch_size (int): |
| | Optional. The number of the records (e.g. instances) of the operation given in each batch |
| | to a machine replica. Machine type, and size of a single record should be considered |
| | when setting this parameter, higher value speeds up the batch operation's execution, |
| | but too high value will result in a whole batch not fitting in a machine's memory, |
| | and the whole operation will fail. |
| | The default value is 64. |
| | model_monitoring_objective_config (aiplatform.model_monitoring.ObjectiveConfig): |
| | Optional. The objective config for model monitoring. Passing this parameter enables |
| | monitoring on the model associated with this batch prediction job. |
| | model_monitoring_alert_config (aiplatform.model_monitoring.EmailAlertConfig): |
| | Optional. Configures how model monitoring alerts are sent to the user. Right now |
| | only email alert is supported. |
| | analysis_instance_schema_uri (str): |
| | Optional. Only applicable if model_monitoring_objective_config is also passed. |
| | This parameter specifies the YAML schema file uri describing the format of a single |
| | instance that you want Tensorflow Data Validation (TFDV) to |
| | analyze. If this field is empty, all the feature data types are |
| | inferred from predict_instance_schema_uri, meaning that TFDV |
| | will use the data in the exact format as prediction request/response. |
| | If there are any data type differences between predict instance |
| | and TFDV instance, this field can be used to override the schema. |
| | For models trained with Vertex AI, this field must be set as all the |
| | fields in predict instance formatted as string. |
| | service_account (str): |
| | Optional. Specifies the service account for workload run-as account. |
| | Users submitting jobs must have act-as permission on this run-as account. |
| | Returns: |
| | (jobs.BatchPredictionJob): |
| | Instantiated representation of the created batch prediction job. |
| | """ |
| | return cls._submit_impl( |
| | job_display_name=job_display_name, |
| | model_name=model_name, |
| | instances_format=instances_format, |
| | predictions_format=predictions_format, |
| | gcs_source=gcs_source, |
| | bigquery_source=bigquery_source, |
| | gcs_destination_prefix=gcs_destination_prefix, |
| | bigquery_destination_prefix=bigquery_destination_prefix, |
| | model_parameters=model_parameters, |
| | machine_type=machine_type, |
| | accelerator_type=accelerator_type, |
| | accelerator_count=accelerator_count, |
| | starting_replica_count=starting_replica_count, |
| | max_replica_count=max_replica_count, |
| | generate_explanation=generate_explanation, |
| | explanation_metadata=explanation_metadata, |
| | explanation_parameters=explanation_parameters, |
| | labels=labels, |
| | project=project, |
| | location=location, |
| | credentials=credentials, |
| | encryption_spec_key_name=encryption_spec_key_name, |
| | create_request_timeout=create_request_timeout, |
| | batch_size=batch_size, |
| | model_monitoring_objective_config=model_monitoring_objective_config, |
| | model_monitoring_alert_config=model_monitoring_alert_config, |
| | analysis_instance_schema_uri=analysis_instance_schema_uri, |
| | service_account=service_account, |
| | |
| | wait_for_completion=False, |
| | sync=True, |
| | ) |
| |
|
| | @classmethod |
| | def _submit_impl( |
| | cls, |
| | *, |
| | job_display_name: Optional[str] = None, |
| | model_name: Union[str, "aiplatform.Model"], |
| | instances_format: str = "jsonl", |
| | predictions_format: str = "jsonl", |
| | gcs_source: Optional[Union[str, Sequence[str]]] = None, |
| | bigquery_source: Optional[str] = None, |
| | gcs_destination_prefix: Optional[str] = None, |
| | bigquery_destination_prefix: Optional[str] = None, |
| | model_parameters: Optional[Dict] = None, |
| | machine_type: Optional[str] = None, |
| | accelerator_type: Optional[str] = None, |
| | accelerator_count: Optional[int] = None, |
| | starting_replica_count: Optional[int] = None, |
| | max_replica_count: Optional[int] = None, |
| | generate_explanation: Optional[bool] = False, |
| | explanation_metadata: Optional["aiplatform.explain.ExplanationMetadata"] = None, |
| | explanation_parameters: Optional[ |
| | "aiplatform.explain.ExplanationParameters" |
| | ] = None, |
| | labels: Optional[Dict[str, str]] = None, |
| | project: Optional[str] = None, |
| | location: Optional[str] = None, |
| | credentials: Optional[auth_credentials.Credentials] = None, |
| | encryption_spec_key_name: Optional[str] = None, |
| | sync: bool = True, |
| | create_request_timeout: Optional[float] = None, |
| | batch_size: Optional[int] = None, |
| | model_monitoring_objective_config: Optional[ |
| | "aiplatform.model_monitoring.ObjectiveConfig" |
| | ] = None, |
| | model_monitoring_alert_config: Optional[ |
| | "aiplatform.model_monitoring.AlertConfig" |
| | ] = None, |
| | analysis_instance_schema_uri: Optional[str] = None, |
| | service_account: Optional[str] = None, |
| | wait_for_completion: bool = False, |
| | ) -> "BatchPredictionJob": |
| | """Create a batch prediction job. |
| | |
| | Args: |
| | job_display_name (str): |
| | Required. The user-defined name of the BatchPredictionJob. |
| | The name can be up to 128 characters long and can be consist |
| | of any UTF-8 characters. |
| | model_name (Union[str, aiplatform.Model]): |
| | Required. A fully-qualified model resource name or model ID. |
| | Example: "projects/123/locations/us-central1/models/456" or |
| | "456" when project and location are initialized or passed. |
| | May optionally contain a version ID or alias in |
| | {model_name}@{version} form. |
| | |
| | Or an instance of aiplatform.Model. |
| | instances_format (str): |
| | Required. The format in which instances are provided. Must be one |
| | of the formats listed in `Model.supported_input_storage_formats`. |
| | Default is "jsonl" when using `gcs_source`. If a `bigquery_source` |
| | is provided, this is overridden to "bigquery". |
| | predictions_format (str): |
| | Required. The format in which Vertex AI outputs the |
| | predictions, must be one of the formats specified in |
| | `Model.supported_output_storage_formats`. |
| | Default is "jsonl" when using `gcs_destination_prefix`. If a |
| | `bigquery_destination_prefix` is provided, this is overridden to |
| | "bigquery". |
| | gcs_source (Optional[Sequence[str]]): |
| | Google Cloud Storage URI(-s) to your instances to run |
| | batch prediction on. They must match `instances_format`. |
| | |
| | bigquery_source (Optional[str]): |
| | BigQuery URI to a table, up to 2000 characters long. For example: |
| | `bq://projectId.bqDatasetId.bqTableId` |
| | gcs_destination_prefix (Optional[str]): |
| | The Google Cloud Storage location of the directory where the |
| | output is to be written to. In the given directory a new |
| | directory is created. Its name is |
| | ``prediction-<model-display-name>-<job-create-time>``, where |
| | timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format. |
| | Inside of it files ``predictions_0001.<extension>``, |
| | ``predictions_0002.<extension>``, ..., |
| | ``predictions_N.<extension>`` are created where |
| | ``<extension>`` depends on chosen ``predictions_format``, |
| | and N may equal 0001 and depends on the total number of |
| | successfully predicted instances. If the Model has both |
| | ``instance`` and ``prediction`` schemata defined then each such |
| | file contains predictions as per the ``predictions_format``. |
| | If prediction for any instance failed (partially or |
| | completely), then an additional ``errors_0001.<extension>``, |
| | ``errors_0002.<extension>``,..., ``errors_N.<extension>`` |
| | files are created (N depends on total number of failed |
| | predictions). These files contain the failed instances, as |
| | per their schema, followed by an additional ``error`` field |
| | which as value has ```google.rpc.Status`` <Status>`__ |
| | containing only ``code`` and ``message`` fields. |
| | bigquery_destination_prefix (Optional[str]): |
| | The BigQuery project or dataset location where the output is |
| | to be written to. If project is provided, a new dataset is |
| | created with name |
| | ``prediction_<model-display-name>_<job-create-time>`` where |
| | is made BigQuery-dataset-name compatible (for example, most |
| | special characters become underscores), and timestamp is in |
| | YYYY_MM_DDThh_mm_ss_sssZ "based on ISO-8601" format. In the |
| | dataset two tables will be created, ``predictions``, and |
| | ``errors``. If the Model has both |
| | [instance][google.cloud.aiplatform.v1.PredictSchemata.instance_schema_uri] |
| | and |
| | [prediction][google.cloud.aiplatform.v1.PredictSchemata.parameters_schema_uri] |
| | schemata defined then the tables have columns as follows: |
| | The ``predictions`` table contains instances for which the |
| | prediction succeeded, it has columns as per a concatenation |
| | of the Model's instance and prediction schemata. The |
| | ``errors`` table contains rows for which the prediction has |
| | failed, it has instance columns, as per the instance schema, |
| | followed by a single "errors" column, which as values has |
| | [google.rpc.Status][google.rpc.Status] represented as a |
| | STRUCT, and containing only ``code`` and ``message``. |
| | model_parameters (Optional[Dict]): |
| | The parameters that govern the predictions. The schema of |
| | the parameters may be specified via the Model's `parameters_schema_uri`. |
| | machine_type (Optional[str]): |
| | The type of machine for running batch prediction on |
| | dedicated resources. Not specifying machine type will result in |
| | batch prediction job being run with automatic resources. |
| | accelerator_type (Optional[str]): |
| | The type of accelerator(s) that may be attached |
| | to the machine as per `accelerator_count`. Only used if |
| | `machine_type` is set. |
| | accelerator_count (Optional[int]): |
| | The number of accelerators to attach to the |
| | `machine_type`. Only used if `machine_type` is set. |
| | starting_replica_count (Optional[int]): |
| | The number of machine replicas used at the start of the batch |
| | operation. If not set, Vertex AI decides starting number, not |
| | greater than `max_replica_count`. Only used if `machine_type` is |
| | set. |
| | max_replica_count (Optional[int]): |
| | The maximum number of machine replicas the batch operation may |
| | be scaled to. Only used if `machine_type` is set. |
| | Default is 10. |
| | generate_explanation (bool): |
| | Optional. Generate explanation along with the batch prediction |
| | results. This will cause the batch prediction output to include |
| | explanations based on the `prediction_format`: |
| | - `bigquery`: output includes a column named `explanation`. The value |
| | is a struct that conforms to the [aiplatform.gapic.Explanation] object. |
| | - `jsonl`: The JSON objects on each line include an additional entry |
| | keyed `explanation`. The value of the entry is a JSON object that |
| | conforms to the [aiplatform.gapic.Explanation] object. |
| | - `csv`: Generating explanations for CSV format is not supported. |
| | explanation_metadata (aiplatform.explain.ExplanationMetadata): |
| | Optional. Explanation metadata configuration for this BatchPredictionJob. |
| | Can be specified only if `generate_explanation` is set to `True`. |
| | |
| | This value overrides the value of `Model.explanation_metadata`. |
| | All fields of `explanation_metadata` are optional in the request. If |
| | a field of the `explanation_metadata` object is not populated, the |
| | corresponding field of the `Model.explanation_metadata` object is inherited. |
| | For more details, see `Ref docs <http://tinyurl.com/1igh60kt>` |
| | explanation_parameters (aiplatform.explain.ExplanationParameters): |
| | Optional. Parameters to configure explaining for Model's predictions. |
| | Can be specified only if `generate_explanation` is set to `True`. |
| | |
| | This value overrides the value of `Model.explanation_parameters`. |
| | All fields of `explanation_parameters` are optional in the request. If |
| | a field of the `explanation_parameters` object is not populated, the |
| | corresponding field of the `Model.explanation_parameters` object is inherited. |
| | For more details, see `Ref docs <http://tinyurl.com/1an4zake>` |
| | labels (Dict[str, str]): |
| | Optional. The labels with user-defined metadata to organize your |
| | BatchPredictionJobs. Label keys and values can be no longer than |
| | 64 characters (Unicode codepoints), can only contain lowercase |
| | letters, numeric characters, underscores and dashes. |
| | International characters are allowed. See https://goo.gl/xmQnxf |
| | for more information and examples of labels. |
| | credentials (Optional[auth_credentials.Credentials]): |
| | Custom credentials to use to create this batch prediction |
| | job. Overrides credentials set in aiplatform.init. |
| | encryption_spec_key_name (Optional[str]): |
| | Optional. The Cloud KMS resource identifier of the customer |
| | managed encryption key used to protect the job. Has the |
| | form: |
| | ``projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key``. |
| | The key needs to be in the same region as where the compute |
| | resource is created. |
| | |
| | If this is set, then all |
| | resources created by the BatchPredictionJob will |
| | be encrypted with the provided encryption key. |
| | |
| | Overrides encryption_spec_key_name set in aiplatform.init. |
| | sync (bool): |
| | Whether to execute this method synchronously. If False, this method |
| | will be executed in concurrent Future and any downstream object will |
| | be immediately returned and synced when the Future has completed. |
| | create_request_timeout (float): |
| | Optional. The timeout for the create request in seconds. |
| | batch_size (int): |
| | Optional. The number of the records (e.g. instances) of the operation given in each batch |
| | to a machine replica. Machine type, and size of a single record should be considered |
| | when setting this parameter, higher value speeds up the batch operation's execution, |
| | but too high value will result in a whole batch not fitting in a machine's memory, |
| | and the whole operation will fail. |
| | The default value is 64. |
| | model_monitoring_objective_config (aiplatform.model_monitoring.ObjectiveConfig): |
| | Optional. The objective config for model monitoring. Passing this parameter enables |
| | monitoring on the model associated with this batch prediction job. |
| | model_monitoring_alert_config (aiplatform.model_monitoring.EmailAlertConfig): |
| | Optional. Configures how model monitoring alerts are sent to the user. Right now |
| | only email alert is supported. |
| | analysis_instance_schema_uri (str): |
| | Optional. Only applicable if model_monitoring_objective_config is also passed. |
| | This parameter specifies the YAML schema file uri describing the format of a single |
| | instance that you want Tensorflow Data Validation (TFDV) to |
| | analyze. If this field is empty, all the feature data types are |
| | inferred from predict_instance_schema_uri, meaning that TFDV |
| | will use the data in the exact format as prediction request/response. |
| | If there are any data type differences between predict instance |
| | and TFDV instance, this field can be used to override the schema. |
| | For models trained with Vertex AI, this field must be set as all the |
| | fields in predict instance formatted as string. |
| | service_account (str): |
| | Optional. Specifies the service account for workload run-as account. |
| | Users submitting jobs must have act-as permission on this run-as account. |
| | wait_for_completion (bool): |
| | Whether to wait for the job completion. |
| | Returns: |
| | (jobs.BatchPredictionJob): |
| | Instantiated representation of the created batch prediction job. |
| | """ |
| | |
| | if model_monitoring_objective_config: |
| | from google.cloud.aiplatform.compat.types import ( |
| | batch_prediction_job_v1beta1 as gca_bp_job_compat, |
| | io_v1beta1 as gca_io_compat, |
| | explanation_v1beta1 as gca_explanation_v1beta1, |
| | machine_resources_v1beta1 as gca_machine_resources_compat, |
| | manual_batch_tuning_parameters_v1beta1 as gca_manual_batch_tuning_parameters_compat, |
| | ) |
| | else: |
| | from google.cloud.aiplatform.compat.types import ( |
| | batch_prediction_job as gca_bp_job_compat, |
| | io as gca_io_compat, |
| | explanation as gca_explanation_v1beta1, |
| | machine_resources as gca_machine_resources_compat, |
| | manual_batch_tuning_parameters as gca_manual_batch_tuning_parameters_compat, |
| | ) |
| | if not job_display_name: |
| | job_display_name = cls._generate_display_name() |
| |
|
| | utils.validate_display_name(job_display_name) |
| |
|
| | if labels: |
| | utils.validate_labels(labels) |
| |
|
| | if isinstance(model_name, str): |
| | try: |
| | model_name = utils.full_resource_name( |
| | resource_name=model_name, |
| | resource_noun="models", |
| | parse_resource_name_method=aiplatform.Model._parse_resource_name, |
| | format_resource_name_method=aiplatform.Model._format_resource_name, |
| | project=project, |
| | location=location, |
| | resource_id_validator=super()._revisioned_resource_id_validator, |
| | ) |
| | except ValueError: |
| | |
| | if not _publisher_models._PublisherModel._parse_resource_name( |
| | model_name |
| | ): |
| | raise |
| |
|
| | |
| | if bool(gcs_source) == bool(bigquery_source): |
| | raise ValueError( |
| | "Please provide either a gcs_source or bigquery_source, " |
| | "but not both." |
| | ) |
| |
|
| | |
| | if bool(gcs_destination_prefix) == bool(bigquery_destination_prefix): |
| | raise ValueError( |
| | "Please provide either a gcs_destination_prefix or " |
| | "bigquery_destination_prefix, but not both." |
| | ) |
| |
|
| | |
| | if instances_format not in constants.BATCH_PREDICTION_INPUT_STORAGE_FORMATS: |
| | raise ValueError( |
| | f"{predictions_format} is not an accepted instances format " |
| | f"type. Please choose from: {constants.BATCH_PREDICTION_INPUT_STORAGE_FORMATS}" |
| | ) |
| |
|
| | |
| | if predictions_format not in constants.BATCH_PREDICTION_OUTPUT_STORAGE_FORMATS: |
| | raise ValueError( |
| | f"{predictions_format} is not an accepted prediction format " |
| | f"type. Please choose from: {constants.BATCH_PREDICTION_OUTPUT_STORAGE_FORMATS}" |
| | ) |
| |
|
| | gapic_batch_prediction_job = gca_bp_job_compat.BatchPredictionJob() |
| |
|
| | |
| | gapic_batch_prediction_job.display_name = job_display_name |
| |
|
| | input_config = gca_bp_job_compat.BatchPredictionJob.InputConfig() |
| | output_config = gca_bp_job_compat.BatchPredictionJob.OutputConfig() |
| |
|
| | if bigquery_source: |
| | input_config.instances_format = "bigquery" |
| | input_config.bigquery_source = gca_io_compat.BigQuerySource() |
| | input_config.bigquery_source.input_uri = bigquery_source |
| | else: |
| | input_config.instances_format = instances_format |
| | input_config.gcs_source = gca_io_compat.GcsSource( |
| | uris=gcs_source if isinstance(gcs_source, list) else [gcs_source] |
| | ) |
| |
|
| | if bigquery_destination_prefix: |
| | output_config.predictions_format = "bigquery" |
| | output_config.bigquery_destination = gca_io_compat.BigQueryDestination() |
| |
|
| | bq_dest_prefix = bigquery_destination_prefix |
| |
|
| | if not bq_dest_prefix.startswith("bq://"): |
| | bq_dest_prefix = f"bq://{bq_dest_prefix}" |
| |
|
| | output_config.bigquery_destination.output_uri = bq_dest_prefix |
| | else: |
| | output_config.predictions_format = predictions_format |
| | output_config.gcs_destination = gca_io_compat.GcsDestination( |
| | output_uri_prefix=gcs_destination_prefix |
| | ) |
| |
|
| | gapic_batch_prediction_job.input_config = input_config |
| | gapic_batch_prediction_job.output_config = output_config |
| |
|
| | |
| | gapic_batch_prediction_job.encryption_spec = ( |
| | initializer.global_config.get_encryption_spec( |
| | encryption_spec_key_name=encryption_spec_key_name |
| | ) |
| | ) |
| |
|
| | if model_parameters: |
| | gapic_batch_prediction_job.model_parameters = model_parameters |
| |
|
| | |
| | if machine_type: |
| |
|
| | machine_spec = gca_machine_resources_compat.MachineSpec() |
| | machine_spec.machine_type = machine_type |
| | machine_spec.accelerator_type = accelerator_type |
| | machine_spec.accelerator_count = accelerator_count |
| |
|
| | dedicated_resources = gca_machine_resources_compat.BatchDedicatedResources() |
| |
|
| | dedicated_resources.machine_spec = machine_spec |
| | dedicated_resources.starting_replica_count = starting_replica_count |
| | dedicated_resources.max_replica_count = max_replica_count |
| |
|
| | gapic_batch_prediction_job.dedicated_resources = dedicated_resources |
| |
|
| | manual_batch_tuning_parameters = ( |
| | gca_manual_batch_tuning_parameters_compat.ManualBatchTuningParameters() |
| | ) |
| | manual_batch_tuning_parameters.batch_size = batch_size |
| |
|
| | gapic_batch_prediction_job.manual_batch_tuning_parameters = ( |
| | manual_batch_tuning_parameters |
| | ) |
| |
|
| | |
| | gapic_batch_prediction_job.labels = labels |
| |
|
| | |
| | if generate_explanation: |
| | gapic_batch_prediction_job.generate_explanation = generate_explanation |
| |
|
| | if explanation_metadata or explanation_parameters: |
| | explanation_spec = gca_explanation_compat.ExplanationSpec( |
| | metadata=explanation_metadata, parameters=explanation_parameters |
| | ) |
| | |
| | if model_monitoring_objective_config: |
| |
|
| | explanation_spec = gca_explanation_v1beta1.ExplanationSpec.deserialize( |
| | gca_explanation_compat.ExplanationSpec.serialize(explanation_spec) |
| | ) |
| | gapic_batch_prediction_job.explanation_spec = explanation_spec |
| |
|
| | service_account = service_account or initializer.global_config.service_account |
| | if service_account: |
| | gapic_batch_prediction_job.service_account = service_account |
| |
|
| | empty_batch_prediction_job = cls._empty_constructor( |
| | project=project, |
| | location=location, |
| | credentials=credentials, |
| | ) |
| | if model_monitoring_objective_config: |
| | empty_batch_prediction_job.api_client = ( |
| | empty_batch_prediction_job.api_client.select_version("v1beta1") |
| | ) |
| |
|
| | |
| | if model_monitoring_objective_config: |
| | model_monitoring_objective_config._config_for_bp = True |
| | if model_monitoring_alert_config is not None: |
| | model_monitoring_alert_config._config_for_bp = True |
| | gapic_mm_config = gca_model_monitoring_v1beta1.ModelMonitoringConfig( |
| | objective_configs=[model_monitoring_objective_config.as_proto()], |
| | alert_config=model_monitoring_alert_config.as_proto() |
| | if model_monitoring_alert_config is not None |
| | else None, |
| | analysis_instance_schema_uri=analysis_instance_schema_uri |
| | if analysis_instance_schema_uri is not None |
| | else None, |
| | ) |
| | gapic_batch_prediction_job.model_monitoring_config = gapic_mm_config |
| |
|
| | |
| | return cls._submit_and_optionally_wait_with_sync_support( |
| | empty_batch_prediction_job=empty_batch_prediction_job, |
| | model_or_model_name=model_name, |
| | gca_batch_prediction_job=gapic_batch_prediction_job, |
| | generate_explanation=generate_explanation, |
| | sync=sync, |
| | create_request_timeout=create_request_timeout, |
| | wait_for_completion=wait_for_completion, |
| | ) |
| |
|
| | @classmethod |
| | @base.optional_sync(return_input_arg="empty_batch_prediction_job") |
| | def _submit_and_optionally_wait_with_sync_support( |
| | cls, |
| | empty_batch_prediction_job: "BatchPredictionJob", |
| | model_or_model_name: Union[str, "aiplatform.Model"], |
| | gca_batch_prediction_job: gca_bp_job_compat.BatchPredictionJob, |
| | generate_explanation: bool, |
| | sync: bool = True, |
| | create_request_timeout: Optional[float] = None, |
| | wait_for_completion: bool = True, |
| | ) -> "BatchPredictionJob": |
| | """Create a batch prediction job. |
| | |
| | Args: |
| | empty_batch_prediction_job (BatchPredictionJob): |
| | Required. BatchPredictionJob without _gca_resource populated. |
| | model_or_model_name (Union[str, aiplatform.Model]): |
| | Required. Required. A fully-qualified model resource name or |
| | an instance of aiplatform.Model. If a resource name, it may |
| | optionally contain a version ID or alias in |
| | {model_name}@{version} form. |
| | gca_batch_prediction_job (gca_bp_job.BatchPredictionJob): |
| | Required. a batch prediction job proto for creating a batch prediction job on Vertex AI. |
| | generate_explanation (bool): |
| | Required. Generate explanation along with the batch prediction |
| | results. |
| | create_request_timeout (float): |
| | Optional. The timeout for the create request in seconds. |
| | wait_for_completion (bool): |
| | Whether to wait for the job completion. |
| | Returns: |
| | (jobs.BatchPredictionJob): |
| | Instantiated representation of the created batch prediction job. |
| | |
| | Raises: |
| | ValueError: |
| | If no or multiple source or destinations are provided. Also, if |
| | provided instances_format or predictions_format are not supported |
| | by Vertex AI. |
| | """ |
| |
|
| | parent = initializer.global_config.common_location_path( |
| | project=empty_batch_prediction_job.project, |
| | location=empty_batch_prediction_job.location, |
| | ) |
| |
|
| | model_resource_name = ( |
| | model_or_model_name |
| | if isinstance(model_or_model_name, str) |
| | else model_or_model_name.versioned_resource_name |
| | ) |
| |
|
| | gca_batch_prediction_job.model = model_resource_name |
| |
|
| | api_client = empty_batch_prediction_job.api_client |
| |
|
| | _LOGGER.log_create_with_lro(cls) |
| |
|
| | gca_batch_prediction_job = api_client.create_batch_prediction_job( |
| | parent=parent, |
| | batch_prediction_job=gca_batch_prediction_job, |
| | timeout=create_request_timeout, |
| | ) |
| |
|
| | empty_batch_prediction_job._gca_resource = gca_batch_prediction_job |
| |
|
| | batch_prediction_job = empty_batch_prediction_job |
| |
|
| | _LOGGER.log_create_complete(cls, batch_prediction_job._gca_resource, "bpj") |
| |
|
| | _LOGGER.info( |
| | "View Batch Prediction Job:\n%s" % batch_prediction_job._dashboard_uri() |
| | ) |
| | if wait_for_completion: |
| | batch_prediction_job._block_until_complete() |
| |
|
| | return batch_prediction_job |
| |
|
| | def iter_outputs( |
| | self, bq_max_results: Optional[int] = 100 |
| | ) -> Union[ |
| | Iterable["storage.Blob"], Iterable["bigquery.table.RowIterator"] |
| | ]: |
| | """Returns an Iterable object to traverse the output files, either a |
| | list of GCS Blobs or a BigQuery RowIterator depending on the output |
| | config set when the BatchPredictionJob was created. |
| | |
| | Args: |
| | bq_max_results: Optional[int] = 100 |
| | Limit on rows to retrieve from prediction table in BigQuery dataset. |
| | Only used when retrieving predictions from a bigquery_destination_prefix. |
| | Default is 100. |
| | |
| | Returns: |
| | Union[Iterable[storage.Blob], Iterable[bigquery.table.RowIterator]]: |
| | Either a list of GCS Blob objects within the prediction output |
| | directory or an iterable BigQuery RowIterator with predictions. |
| | |
| | Raises: |
| | RuntimeError: |
| | If BatchPredictionJob is in a JobState other than SUCCEEDED, |
| | since outputs cannot be retrieved until the Job has finished. |
| | NotImplementedError: |
| | If BatchPredictionJob succeeded and output_info does not have a |
| | GCS or BQ output provided. |
| | """ |
| | |
| | from google.cloud import bigquery |
| | from google.cloud import storage |
| |
|
| | self._assert_gca_resource_is_available() |
| |
|
| | if self.state != gca_job_state.JobState.JOB_STATE_SUCCEEDED: |
| | raise RuntimeError( |
| | f"Cannot read outputs until BatchPredictionJob has succeeded, " |
| | f"current state: {self._gca_resource.state}" |
| | ) |
| |
|
| | output_info = self._gca_resource.output_info |
| |
|
| | |
| | if output_info.gcs_output_directory: |
| |
|
| | |
| | storage_client = storage.Client( |
| | project=self.project, |
| | credentials=self.api_client._transport._credentials, |
| | ) |
| |
|
| | gcs_bucket, gcs_prefix = utils.extract_bucket_and_prefix_from_gcs_path( |
| | output_info.gcs_output_directory |
| | ) |
| |
|
| | blobs = storage_client.list_blobs(gcs_bucket, prefix=gcs_prefix) |
| |
|
| | return blobs |
| |
|
| | |
| | elif output_info.bigquery_output_dataset: |
| |
|
| | |
| | bq_dataset = output_info.bigquery_output_dataset |
| | bq_table = output_info.bigquery_output_table |
| |
|
| | if not bq_table: |
| | raise RuntimeError( |
| | "A BigQuery table with predictions was not found, this " |
| | f"might be due to errors. Visit {self._dashboard_uri()} for details." |
| | ) |
| |
|
| | if bq_dataset.startswith("bq://"): |
| | bq_dataset = bq_dataset[5:] |
| |
|
| | |
| | bq_client = bigquery.Client( |
| | project=self.project, |
| | credentials=self.api_client._transport._credentials, |
| | ) |
| |
|
| | row_iterator = bq_client.list_rows( |
| | table=f"{bq_dataset}.{bq_table}", max_results=bq_max_results |
| | ) |
| |
|
| | return row_iterator |
| |
|
| | |
| | else: |
| | raise NotImplementedError( |
| | f"Unsupported batch prediction output location, here are details" |
| | f"on your prediction output:\n{output_info}" |
| | ) |
| |
|
| | def wait_for_resource_creation(self) -> None: |
| | """Waits until resource has been created.""" |
| | self._wait_for_resource_creation() |
| |
|
| |
|
| | class _RunnableJob(_Job): |
| | """ABC to interface job as a runnable training class.""" |
| |
|
| | def __init__( |
| | self, |
| | project: Optional[str] = None, |
| | location: Optional[str] = None, |
| | credentials: Optional[auth_credentials.Credentials] = None, |
| | ): |
| | """Initializes job with project, location, and api_client. |
| | |
| | Args: |
| | project(str): Project of the resource noun. |
| | location(str): The location of the resource noun. |
| | credentials(google.auth.credentials.Credentials): Optional. custom |
| | credentials to use when accessing interacting with resource noun. |
| | """ |
| |
|
| | base.VertexAiResourceNounWithFutureManager.__init__( |
| | self, project=project, location=location, credentials=credentials |
| | ) |
| |
|
| | self._parent = aiplatform.initializer.global_config.common_location_path( |
| | project=project, location=location |
| | ) |
| |
|
| | self._logged_web_access_uris = set() |
| |
|
| | @classmethod |
| | def _empty_constructor( |
| | cls, |
| | project: Optional[str] = None, |
| | location: Optional[str] = None, |
| | credentials: Optional[auth_credentials.Credentials] = None, |
| | resource_name: Optional[str] = None, |
| | ) -> "_RunnableJob": |
| | """Initializes with all attributes set to None. |
| | |
| | The attributes should be populated after a future is complete. This allows |
| | scheduling of additional API calls before the resource is created. |
| | |
| | Args: |
| | project (str): Optional. Project of the resource noun. |
| | location (str): Optional. The location of the resource noun. |
| | credentials(google.auth.credentials.Credentials): |
| | Optional. custom credentials to use when accessing interacting with |
| | resource noun. |
| | resource_name(str): Optional. A fully-qualified resource name or ID. |
| | Returns: |
| | An instance of this class with attributes set to None. |
| | """ |
| | self = super()._empty_constructor( |
| | project=project, |
| | location=location, |
| | credentials=credentials, |
| | resource_name=resource_name, |
| | ) |
| |
|
| | self._logged_web_access_uris = set() |
| |
|
| | if isinstance(self, CustomJob): |
| | self._enable_autolog = False |
| |
|
| | return self |
| |
|
| | @property |
| | def web_access_uris(self) -> Dict[str, Union[str, Dict[str, str]]]: |
| | """Fetch the runnable job again and return the latest web access uris. |
| | |
| | Returns: |
| | (Dict[str, Union[str, Dict[str, str]]]): |
| | Web access uris of the runnable job. |
| | """ |
| |
|
| | |
| | self._sync_gca_resource() |
| | return self._get_web_access_uris() |
| |
|
| | @abc.abstractmethod |
| | def _get_web_access_uris(self): |
| | """Helper method to get the web access uris of the runnable job""" |
| | pass |
| |
|
| | @abc.abstractmethod |
| | def _log_web_access_uris(self): |
| | """Helper method to log the web access uris of the runnable job""" |
| | pass |
| |
|
| | def _block_until_complete(self): |
| | """Helper method to block and check on runnable job until complete. |
| | |
| | Raises: |
| | RuntimeError: If job failed or cancelled. |
| | """ |
| |
|
| | log_wait = _LOG_WAIT_TIME |
| |
|
| | previous_time = time.time() |
| | while self.state not in _JOB_COMPLETE_STATES: |
| | current_time = time.time() |
| | if current_time - previous_time >= _LOG_WAIT_TIME: |
| | self._log_job_state() |
| | log_wait = min(log_wait * _WAIT_TIME_MULTIPLIER, _MAX_WAIT_TIME) |
| | previous_time = current_time |
| | self._log_web_access_uris() |
| | time.sleep(_JOB_WAIT_TIME) |
| |
|
| | self._log_job_state() |
| |
|
| | if isinstance(self, CustomJob): |
| | |
| | experiment_runs = [] |
| | if self._gca_resource.job_spec.experiment_run: |
| | experiment_runs = [self._gca_resource.job_spec.experiment_run] |
| | elif self._gca_resource.job_spec.tensorboard: |
| | tensorboard_id = self._gca_resource.job_spec.tensorboard.split("/")[-1] |
| | try: |
| | tb_runs = aiplatform.TensorboardRun.list( |
| | tensorboard_experiment_name=self.name, |
| | tensorboard_id=tensorboard_id, |
| | ) |
| | experiment_runs = [ |
| | f"{self.name}-{tb_run.name.split('/')[-1]}" |
| | for tb_run in tb_runs |
| | ] |
| | except (ValueError, api_exceptions.GoogleAPIError) as e: |
| | _LOGGER.warning( |
| | f"Failed to list experiment runs for tensorboard " |
| | f"{tensorboard_id} due to: {e}" |
| | ) |
| | for experiment_run in experiment_runs: |
| | try: |
| | |
| | experiment_run_context = aiplatform.Context(experiment_run) |
| | experiment_run_context.update( |
| | metadata={ |
| | metadata_constants._STATE_KEY: ( |
| | gca_execution_compat.Execution.State.COMPLETE.name |
| | ) |
| | } |
| | ) |
| | except (ValueError, api_exceptions.GoogleAPIError) as e: |
| | _LOGGER.warning( |
| | f"Failed to end experiment run {experiment_run} due to: {e}" |
| | ) |
| |
|
| | |
| | |
| | if self._gca_resource.state in _JOB_ERROR_STATES: |
| | raise RuntimeError("Job failed with:\n%s" % self._gca_resource.error) |
| | else: |
| | _LOGGER.log_action_completed_against_resource("run", "completed", self) |
| |
|
| | @abc.abstractmethod |
| | def run(self) -> None: |
| | pass |
| |
|
| | @classmethod |
| | def get( |
| | cls, |
| | resource_name: str, |
| | project: Optional[str] = None, |
| | location: Optional[str] = None, |
| | credentials: Optional[auth_credentials.Credentials] = None, |
| | ) -> "_RunnableJob": |
| | """Get a Vertex AI Job for the given resource_name. |
| | |
| | Args: |
| | resource_name (str): |
| | Required. A fully-qualified resource name or ID. |
| | project (str): |
| | Optional. project to retrieve dataset from. If not set, project |
| | set in aiplatform.init will be used. |
| | location (str): |
| | Optional. location to retrieve dataset from. If not set, location |
| | set in aiplatform.init will be used. |
| | credentials (auth_credentials.Credentials): |
| | Custom credentials to use to upload this model. Overrides |
| | credentials set in aiplatform.init. |
| | |
| | Returns: |
| | A Vertex AI Job. |
| | """ |
| | self = cls._empty_constructor( |
| | project=project, |
| | location=location, |
| | credentials=credentials, |
| | resource_name=resource_name, |
| | ) |
| |
|
| | self._gca_resource = self._get_gca_resource(resource_name=resource_name) |
| |
|
| | return self |
| |
|
| | def wait_for_resource_creation(self) -> None: |
| | """Waits until resource has been created.""" |
| | self._wait_for_resource_creation() |
| |
|
| |
|
| | class DataLabelingJob(_Job): |
| | _resource_noun = "dataLabelingJobs" |
| | _getter_method = "get_data_labeling_job" |
| | _list_method = "list_data_labeling_jobs" |
| | _cancel_method = "cancel_data_labeling_job" |
| | _delete_method = "delete_data_labeling_job" |
| | _job_type = "labeling-tasks" |
| | _parse_resource_name_method = "parse_data_labeling_job_path" |
| | _format_resource_name_method = "data_labeling_job_path" |
| | pass |
| |
|
| |
|
| | class CustomJob(_RunnableJob, base.PreviewMixin): |
| | """Vertex AI Custom Job.""" |
| |
|
| | _resource_noun = "customJobs" |
| | _getter_method = "get_custom_job" |
| | _list_method = "list_custom_jobs" |
| | _cancel_method = "cancel_custom_job" |
| | _delete_method = "delete_custom_job" |
| | _parse_resource_name_method = "parse_custom_job_path" |
| | _format_resource_name_method = "custom_job_path" |
| | _job_type = "training" |
| | _preview_class = "google.cloud.aiplatform.aiplatform.preview.jobs.CustomJob" |
| |
|
| | def __init__( |
| | self, |
| | |
| | display_name: str, |
| | worker_pool_specs: Union[List[Dict], List[custom_job_v1.WorkerPoolSpec]], |
| | base_output_dir: Optional[str] = None, |
| | project: Optional[str] = None, |
| | location: Optional[str] = None, |
| | credentials: Optional[auth_credentials.Credentials] = None, |
| | labels: Optional[Dict[str, str]] = None, |
| | encryption_spec_key_name: Optional[str] = None, |
| | staging_bucket: Optional[str] = None, |
| | persistent_resource_id: Optional[str] = None, |
| | ): |
| | """Constructs a Custom Job with Worker Pool Specs. |
| | |
| | ``` |
| | Example usage: |
| | worker_pool_specs = [ |
| | { |
| | "machine_spec": { |
| | "machine_type": "n1-standard-4", |
| | "accelerator_type": "NVIDIA_TESLA_K80", |
| | "accelerator_count": 1, |
| | }, |
| | "replica_count": 1, |
| | "container_spec": { |
| | "image_uri": container_image_uri, |
| | "command": [], |
| | "args": [], |
| | }, |
| | } |
| | ] |
| | |
| | my_job = aiplatform.CustomJob( |
| | display_name='my_job', |
| | worker_pool_specs=worker_pool_specs, |
| | labels={'my_key': 'my_value'}, |
| | ) |
| | |
| | my_job.run() |
| | ``` |
| | |
| | |
| | For more information on configuring worker pool specs please visit: |
| | https://cloud.google.com/ai-platform-unified/docs/training/create-custom-job |
| | |
| | |
| | Args: |
| | display_name (str): |
| | Required. The user-defined name of the HyperparameterTuningJob. |
| | The name can be up to 128 characters long and can be consist |
| | of any UTF-8 characters. |
| | worker_pool_specs (Union[List[Dict], List[aiplatform.gapic.WorkerPoolSpec]]): |
| | Required. The spec of the worker pools including machine type and Docker image. |
| | Can provided as a list of dictionaries or list of WorkerPoolSpec proto messages. |
| | base_output_dir (str): |
| | Optional. GCS output directory of job. If not provided a |
| | timestamped directory in the staging directory will be used. |
| | project (str): |
| | Optional.Project to run the custom job in. Overrides project set in aiplatform.init. |
| | location (str): |
| | Optional.Location to run the custom job in. Overrides location set in aiplatform.init. |
| | credentials (auth_credentials.Credentials): |
| | Optional.Custom credentials to use to run call custom job service. Overrides |
| | credentials set in aiplatform.init. |
| | labels (Dict[str, str]): |
| | Optional. The labels with user-defined metadata to |
| | organize CustomJobs. |
| | Label keys and values can be no longer than 64 |
| | characters (Unicode codepoints), can only |
| | contain lowercase letters, numeric characters, |
| | underscores and dashes. International characters |
| | are allowed. |
| | See https://goo.gl/xmQnxf for more information |
| | and examples of labels. |
| | encryption_spec_key_name (str): |
| | Optional.Customer-managed encryption key name for a |
| | CustomJob. If this is set, then all resources |
| | created by the CustomJob will be encrypted with |
| | the provided encryption key. |
| | staging_bucket (str): |
| | Optional. Bucket for produced custom job artifacts. Overrides |
| | staging_bucket set in aiplatform.init. |
| | persistent_resource_id (str): |
| | Optional. The ID of the PersistentResource in the same Project |
| | and Location. If this is specified, the job will be run on |
| | existing machines held by the PersistentResource instead of |
| | on-demand short-live machines. The network and CMEK configs on |
| | the job should be consistent with those on the PersistentResource, |
| | otherwise, the job will be rejected. |
| | |
| | Raises: |
| | RuntimeError: If staging bucket was not set using aiplatform.init |
| | and a staging bucket was not passed in. |
| | """ |
| |
|
| | super().__init__(project=project, location=location, credentials=credentials) |
| |
|
| | staging_bucket = staging_bucket or initializer.global_config.staging_bucket |
| |
|
| | if not staging_bucket: |
| | raise RuntimeError( |
| | "staging_bucket should be passed to CustomJob constructor or " |
| | "should be set using aiplatform.init(staging_bucket='gs://my-bucket')" |
| | ) |
| |
|
| | if labels: |
| | utils.validate_labels(labels) |
| |
|
| | |
| | base_output_dir = base_output_dir or utils._timestamped_gcs_dir( |
| | staging_bucket, "aiplatform-custom-job" |
| | ) |
| |
|
| | if not display_name: |
| | display_name = self.__class__._generate_display_name() |
| |
|
| | self._gca_resource = gca_custom_job_compat.CustomJob( |
| | display_name=display_name, |
| | job_spec=gca_custom_job_compat.CustomJobSpec( |
| | worker_pool_specs=worker_pool_specs, |
| | base_output_directory=gca_io_compat.GcsDestination( |
| | output_uri_prefix=base_output_dir |
| | ), |
| | persistent_resource_id=persistent_resource_id, |
| | ), |
| | labels=labels, |
| | encryption_spec=initializer.global_config.get_encryption_spec( |
| | encryption_spec_key_name=encryption_spec_key_name |
| | ), |
| | ) |
| |
|
| | self._enable_autolog = False |
| |
|
| | @property |
| | def network(self) -> Optional[str]: |
| | """The full name of the Google Compute Engine |
| | [network](https://cloud.google.com/vpc/docs/vpc#networks) to which this |
| | CustomJob should be peered. |
| | |
| | Takes the format `projects/{project}/global/networks/{network}`. Where |
| | {project} is a project number, as in `12345`, and {network} is a network name. |
| | |
| | Private services access must already be configured for the network. If left |
| | unspecified, the CustomJob is not peered with any network. |
| | """ |
| | self._assert_gca_resource_is_available() |
| | return self._gca_resource.job_spec.network |
| |
|
| | def _get_web_access_uris(self) -> Dict[str, str]: |
| | """Helper method to get the web access uris of the custom job |
| | |
| | Returns: |
| | (Dict[str, str]): |
| | Web access uris of the custom job. |
| | """ |
| | return dict(self._gca_resource.web_access_uris) |
| |
|
| | def _log_web_access_uris(self): |
| | """Helper method to log the web access uris of the custom job""" |
| |
|
| | for worker, uri in self._get_web_access_uris().items(): |
| | if uri not in self._logged_web_access_uris: |
| | _LOGGER.info( |
| | "%s %s access the interactive shell terminals for the custom job:\n%s:\n%s" |
| | % ( |
| | self.__class__.__name__, |
| | self._gca_resource.name, |
| | worker, |
| | uri, |
| | ), |
| | ) |
| | self._logged_web_access_uris.add(uri) |
| |
|
| | @classmethod |
| | def from_local_script( |
| | cls, |
| | |
| | display_name: str, |
| | script_path: str, |
| | container_uri: str, |
| | enable_autolog: bool = False, |
| | args: Optional[Sequence[str]] = None, |
| | requirements: Optional[Sequence[str]] = None, |
| | environment_variables: Optional[Dict[str, str]] = None, |
| | replica_count: int = 1, |
| | machine_type: str = "n1-standard-4", |
| | accelerator_type: str = "ACCELERATOR_TYPE_UNSPECIFIED", |
| | accelerator_count: int = 0, |
| | boot_disk_type: str = "pd-ssd", |
| | boot_disk_size_gb: int = 100, |
| | reduction_server_replica_count: int = 0, |
| | reduction_server_machine_type: Optional[str] = None, |
| | reduction_server_container_uri: Optional[str] = None, |
| | base_output_dir: Optional[str] = None, |
| | project: Optional[str] = None, |
| | location: Optional[str] = None, |
| | credentials: Optional[auth_credentials.Credentials] = None, |
| | labels: Optional[Dict[str, str]] = None, |
| | encryption_spec_key_name: Optional[str] = None, |
| | staging_bucket: Optional[str] = None, |
| | persistent_resource_id: Optional[str] = None, |
| | tpu_topology: Optional[str] = None, |
| | ) -> "CustomJob": |
| | """Configures a custom job from a local script. |
| | |
| | Example usage: |
| | ``` |
| | job = aiplatform.CustomJob.from_local_script( |
| | display_name="my-custom-job", |
| | script_path="training_script.py", |
| | container_uri="gcr.io/cloud-aiplatform/training/tf-cpu.2-2:latest", |
| | requirements=["gcsfs==0.7.1"], |
| | replica_count=1, |
| | args=['--dataset', 'gs://my-bucket/my-dataset', |
| | '--model_output_uri', 'gs://my-bucket/model'] |
| | labels={'my_key': 'my_value'}, |
| | ) |
| | |
| | job.run() |
| | ``` |
| | |
| | Args: |
| | display_name (str): |
| | Required. The user-defined name of this CustomJob. |
| | script_path (str): |
| | Required. Local path to training script. |
| | container_uri (str): |
| | Required. Uri of the training container image to use for custom job. |
| | Support images in Artifact Registry, Container Registry, or Docker Hub. |
| | Vertex AI provides a wide range of executor images with pre-installed |
| | packages to meet users' various use cases. See the list of `pre-built containers |
| | for training <https://cloud.google.com/vertex-ai/docs/training/pre-built-containers>`. |
| | If not using image from this list, please make sure python3 and pip3 are installed in your container. |
| | enable_autolog (bool): |
| | Optional. If True, the Vertex Experiments autologging feature will be |
| | enabled in the CustomJob. Note that this will wrap your training script |
| | with some autologging-related code. |
| | args (Optional[Sequence[str]]): |
| | Optional. Command line arguments to be passed to the Python task. |
| | requirements (Sequence[str]): |
| | Optional. List of python packages dependencies of script. |
| | environment_variables (Dict[str, str]): |
| | Optional. Environment variables to be passed to the container. |
| | Should be a dictionary where keys are environment variable names |
| | and values are environment variable values for those names. |
| | At most 10 environment variables can be specified. |
| | The Name of the environment variable must be unique. |
| | |
| | environment_variables = { |
| | 'MY_KEY': 'MY_VALUE' |
| | } |
| | replica_count (int): |
| | Optional. The number of worker replicas. If replica count = 1 then one chief |
| | replica will be provisioned. If replica_count > 1 the remainder will be |
| | provisioned as a worker replica pool. |
| | machine_type (str): |
| | Optional. The type of machine to use for training. |
| | accelerator_type (str): |
| | Optional. Hardware accelerator type. One of ACCELERATOR_TYPE_UNSPECIFIED, |
| | NVIDIA_TESLA_K80, NVIDIA_TESLA_P100, NVIDIA_TESLA_V100, NVIDIA_TESLA_P4, |
| | NVIDIA_TESLA_T4 |
| | accelerator_count (int): |
| | Optional. The number of accelerators to attach to a worker replica. |
| | boot_disk_type (str): |
| | Optional. Type of the boot disk, default is `pd-ssd`. |
| | Valid values: `pd-ssd` (Persistent Disk Solid State Drive) or |
| | `pd-standard` (Persistent Disk Hard Disk Drive). |
| | boot_disk_size_gb (int): |
| | Optional. Size in GB of the boot disk, default is 100GB. |
| | boot disk size must be within the range of [100, 64000]. |
| | reduction_server_replica_count (int): |
| | The number of reduction server replicas, default is 0. |
| | reduction_server_machine_type (str): |
| | Optional. The type of machine to use for reduction server. |
| | reduction_server_container_uri (str): |
| | Optional. The Uri of the reduction server container image. |
| | See details: https://cloud.google.com/vertex-ai/docs/training/distributed-training#reduce_training_time_with_reduction_server |
| | base_output_dir (str): |
| | Optional. GCS output directory of job. If not provided a |
| | timestamped directory in the staging directory will be used. |
| | project (str): |
| | Optional. Project to run the custom job in. Overrides project set in aiplatform.init. |
| | location (str): |
| | Optional. Location to run the custom job in. Overrides location set in aiplatform.init. |
| | credentials (auth_credentials.Credentials): |
| | Optional. Custom credentials to use to run call custom job service. Overrides |
| | credentials set in aiplatform.init. |
| | labels (Dict[str, str]): |
| | Optional. The labels with user-defined metadata to |
| | organize CustomJobs. |
| | Label keys and values can be no longer than 64 |
| | characters (Unicode codepoints), can only |
| | contain lowercase letters, numeric characters, |
| | underscores and dashes. International characters |
| | are allowed. |
| | See https://goo.gl/xmQnxf for more information |
| | and examples of labels. |
| | encryption_spec_key_name (str): |
| | Optional. Customer-managed encryption key name for a |
| | CustomJob. If this is set, then all resources |
| | created by the CustomJob will be encrypted with |
| | the provided encryption key. |
| | staging_bucket (str): |
| | Optional. Bucket for produced custom job artifacts. Overrides |
| | staging_bucket set in aiplatform.init. |
| | persistent_resource_id (str): |
| | Optional. The ID of the PersistentResource in the same Project |
| | and Location. If this is specified, the job will be run on |
| | existing machines held by the PersistentResource instead of |
| | on-demand short-live machines. The network, CMEK, and node pool |
| | configs on the job should be consistent with those on the |
| | PersistentResource, otherwise, the job will be rejected. |
| | tpu_topology (str): |
| | Optional. Specifies the tpu topology to be used for |
| | TPU training job. This field is required for TPU v5 versions. For |
| | details on the TPU topology, refer to |
| | https://cloud.google.com/tpu/docs/v5e#tpu-v5e-config. The topology |
| | must be a supported value for the TPU machine type. |
| | |
| | Raises: |
| | RuntimeError: If staging bucket was not set using aiplatform.init |
| | and a staging bucket was not passed in. |
| | """ |
| |
|
| | project = project or initializer.global_config.project |
| | location = location or initializer.global_config.location |
| | staging_bucket = staging_bucket or initializer.global_config.staging_bucket |
| |
|
| | if not staging_bucket: |
| | raise RuntimeError( |
| | "staging_bucket should be passed to CustomJob.from_local_script or " |
| | "should be set using aiplatform.init(staging_bucket='gs://my-bucket')" |
| | ) |
| |
|
| | if labels: |
| | utils.validate_labels(labels) |
| |
|
| | worker_pool_specs = ( |
| | worker_spec_utils._DistributedTrainingSpec.chief_worker_pool( |
| | replica_count=replica_count, |
| | machine_type=machine_type, |
| | accelerator_count=accelerator_count, |
| | accelerator_type=accelerator_type, |
| | boot_disk_type=boot_disk_type, |
| | boot_disk_size_gb=boot_disk_size_gb, |
| | reduction_server_replica_count=reduction_server_replica_count, |
| | reduction_server_machine_type=reduction_server_machine_type, |
| | tpu_topology=tpu_topology, |
| | ).pool_specs |
| | ) |
| |
|
| | |
| | |
| | if enable_autolog: |
| | experiment_requirements = [constants.AIPLATFORM_AUTOLOG_DEPENDENCY_PATH] |
| | else: |
| | experiment_requirements = [] |
| |
|
| | if requirements: |
| | requirements.extend(experiment_requirements) |
| | else: |
| | requirements = experiment_requirements |
| |
|
| | if enable_autolog: |
| | with tempfile.TemporaryDirectory() as temp_dir: |
| | autolog_script_path = f"{temp_dir}/trainer_with_autolog.py" |
| | with open(autolog_script_path, "w") as f: |
| | autolog_script = ( |
| | "# Start a Vertex Experiments autolog session...\n" |
| | "from google.cloud " |
| | "import aiplatform\n" |
| | "aiplatform.autolog()\n\n" |
| | "# Training script...\n" |
| | ) |
| | f.write(autolog_script) |
| |
|
| | trainer_script = open(script_path, "r").read() |
| | f.write(trainer_script) |
| |
|
| | python_packager = source_utils._TrainingScriptPythonPackager( |
| | script_path=autolog_script_path, requirements=requirements |
| | ) |
| |
|
| | package_gcs_uri = python_packager.package_and_copy_to_gcs( |
| | gcs_staging_dir=staging_bucket, |
| | project=project, |
| | credentials=credentials, |
| | ) |
| | else: |
| | python_packager = source_utils._TrainingScriptPythonPackager( |
| | script_path=script_path, requirements=requirements |
| | ) |
| |
|
| | package_gcs_uri = python_packager.package_and_copy_to_gcs( |
| | gcs_staging_dir=staging_bucket, |
| | project=project, |
| | credentials=credentials, |
| | ) |
| |
|
| | for spec_order, spec in enumerate(worker_pool_specs): |
| |
|
| | if not spec: |
| | continue |
| |
|
| | if ( |
| | spec_order == worker_spec_utils._SPEC_ORDERS["server_spec"] |
| | and reduction_server_replica_count > 0 |
| | ): |
| | spec["container_spec"] = { |
| | "image_uri": reduction_server_container_uri, |
| | } |
| | |
| | elif ("docker.pkg.dev/vertex-ai/" in container_uri) or ( |
| | "gcr.io/cloud-aiplatform/" in container_uri |
| | ): |
| | spec["python_package_spec"] = { |
| | "executor_image_uri": container_uri, |
| | "python_module": python_packager.module_name, |
| | "package_uris": [package_gcs_uri], |
| | } |
| |
|
| | if args: |
| | spec["python_package_spec"]["args"] = args |
| |
|
| | if environment_variables: |
| | spec["python_package_spec"]["env"] = [ |
| | {"name": key, "value": value} |
| | for key, value in environment_variables.items() |
| | ] |
| | else: |
| | command = [ |
| | "sh", |
| | "-c", |
| | "pip install --upgrade pip && " |
| | + f"pip3 install -q --user {package_gcs_uri} && ".replace( |
| | "gs://", "/gcs/" |
| | ) |
| | + f"python3 -m {python_packager.module_name}", |
| | ] |
| |
|
| | if args: |
| | command[-1] += " " + " ".join(args) |
| |
|
| | spec["container_spec"] = { |
| | "image_uri": container_uri, |
| | "command": command, |
| | } |
| |
|
| | if environment_variables: |
| | spec["container_spec"]["env"] = [ |
| | {"name": key, "value": value} |
| | for key, value in environment_variables.items() |
| | ] |
| |
|
| | job = cls( |
| | display_name=display_name, |
| | worker_pool_specs=worker_pool_specs, |
| | base_output_dir=base_output_dir, |
| | project=project, |
| | location=location, |
| | credentials=credentials, |
| | labels=labels, |
| | encryption_spec_key_name=encryption_spec_key_name, |
| | staging_bucket=staging_bucket, |
| | persistent_resource_id=persistent_resource_id, |
| | ) |
| |
|
| | if enable_autolog: |
| | job._enable_autolog = True |
| |
|
| | return job |
| |
|
| | def run( |
| | self, |
| | service_account: Optional[str] = None, |
| | network: Optional[str] = None, |
| | timeout: Optional[int] = None, |
| | restart_job_on_worker_restart: bool = False, |
| | enable_web_access: bool = False, |
| | experiment: Optional[Union["aiplatform.Experiment", str]] = None, |
| | experiment_run: Optional[Union["aiplatform.ExperimentRun", str]] = None, |
| | tensorboard: Optional[str] = None, |
| | sync: bool = True, |
| | create_request_timeout: Optional[float] = None, |
| | disable_retries: bool = False, |
| | persistent_resource_id: Optional[str] = None, |
| | scheduling_strategy: Optional[gca_custom_job_compat.Scheduling.Strategy] = None, |
| | max_wait_duration: Optional[int] = None, |
| | ) -> None: |
| | """Run this configured CustomJob. |
| | |
| | Args: |
| | service_account (str): |
| | Optional. Specifies the service account for workload run-as account. |
| | Users submitting jobs must have act-as permission on this run-as account. |
| | network (str): |
| | Optional. The full name of the Compute Engine network to which the job |
| | should be peered. For example, projects/12345/global/networks/myVPC. |
| | Private services access must already be configured for the network. |
| | If left unspecified, the network set in aiplatform.init will be used. |
| | Otherwise, the job is not peered with any network. |
| | timeout (int): |
| | The maximum job running time in seconds. The default is 7 days. |
| | restart_job_on_worker_restart (bool): |
| | Restarts the entire CustomJob if a worker |
| | gets restarted. This feature can be used by |
| | distributed training jobs that are not resilient |
| | to workers leaving and joining a job. |
| | enable_web_access (bool): |
| | Whether you want Vertex AI to enable interactive shell access |
| | to training containers. |
| | https://cloud.google.com/vertex-ai/docs/training/monitor-debug-interactive-shell |
| | experiment (Union[aiplatform.Experiment, str]): |
| | Optional. The instance or name of an Experiment resource to which |
| | this CustomJob will upload training parameters and metrics. |
| | |
| | `service_account` is required with provided `experiment`. |
| | For more information on configuring your service account please visit: |
| | https://cloud.google.com/vertex-ai/docs/experiments/tensorboard-training |
| | experiment_run (Union[aiplatform.ExperimentRun, str]): |
| | Optional. The instance or name of an ExperimentRun resource to which |
| | this CustomJob will upload training parameters and metrics. |
| | This arg can only be set when `experiment` is set. If 'experiment' |
| | is set but 'experiment_run` is not, an ExperimentRun resource |
| | will still be auto-generated. |
| | tensorboard (str): |
| | Optional. The name of a Vertex AI |
| | [Tensorboard][google.cloud.aiplatform.v1beta1.Tensorboard] |
| | resource to which this CustomJob will upload Tensorboard |
| | logs. Format: |
| | ``projects/{project}/locations/{location}/tensorboards/{tensorboard}`` |
| | |
| | The training script should write Tensorboard to following Vertex AI environment |
| | variable: |
| | |
| | AIP_TENSORBOARD_LOG_DIR |
| | |
| | `service_account` is required with provided `tensorboard`. |
| | For more information on configuring your service account please visit: |
| | https://cloud.google.com/vertex-ai/docs/experiments/tensorboard-training |
| | sync (bool): |
| | Whether to execute this method synchronously. If False, this method |
| | will unblock and it will be executed in a concurrent Future. |
| | create_request_timeout (float): |
| | Optional. The timeout for the create request in seconds. |
| | disable_retries (bool): |
| | Indicates if the job should retry for internal errors after the |
| | job starts running. If True, overrides |
| | `restart_job_on_worker_restart` to False. |
| | persistent_resource_id (str): |
| | Optional. The ID of the PersistentResource in the same Project |
| | and Location. If this is specified, the job will be run on |
| | existing machines held by the PersistentResource instead of |
| | on-demand short-live machines. The network, CMEK, and node pool |
| | configs on the job should be consistent with those on the |
| | PersistentResource, otherwise, the job will be rejected. |
| | scheduling_strategy (gca_custom_job_compat.Scheduling.Strategy): |
| | Optional. Indicates the job scheduling strategy. |
| | max_wait_duration (int): |
| | This is the maximum duration that a job will wait for the |
| | requested resources to be provisioned in seconds. If set to 0, |
| | the job will wait indefinitely. The default is 1 day. |
| | """ |
| | network = network or initializer.global_config.network |
| | service_account = service_account or initializer.global_config.service_account |
| |
|
| | self._run( |
| | service_account=service_account, |
| | network=network, |
| | timeout=timeout, |
| | restart_job_on_worker_restart=restart_job_on_worker_restart, |
| | enable_web_access=enable_web_access, |
| | experiment=experiment, |
| | experiment_run=experiment_run, |
| | tensorboard=tensorboard, |
| | sync=sync, |
| | create_request_timeout=create_request_timeout, |
| | disable_retries=disable_retries, |
| | persistent_resource_id=persistent_resource_id, |
| | scheduling_strategy=scheduling_strategy, |
| | max_wait_duration=max_wait_duration, |
| | ) |
| |
|
| | @base.optional_sync() |
| | def _run( |
| | self, |
| | service_account: Optional[str] = None, |
| | network: Optional[str] = None, |
| | timeout: Optional[int] = None, |
| | restart_job_on_worker_restart: bool = False, |
| | enable_web_access: bool = False, |
| | experiment: Optional[Union["aiplatform.Experiment", str]] = None, |
| | experiment_run: Optional[Union["aiplatform.ExperimentRun", str]] = None, |
| | tensorboard: Optional[str] = None, |
| | sync: bool = True, |
| | create_request_timeout: Optional[float] = None, |
| | disable_retries: bool = False, |
| | persistent_resource_id: Optional[str] = None, |
| | scheduling_strategy: Optional[gca_custom_job_compat.Scheduling.Strategy] = None, |
| | max_wait_duration: Optional[int] = None, |
| | ) -> None: |
| | """Helper method to ensure network synchronization and to run the configured CustomJob. |
| | |
| | Args: |
| | service_account (str): |
| | Optional. Specifies the service account for workload run-as account. |
| | Users submitting jobs must have act-as permission on this run-as account. |
| | network (str): |
| | Optional. The full name of the Compute Engine network to which the job |
| | should be peered. For example, projects/12345/global/networks/myVPC. |
| | Private services access must already be configured for the network. |
| | timeout (int): |
| | The maximum job running time in seconds. The default is 7 days. |
| | restart_job_on_worker_restart (bool): |
| | Restarts the entire CustomJob if a worker |
| | gets restarted. This feature can be used by |
| | distributed training jobs that are not resilient |
| | to workers leaving and joining a job. |
| | enable_web_access (bool): |
| | Whether you want Vertex AI to enable interactive shell access |
| | to training containers. |
| | https://cloud.google.com/vertex-ai/docs/training/monitor-debug-interactive-shell |
| | experiment (Union[aiplatform.Experiment, str]): |
| | Optional. The instance or name of an Experiment resource to which |
| | this CustomJob will upload training parameters and metrics. |
| | |
| | `service_account` is required with provided `experiment`. |
| | For more information on configuring your service account please visit: |
| | https://cloud.google.com/vertex-ai/docs/experiments/tensorboard-training |
| | experiment_run (Union[aiplatform.ExperimentRun, str]): |
| | Optional. The instance or name of an ExperimentRun resource to which |
| | this CustomJob will upload training parameters and metrics. |
| | This arg can only be set when `experiment` is set. If 'experiment' |
| | is set but 'experiment_run` is not, an ExperimentRun resource |
| | will still be auto-generated. |
| | tensorboard (str): |
| | Optional. The name of a Vertex AI |
| | [Tensorboard][google.cloud.aiplatform.v1beta1.Tensorboard] |
| | resource to which this CustomJob will upload Tensorboard |
| | logs. Format: |
| | ``projects/{project}/locations/{location}/tensorboards/{tensorboard}`` |
| | |
| | The training script should write Tensorboard to following Vertex AI environment |
| | variable: |
| | |
| | AIP_TENSORBOARD_LOG_DIR |
| | |
| | `service_account` is required with provided `tensorboard`. |
| | For more information on configuring your service account please visit: |
| | https://cloud.google.com/vertex-ai/docs/experiments/tensorboard-training |
| | sync (bool): |
| | Whether to execute this method synchronously. If False, this method |
| | will unblock and it will be executed in a concurrent Future. |
| | create_request_timeout (float): |
| | Optional. The timeout for the create request in seconds. |
| | disable_retries (bool): |
| | Indicates if the job should retry for internal errors after the |
| | job starts running. If True, overrides |
| | `restart_job_on_worker_restart` to False. |
| | persistent_resource_id (str): |
| | Optional. The ID of the PersistentResource in the same Project |
| | and Location. If this is specified, the job will be run on |
| | existing machines held by the PersistentResource instead of |
| | on-demand short-live machines. The network, CMEK, and node pool |
| | configs on the job should be consistent with those on the |
| | PersistentResource, otherwise, the job will be rejected. |
| | scheduling_strategy (gca_custom_job_compat.Scheduling.Strategy): |
| | Optional. Indicates the job scheduling strategy. |
| | max_wait_duration (int): |
| | This is the maximum duration that a job will wait for the |
| | requested resources to be provisioned in seconds. If set to 0, |
| | the job will wait indefinitely. The default is 1 day. |
| | """ |
| | self.submit( |
| | service_account=service_account, |
| | network=network, |
| | timeout=timeout, |
| | restart_job_on_worker_restart=restart_job_on_worker_restart, |
| | enable_web_access=enable_web_access, |
| | experiment=experiment, |
| | experiment_run=experiment_run, |
| | tensorboard=tensorboard, |
| | create_request_timeout=create_request_timeout, |
| | disable_retries=disable_retries, |
| | persistent_resource_id=persistent_resource_id, |
| | scheduling_strategy=scheduling_strategy, |
| | max_wait_duration=max_wait_duration, |
| | ) |
| |
|
| | self._block_until_complete() |
| |
|
| | def submit( |
| | self, |
| | *, |
| | service_account: Optional[str] = None, |
| | network: Optional[str] = None, |
| | timeout: Optional[int] = None, |
| | restart_job_on_worker_restart: bool = False, |
| | enable_web_access: bool = False, |
| | experiment: Optional[Union["aiplatform.Experiment", str]] = None, |
| | experiment_run: Optional[Union["aiplatform.ExperimentRun", str]] = None, |
| | tensorboard: Optional[str] = None, |
| | create_request_timeout: Optional[float] = None, |
| | disable_retries: bool = False, |
| | persistent_resource_id: Optional[str] = None, |
| | scheduling_strategy: Optional[gca_custom_job_compat.Scheduling.Strategy] = None, |
| | max_wait_duration: Optional[int] = None, |
| | ) -> None: |
| | """Submit the configured CustomJob. |
| | |
| | Args: |
| | service_account (str): |
| | Optional. Specifies the service account for workload run-as account. |
| | Users submitting jobs must have act-as permission on this run-as account. |
| | network (str): |
| | Optional. The full name of the Compute Engine network to which the job |
| | should be peered. For example, projects/12345/global/networks/myVPC. |
| | Private services access must already be configured for the network. |
| | timeout (int): |
| | The maximum job running time in seconds. The default is 7 days. |
| | restart_job_on_worker_restart (bool): |
| | Restarts the entire CustomJob if a worker |
| | gets restarted. This feature can be used by |
| | distributed training jobs that are not resilient |
| | to workers leaving and joining a job. |
| | enable_web_access (bool): |
| | Whether you want Vertex AI to enable interactive shell access |
| | to training containers. |
| | https://cloud.google.com/vertex-ai/docs/training/monitor-debug-interactive-shell |
| | experiment (Union[aiplatform.Experiment, str]): |
| | Optional. The instance or name of an Experiment resource to which |
| | this CustomJob will upload training parameters and metrics. |
| | |
| | `service_account` is required with provided `experiment`. |
| | For more information on configuring your service account please visit: |
| | https://cloud.google.com/vertex-ai/docs/experiments/tensorboard-training |
| | experiment_run (Union[aiplatform.ExperimentRun, str]): |
| | Optional. The instance or name of an ExperimentRun resource to which |
| | this CustomJob will upload training parameters and metrics. |
| | This arg can only be set when `experiment` is set. If 'experiment' |
| | is set but 'experiment_run` is not, an ExperimentRun resource |
| | will still be auto-generated. |
| | tensorboard (str): |
| | Optional. The name of a Vertex AI |
| | [Tensorboard][google.cloud.aiplatform.v1beta1.Tensorboard] |
| | resource to which this CustomJob will upload Tensorboard |
| | logs. Format: |
| | ``projects/{project}/locations/{location}/tensorboards/{tensorboard}`` |
| | |
| | The training script should write Tensorboard to following Vertex AI environment |
| | variable: |
| | |
| | AIP_TENSORBOARD_LOG_DIR |
| | |
| | `service_account` is required with provided `tensorboard`. |
| | For more information on configuring your service account please visit: |
| | https://cloud.google.com/vertex-ai/docs/experiments/tensorboard-training |
| | create_request_timeout (float): |
| | Optional. The timeout for the create request in seconds. |
| | disable_retries (bool): |
| | Indicates if the job should retry for internal errors after the |
| | job starts running. If True, overrides |
| | `restart_job_on_worker_restart` to False. |
| | persistent_resource_id (str): |
| | Optional. The ID of the PersistentResource in the same Project |
| | and Location. If this is specified, the job will be run on |
| | existing machines held by the PersistentResource instead of |
| | on-demand short-live machines. The network, CMEK, and node pool |
| | configs on the job should be consistent with those on the |
| | PersistentResource, otherwise, the job will be rejected. |
| | scheduling_strategy (gca_custom_job_compat.Scheduling.Strategy): |
| | Optional. Indicates the job scheduling strategy. |
| | max_wait_duration (int): |
| | This is the maximum duration that a job will wait for the |
| | requested resources to be provisioned in seconds. If set to 0, |
| | the job will wait indefinitely. The default is 1 day. |
| | |
| | Raises: |
| | ValueError: |
| | If both `experiment` and `tensorboard` are specified or if |
| | `enable_autolog` is True in `CustomJob.from_local_script` but |
| | `experiment` is not specified or the specified experiment |
| | doesn't have a backing tensorboard. |
| | """ |
| | if experiment and tensorboard: |
| | raise ValueError("'experiment' and 'tensorboard' cannot be set together.") |
| | if self._enable_autolog and (not experiment): |
| | raise ValueError( |
| | "'experiment' is required since you've enabled autolog in 'from_local_script'." |
| | ) |
| |
|
| | service_account = service_account or initializer.global_config.service_account |
| | if service_account: |
| | self._gca_resource.job_spec.service_account = service_account |
| |
|
| | if network: |
| | self._gca_resource.job_spec.network = network |
| |
|
| | if ( |
| | timeout |
| | or restart_job_on_worker_restart |
| | or disable_retries |
| | or scheduling_strategy |
| | or max_wait_duration |
| | ): |
| | timeout = duration_pb2.Duration(seconds=timeout) if timeout else None |
| | max_wait_duration = ( |
| | duration_pb2.Duration(seconds=max_wait_duration) |
| | if max_wait_duration |
| | else None |
| | ) |
| | self._gca_resource.job_spec.scheduling = gca_custom_job_compat.Scheduling( |
| | timeout=timeout, |
| | restart_job_on_worker_restart=restart_job_on_worker_restart, |
| | disable_retries=disable_retries, |
| | strategy=scheduling_strategy, |
| | max_wait_duration=max_wait_duration, |
| | ) |
| |
|
| | if enable_web_access: |
| | self._gca_resource.job_spec.enable_web_access = enable_web_access |
| |
|
| | if tensorboard: |
| | self._gca_resource.job_spec.tensorboard = tensorboard |
| |
|
| | if persistent_resource_id: |
| | self._gca_resource.job_spec.persistent_resource_id = persistent_resource_id |
| |
|
| | ( |
| | self._gca_resource.job_spec.experiment, |
| | self._gca_resource.job_spec.experiment_run, |
| | ) = self._get_experiment_and_run_resource_name(experiment, experiment_run) |
| |
|
| | _LOGGER.log_create_with_lro(self.__class__) |
| |
|
| | self._gca_resource = self.api_client.create_custom_job( |
| | parent=self._parent, |
| | custom_job=self._gca_resource, |
| | timeout=create_request_timeout, |
| | ) |
| |
|
| | _LOGGER.log_create_complete_with_getter( |
| | self.__class__, self._gca_resource, "custom_job" |
| | ) |
| |
|
| | _LOGGER.info("View Custom Job:\n%s" % self._dashboard_uri()) |
| |
|
| | if tensorboard: |
| | _LOGGER.info( |
| | "View Tensorboard:\n%s" |
| | % console_utils.custom_job_tensorboard_console_uri( |
| | tensorboard, self.resource_name |
| | ) |
| | ) |
| |
|
| | @property |
| | def job_spec(self): |
| | return self._gca_resource.job_spec |
| |
|
| | @staticmethod |
| | def _get_experiment_and_run_resource_name( |
| | experiment: Optional[Union["aiplatform.Experiment", str]] = None, |
| | experiment_run: Optional[Union["aiplatform.ExperimentRun", str]] = None, |
| | ) -> Tuple[str, str]: |
| | """Helper method to get the experiment and run resource name for the custom job.""" |
| | if not experiment: |
| | return None, None |
| |
|
| | experiment_resource = ( |
| | aiplatform.Experiment(experiment) |
| | if isinstance(experiment, str) |
| | else experiment |
| | ) |
| |
|
| | if not experiment_run: |
| | return experiment_resource.resource_name, None |
| |
|
| | experiment_run_resource = ( |
| | aiplatform.ExperimentRun(experiment_run, experiment_resource) |
| | if isinstance(experiment_run, str) |
| | else experiment_run |
| | ) |
| | return ( |
| | experiment_resource.resource_name, |
| | experiment_run_resource.resource_name, |
| | ) |
| |
|
| |
|
| | class HyperparameterTuningJob(_RunnableJob, base.PreviewMixin): |
| | """Vertex AI Hyperparameter Tuning Job.""" |
| |
|
| | _resource_noun = "hyperparameterTuningJobs" |
| | _getter_method = "get_hyperparameter_tuning_job" |
| | _list_method = "list_hyperparameter_tuning_jobs" |
| | _cancel_method = "cancel_hyperparameter_tuning_job" |
| | _delete_method = "delete_hyperparameter_tuning_job" |
| | _parse_resource_name_method = "parse_hyperparameter_tuning_job_path" |
| | _format_resource_name_method = "hyperparameter_tuning_job_path" |
| | _job_type = "training" |
| | _preview_class = ( |
| | "google.cloud.aiplatform.aiplatform.preview.jobs.HyperparameterTuningJob" |
| | ) |
| |
|
| | def __init__( |
| | self, |
| | |
| | display_name: str, |
| | custom_job: CustomJob, |
| | metric_spec: Dict[str, str], |
| | parameter_spec: Dict[str, hyperparameter_tuning._ParameterSpec], |
| | max_trial_count: int, |
| | parallel_trial_count: int, |
| | max_failed_trial_count: int = 0, |
| | search_algorithm: Optional[str] = None, |
| | measurement_selection: Optional[str] = "best", |
| | project: Optional[str] = None, |
| | location: Optional[str] = None, |
| | credentials: Optional[auth_credentials.Credentials] = None, |
| | labels: Optional[Dict[str, str]] = None, |
| | encryption_spec_key_name: Optional[str] = None, |
| | ): |
| | """ |
| | Configures a HyperparameterTuning Job. |
| | |
| | Example usage: |
| | |
| | ``` |
| | from google.cloud.aiplatform import hyperparameter_tuning as hpt |
| | |
| | worker_pool_specs = [ |
| | { |
| | "machine_spec": { |
| | "machine_type": "n1-standard-4", |
| | "accelerator_type": "NVIDIA_TESLA_K80", |
| | "accelerator_count": 1, |
| | }, |
| | "replica_count": 1, |
| | "container_spec": { |
| | "image_uri": container_image_uri, |
| | "command": [], |
| | "args": [], |
| | }, |
| | } |
| | ] |
| | |
| | custom_job = aiplatform.CustomJob( |
| | display_name='my_job', |
| | worker_pool_specs=worker_pool_specs, |
| | labels={'my_key': 'my_value'}, |
| | ) |
| | |
| | |
| | hp_job = aiplatform.HyperparameterTuningJob( |
| | display_name='hp-test', |
| | custom_job=job, |
| | metric_spec={ |
| | 'loss': 'minimize', |
| | }, |
| | parameter_spec={ |
| | 'lr': hpt.DoubleParameterSpec(min=0.001, max=0.1, scale='log'), |
| | 'units': hpt.IntegerParameterSpec(min=4, max=128, scale='linear'), |
| | 'activation': hpt.CategoricalParameterSpec(values=['relu', 'selu']), |
| | 'batch_size': hpt.DiscreteParameterSpec(values=[128, 256], scale='linear') |
| | }, |
| | max_trial_count=128, |
| | parallel_trial_count=8, |
| | labels={'my_key': 'my_value'}, |
| | ) |
| | |
| | hp_job.run() |
| | |
| | print(hp_job.trials) |
| | ``` |
| | |
| | |
| | For more information on using hyperparameter tuning please visit: |
| | https://cloud.google.com/ai-platform-unified/docs/training/using-hyperparameter-tuning |
| | |
| | Args: |
| | display_name (str): |
| | Required. The user-defined name of the HyperparameterTuningJob. |
| | The name can be up to 128 characters long and can be consist |
| | of any UTF-8 characters. |
| | custom_job (aiplatform.CustomJob): |
| | Required. Configured CustomJob. The worker pool spec from this custom job |
| | applies to the CustomJobs created in all the trials. A persistent_resource_id can be |
| | specified on the custom job to be used when running this Hyperparameter Tuning job. |
| | metric_spec: Dict[str, str] |
| | Required. Dictionary representing metrics to optimize. The dictionary key is the metric_id, |
| | which is reported by your training job, and the dictionary value is the |
| | optimization goal of the metric('minimize' or 'maximize'). example: |
| | |
| | metric_spec = {'loss': 'minimize', 'accuracy': 'maximize'} |
| | |
| | parameter_spec (Dict[str, hyperparameter_tuning._ParameterSpec]): |
| | Required. Dictionary representing parameters to optimize. The dictionary key is the metric_id, |
| | which is passed into your training job as a command line key word argument, and the |
| | dictionary value is the parameter specification of the metric. |
| | |
| | |
| | from google.cloud.aiplatform import hyperparameter_tuning as hpt |
| | |
| | parameter_spec={ |
| | 'decay': hpt.DoubleParameterSpec(min=1e-7, max=1, scale='linear'), |
| | 'learning_rate': hpt.DoubleParameterSpec(min=1e-7, max=1, scale='linear') |
| | 'batch_size': hpt.DiscreteParamterSpec(values=[4, 8, 16, 32, 64, 128], scale='linear') |
| | } |
| | |
| | Supported parameter specifications can be found until aiplatform.hyperparameter_tuning. |
| | These parameter specification are currently supported: |
| | DoubleParameterSpec, IntegerParameterSpec, CategoricalParameterSpace, DiscreteParameterSpec |
| | |
| | max_trial_count (int): |
| | Required. The desired total number of Trials. |
| | parallel_trial_count (int): |
| | Required. The desired number of Trials to run in parallel. |
| | max_failed_trial_count (int): |
| | Optional. The number of failed Trials that need to be |
| | seen before failing the HyperparameterTuningJob. |
| | If set to 0, Vertex AI decides how many Trials |
| | must fail before the whole job fails. |
| | search_algorithm (str): |
| | The search algorithm specified for the Study. |
| | Accepts one of the following: |
| | `None` - If you do not specify an algorithm, your job uses |
| | the default Vertex AI algorithm. The default algorithm |
| | applies Bayesian optimization to arrive at the optimal |
| | solution with a more effective search over the parameter space. |
| | |
| | 'grid' - A simple grid search within the feasible space. This |
| | option is particularly useful if you want to specify a quantity |
| | of trials that is greater than the number of points in the |
| | feasible space. In such cases, if you do not specify a grid |
| | search, the Vertex AI default algorithm may generate duplicate |
| | suggestions. To use grid search, all parameter specs must be |
| | of type `IntegerParameterSpec`, `CategoricalParameterSpace`, |
| | or `DiscreteParameterSpec`. |
| | |
| | 'random' - A simple random search within the feasible space. |
| | measurement_selection (str): |
| | This indicates which measurement to use if/when the service |
| | automatically selects the final measurement from previously reported |
| | intermediate measurements. |
| | |
| | Accepts: 'best', 'last' |
| | |
| | Choose this based on two considerations: |
| | A) Do you expect your measurements to monotonically improve? If so, |
| | choose 'last'. On the other hand, if you're in a situation |
| | where your system can "over-train" and you expect the performance to |
| | get better for a while but then start declining, choose |
| | 'best'. B) Are your measurements significantly noisy |
| | and/or irreproducible? If so, 'best' will tend to be |
| | over-optimistic, and it may be better to choose 'last'. If |
| | both or neither of (A) and (B) apply, it doesn't matter which |
| | selection type is chosen. |
| | project (str): |
| | Optional. Project to run the HyperparameterTuningjob in. Overrides project set in aiplatform.init. |
| | location (str): |
| | Optional. Location to run the HyperparameterTuning in. Overrides location set in aiplatform.init. |
| | credentials (auth_credentials.Credentials): |
| | Optional. Custom credentials to use to run call HyperparameterTuning service. Overrides |
| | credentials set in aiplatform.init. |
| | labels (Dict[str, str]): |
| | Optional. The labels with user-defined metadata to |
| | organize HyperparameterTuningJobs. |
| | Label keys and values can be no longer than 64 |
| | characters (Unicode codepoints), can only |
| | contain lowercase letters, numeric characters, |
| | underscores and dashes. International characters |
| | are allowed. |
| | See https://goo.gl/xmQnxf for more information |
| | and examples of labels. |
| | encryption_spec_key_name (str): |
| | Optional. Customer-managed encryption key options for a |
| | HyperparameterTuningJob. If this is set, then |
| | all resources created by the |
| | HyperparameterTuningJob will be encrypted with |
| | the provided encryption key. |
| | """ |
| | super().__init__(project=project, location=location, credentials=credentials) |
| |
|
| | metrics = [ |
| | gca_study_compat.StudySpec.MetricSpec( |
| | metric_id=metric_id, goal=goal.upper() |
| | ) |
| | for metric_id, goal in metric_spec.items() |
| | ] |
| |
|
| | parameters = [ |
| | parameter._to_parameter_spec(parameter_id=parameter_id) |
| | for parameter_id, parameter in parameter_spec.items() |
| | ] |
| |
|
| | study_spec = gca_study_compat.StudySpec( |
| | metrics=metrics, |
| | parameters=parameters, |
| | algorithm=hyperparameter_tuning.SEARCH_ALGORITHM_TO_PROTO_VALUE[ |
| | search_algorithm |
| | ], |
| | measurement_selection_type=hyperparameter_tuning.MEASUREMENT_SELECTION_TO_PROTO_VALUE[ |
| | measurement_selection |
| | ], |
| | ) |
| |
|
| | if not display_name: |
| | display_name = self.__class__._generate_display_name() |
| |
|
| | self._gca_resource = ( |
| | gca_hyperparameter_tuning_job_compat.HyperparameterTuningJob( |
| | display_name=display_name, |
| | study_spec=study_spec, |
| | max_trial_count=max_trial_count, |
| | parallel_trial_count=parallel_trial_count, |
| | max_failed_trial_count=max_failed_trial_count, |
| | trial_job_spec=copy.deepcopy(custom_job.job_spec), |
| | labels=labels, |
| | encryption_spec=initializer.global_config.get_encryption_spec( |
| | encryption_spec_key_name=encryption_spec_key_name |
| | ), |
| | ) |
| | ) |
| |
|
| | @property |
| | def network(self) -> Optional[str]: |
| | """The full name of the Google Compute Engine |
| | [network](https://cloud.google.com/vpc/docs/vpc#networks) to which this |
| | HyperparameterTuningJob should be peered. |
| | |
| | Takes the format `projects/{project}/global/networks/{network}`. Where |
| | {project} is a project number, as in `12345`, and {network} is a network name. |
| | |
| | Private services access must already be configured for the network. If left |
| | unspecified, the HyperparameterTuningJob is not peered with any network. |
| | """ |
| | self._assert_gca_resource_is_available() |
| | return getattr(self._gca_resource.trial_job_spec, "network") |
| |
|
| | def _get_web_access_uris(self) -> Dict[str, Dict[str, str]]: |
| | """Helper method to get the web access uris of the hyperparameter job |
| | |
| | Returns: |
| | (Dict[str, Dict[str, str]]): |
| | Web access uris of the hyperparameter job. |
| | """ |
| | web_access_uris = dict() |
| | for trial in self.trials: |
| | web_access_uris[trial.id] = web_access_uris.get(trial.id, dict()) |
| | for worker, uri in trial.web_access_uris.items(): |
| | web_access_uris[trial.id][worker] = uri |
| | return web_access_uris |
| |
|
| | def _log_web_access_uris(self): |
| | """Helper method to log the web access uris of the hyperparameter job""" |
| |
|
| | for trial_id, trial_web_access_uris in self._get_web_access_uris().items(): |
| | for worker, uri in trial_web_access_uris.items(): |
| | if uri not in self._logged_web_access_uris: |
| | _LOGGER.info( |
| | "%s %s access the interactive shell terminals for trial - %s:\n%s:\n%s" |
| | % ( |
| | self.__class__.__name__, |
| | self._gca_resource.name, |
| | trial_id, |
| | worker, |
| | uri, |
| | ), |
| | ) |
| | self._logged_web_access_uris.add(uri) |
| |
|
| | def run( |
| | self, |
| | service_account: Optional[str] = None, |
| | network: Optional[str] = None, |
| | timeout: Optional[int] = None, |
| | restart_job_on_worker_restart: bool = False, |
| | enable_web_access: bool = False, |
| | tensorboard: Optional[str] = None, |
| | sync: bool = True, |
| | create_request_timeout: Optional[float] = None, |
| | disable_retries: bool = False, |
| | scheduling_strategy: Optional[gca_custom_job_compat.Scheduling.Strategy] = None, |
| | max_wait_duration: Optional[int] = None, |
| | ) -> None: |
| | """Run this configured CustomJob. |
| | |
| | Args: |
| | service_account (str): |
| | Optional. Specifies the service account for workload run-as account. |
| | Users submitting jobs must have act-as permission on this run-as account. |
| | network (str): |
| | Optional. The full name of the Compute Engine network to which the job |
| | should be peered. For example, projects/12345/global/networks/myVPC. |
| | Private services access must already be configured for the network. |
| | If left unspecified, the network set in aiplatform.init will be used. |
| | Otherwise, the job is not peered with any network. |
| | timeout (int): |
| | Optional. The maximum job running time in seconds. The default is 7 days. |
| | restart_job_on_worker_restart (bool): |
| | Restarts the entire CustomJob if a worker |
| | gets restarted. This feature can be used by |
| | distributed training jobs that are not resilient |
| | to workers leaving and joining a job. |
| | enable_web_access (bool): |
| | Whether you want Vertex AI to enable interactive shell access |
| | to training containers. |
| | https://cloud.google.com/vertex-ai/docs/training/monitor-debug-interactive-shell |
| | tensorboard (str): |
| | Optional. The name of a Vertex AI |
| | [Tensorboard][google.cloud.aiplatform.v1beta1.Tensorboard] |
| | resource to which this CustomJob will upload Tensorboard |
| | logs. Format: |
| | ``projects/{project}/locations/{location}/tensorboards/{tensorboard}`` |
| | |
| | The training script should write Tensorboard to following Vertex AI environment |
| | variable: |
| | |
| | AIP_TENSORBOARD_LOG_DIR |
| | |
| | `service_account` is required with provided `tensorboard`. |
| | For more information on configuring your service account please visit: |
| | https://cloud.google.com/vertex-ai/docs/experiments/tensorboard-training |
| | sync (bool): |
| | Whether to execute this method synchronously. If False, this method |
| | will unblock and it will be executed in a concurrent Future. |
| | create_request_timeout (float): |
| | Optional. The timeout for the create request in seconds. |
| | disable_retries (bool): |
| | Indicates if the job should retry for internal errors after the |
| | job starts running. If True, overrides |
| | `restart_job_on_worker_restart` to False. |
| | scheduling_strategy (gca_custom_job_compat.Scheduling.Strategy): |
| | Optional. Indicates the job scheduling strategy. |
| | max_wait_duration (int): |
| | This is the maximum duration that a job will wait for the |
| | requested resources to be provisioned in seconds. If set to 0, |
| | the job will wait indefinitely. The default is 1 day. |
| | """ |
| | network = network or initializer.global_config.network |
| | service_account = service_account or initializer.global_config.service_account |
| |
|
| | self._run( |
| | service_account=service_account, |
| | network=network, |
| | timeout=timeout, |
| | restart_job_on_worker_restart=restart_job_on_worker_restart, |
| | enable_web_access=enable_web_access, |
| | tensorboard=tensorboard, |
| | sync=sync, |
| | create_request_timeout=create_request_timeout, |
| | disable_retries=disable_retries, |
| | scheduling_strategy=scheduling_strategy, |
| | max_wait_duration=max_wait_duration, |
| | ) |
| |
|
| | @base.optional_sync() |
| | def _run( |
| | self, |
| | service_account: Optional[str] = None, |
| | network: Optional[str] = None, |
| | timeout: Optional[int] = None, |
| | restart_job_on_worker_restart: bool = False, |
| | enable_web_access: bool = False, |
| | tensorboard: Optional[str] = None, |
| | sync: bool = True, |
| | create_request_timeout: Optional[float] = None, |
| | disable_retries: bool = False, |
| | scheduling_strategy: Optional[gca_custom_job_compat.Scheduling.Strategy] = None, |
| | max_wait_duration: Optional[int] = None, |
| | ) -> None: |
| | """Helper method to ensure network synchronization and to run the configured CustomJob. |
| | |
| | Args: |
| | service_account (str): |
| | Optional. Specifies the service account for workload run-as account. |
| | Users submitting jobs must have act-as permission on this run-as account. |
| | network (str): |
| | Optional. The full name of the Compute Engine network to which the job |
| | should be peered. For example, projects/12345/global/networks/myVPC. |
| | Private services access must already be configured for the network. |
| | timeout (int): |
| | Optional. The maximum job running time in seconds. The default is 7 days. |
| | restart_job_on_worker_restart (bool): |
| | Restarts the entire CustomJob if a worker |
| | gets restarted. This feature can be used by |
| | distributed training jobs that are not resilient |
| | to workers leaving and joining a job. |
| | enable_web_access (bool): |
| | Whether you want Vertex AI to enable interactive shell access |
| | to training containers. |
| | https://cloud.google.com/vertex-ai/docs/training/monitor-debug-interactive-shell |
| | tensorboard (str): |
| | Optional. The name of a Vertex AI |
| | [Tensorboard][google.cloud.aiplatform.v1beta1.Tensorboard] |
| | resource to which this CustomJob will upload Tensorboard |
| | logs. Format: |
| | ``projects/{project}/locations/{location}/tensorboards/{tensorboard}`` |
| | |
| | The training script should write Tensorboard to following Vertex AI environment |
| | variable: |
| | |
| | AIP_TENSORBOARD_LOG_DIR |
| | |
| | `service_account` is required with provided `tensorboard`. |
| | For more information on configuring your service account please visit: |
| | https://cloud.google.com/vertex-ai/docs/experiments/tensorboard-training |
| | sync (bool): |
| | Whether to execute this method synchronously. If False, this method |
| | will unblock and it will be executed in a concurrent Future. |
| | create_request_timeout (float): |
| | Optional. The timeout for the create request in seconds. |
| | disable_retries (bool): |
| | Indicates if the job should retry for internal errors after the |
| | job starts running. If True, overrides |
| | `restart_job_on_worker_restart` to False. |
| | scheduling_strategy (gca_custom_job_compat.Scheduling.Strategy): |
| | Optional. Indicates the job scheduling strategy. |
| | max_wait_duration (int): |
| | This is the maximum duration that a job will wait for the |
| | requested resources to be provisioned in seconds. If set to 0, |
| | the job will wait indefinitely. The default is 1 day. |
| | """ |
| | if service_account: |
| | self._gca_resource.trial_job_spec.service_account = service_account |
| |
|
| | if network: |
| | self._gca_resource.trial_job_spec.network = network |
| |
|
| | if ( |
| | timeout |
| | or restart_job_on_worker_restart |
| | or disable_retries |
| | or max_wait_duration |
| | or scheduling_strategy |
| | ): |
| | timeout = duration_pb2.Duration(seconds=timeout) if timeout else None |
| | max_wait_duration = ( |
| | duration_pb2.Duration(seconds=max_wait_duration) |
| | if max_wait_duration |
| | else None |
| | ) |
| | self._gca_resource.trial_job_spec.scheduling = ( |
| | gca_custom_job_compat.Scheduling( |
| | timeout=timeout, |
| | restart_job_on_worker_restart=restart_job_on_worker_restart, |
| | disable_retries=disable_retries, |
| | strategy=scheduling_strategy, |
| | max_wait_duration=max_wait_duration, |
| | ) |
| | ) |
| |
|
| | if enable_web_access: |
| | self._gca_resource.trial_job_spec.enable_web_access = enable_web_access |
| |
|
| | if tensorboard: |
| | self._gca_resource.trial_job_spec.tensorboard = tensorboard |
| |
|
| | _LOGGER.log_create_with_lro(self.__class__) |
| |
|
| | self._gca_resource = self.api_client.create_hyperparameter_tuning_job( |
| | parent=self._parent, |
| | hyperparameter_tuning_job=self._gca_resource, |
| | timeout=create_request_timeout, |
| | ) |
| |
|
| | _LOGGER.log_create_complete_with_getter( |
| | self.__class__, self._gca_resource, "hpt_job" |
| | ) |
| |
|
| | _LOGGER.info("View HyperparameterTuningJob:\n%s" % self._dashboard_uri()) |
| |
|
| | if tensorboard: |
| | _LOGGER.info( |
| | "View Tensorboard:\n%s" |
| | % console_utils.custom_job_tensorboard_console_uri( |
| | tensorboard, self.resource_name |
| | ) |
| | ) |
| |
|
| | self._block_until_complete() |
| |
|
| | @property |
| | def trials(self) -> List[gca_study_compat.Trial]: |
| | self._assert_gca_resource_is_available() |
| | return list(self._gca_resource.trials) |
| |
|
| |
|
| | class ModelDeploymentMonitoringJob(_Job): |
| | """Vertex AI Model Deployment Monitoring Job. |
| | |
| | This class should be used in conjunction with the Endpoint class |
| | in order to configure model monitoring for deployed models. |
| | """ |
| |
|
| | _resource_noun = "modelDeploymentMonitoringJobs" |
| | _getter_method = "get_model_deployment_monitoring_job" |
| | _list_method = "list_model_deployment_monitoring_jobs" |
| | _cancel_method = "cancel_model_deployment_monitoring_jobs" |
| | _delete_method = "delete_model_deployment_monitoring_job" |
| | _job_type = "model-deployment-monitoring" |
| | _parse_resource_name_method = "parse_model_deployment_monitoring_job_path" |
| | _format_resource_name_method = "model_deployment_monitoring_job_path" |
| |
|
| | def __init__( |
| | self, |
| | model_deployment_monitoring_job_name: str, |
| | project: Optional[str] = None, |
| | location: Optional[str] = None, |
| | credentials: Optional[auth_credentials.Credentials] = None, |
| | ): |
| | """Initializer for ModelDeploymentMonitoringJob. |
| | |
| | Args: |
| | model_deployment_monitoring_job_name (str): |
| | Required. A fully-qualified ModelDeploymentMonitoringJob resource name or ID. |
| | Example: "projects/.../locations/.../modelDeploymentMonitoringJobs/456" or |
| | "456" when project and location are initialized or passed. |
| | project: (str), |
| | Optional. project to retrieve ModelDeploymentMonitoringJob from. If not set, |
| | project set in aiplatform.init will be used. |
| | location: (str), |
| | Optional. location to retrieve ModelDeploymentMonitoringJob from. If not set, |
| | location set in aiplatform.init will be used. |
| | credentials: (auth_credentials.Credentials), |
| | Optional. Custom credentials to use. If not set, credentials set in |
| | aiplatform.init will be used. |
| | """ |
| | super().__init__( |
| | job_name=model_deployment_monitoring_job_name, |
| | project=project, |
| | location=location, |
| | credentials=credentials, |
| | ) |
| | self._gca_resource = self._get_gca_resource( |
| | resource_name=model_deployment_monitoring_job_name |
| | ) |
| |
|
| | @classmethod |
| | def _parse_configs( |
| | cls, |
| | objective_configs: Union[ |
| | model_monitoring.ObjectiveConfig, |
| | Dict[str, model_monitoring.ObjectiveConfig], |
| | ], |
| | endpoint: "aiplatform.Endpoint", |
| | deployed_model_ids: Optional[List[str]] = None, |
| | ) -> List[ |
| | gca_model_deployment_monitoring_job_compat.ModelDeploymentMonitoringObjectiveConfig |
| | ]: |
| | """Helper function for matching objective configs with their corresponding models. |
| | |
| | Args: |
| | objective_configs (Union[model_monitoring.objective.ObjectiveConfig, |
| | Dict[str, model_monitoring.objective.ObjectiveConfig]): |
| | Required. A single config if it applies to all models, or a dictionary of |
| | model_id: model_monitoring.objective.ObjectiveConfig if |
| | different model IDs have different configs. |
| | endpoint (aiplatform.Endpoint): |
| | Required. A valid instance of aiplatforn.Endpoint to launch the MDM job on. |
| | deployed_model_ids (Optional[List[str]]): |
| | Optional. A list of deployed model IDs to apply the objective config to. |
| | Note that a model will have a deployed_model_id that is different from the |
| | uploaded model ID, and IDs in this list should consist of deployed model IDs |
| | on the same endpoint passed in the argument. If `objective_configs` is a dictionary, |
| | then this parameter is ignored. If `objective_configs` is an instance of |
| | `model_monitoring.ObjectiveConfig` and `deployed_model_ids` is a non-empty |
| | list of valid IDs, then the same objective config will apply to all models in this list. |
| | |
| | Returns: |
| | A List of ModelDeploymentMonitoringObjectiveConfig objects. |
| | |
| | Raises: |
| | ValueError: |
| | When the model IDs given are invalid. |
| | RuntimeError: |
| | When XAI is enabled on a model that doesn't have XAI parameters |
| | configured. |
| | """ |
| | all_models = [] |
| | xai_enabled = [] |
| | for model in endpoint.list_models(): |
| | all_models.append(model.id) |
| | if str(model.explanation_spec.parameters) != "": |
| | xai_enabled.append(model.id) |
| |
|
| | all_configs = [] |
| |
|
| | |
| | if deployed_model_ids is not None: |
| | if not all(model in all_models for model in deployed_model_ids): |
| | error_string = ( |
| | "Invalid model ID. The model ID must be one of [" |
| | + ",".join(all_models) |
| | + "]. Note that deployed model IDs are different from the uploaded model's ID" |
| | ) |
| | raise ValueError(error_string) |
| | else: |
| | all_models = deployed_model_ids |
| |
|
| | if isinstance(objective_configs, model_monitoring.ObjectiveConfig): |
| | for model in all_models: |
| | if ( |
| | model not in xai_enabled |
| | and objective_configs.explanation_config is not None |
| | ): |
| | raise RuntimeError( |
| | "Invalid config for model ID %s. `explanation_config` should only be enabled if the model has `explanation_spec populated" |
| | % model |
| | ) |
| | all_configs.append( |
| | gca_model_deployment_monitoring_job_compat.ModelDeploymentMonitoringObjectiveConfig( |
| | deployed_model_id=model, |
| | objective_config=objective_configs.as_proto(), |
| | ) |
| | ) |
| |
|
| | |
| | else: |
| | if not all(model in all_models for model in objective_configs.keys()): |
| | error_string = ( |
| | "Invalid model ID. The model ID must be one of [" |
| | + ",".join(all_models) |
| | + "]. Note that deployed model IDs are different from the uploaded model's ID" |
| | ) |
| | raise ValueError(error_string) |
| | for (deployed_model, objective_config) in objective_configs.items(): |
| | if ( |
| | deployed_model not in xai_enabled |
| | and objective_config.explanation_config is not None |
| | ): |
| | raise RuntimeError( |
| | "Invalid config for model ID %s. `explanation_config` should only be enabled if the model has `explanation_spec populated" |
| | % deployed_model |
| | ) |
| | all_configs.append( |
| | gca_model_deployment_monitoring_job_compat.ModelDeploymentMonitoringObjectiveConfig( |
| | deployed_model_id=deployed_model, |
| | objective_config=objective_config.as_proto(), |
| | ) |
| | ) |
| |
|
| | return all_configs |
| |
|
| | @classmethod |
| | def create( |
| | cls, |
| | endpoint: Union[str, "aiplatform.Endpoint"], |
| | objective_configs: Optional[ |
| | Union[ |
| | model_monitoring.ObjectiveConfig, |
| | Dict[str, model_monitoring.ObjectiveConfig], |
| | ] |
| | ] = None, |
| | logging_sampling_strategy: Optional[model_monitoring.RandomSampleConfig] = None, |
| | schedule_config: Optional[model_monitoring.ScheduleConfig] = None, |
| | display_name: Optional[str] = None, |
| | deployed_model_ids: Optional[List[str]] = None, |
| | alert_config: Optional[model_monitoring.EmailAlertConfig] = None, |
| | predict_instance_schema_uri: Optional[str] = None, |
| | sample_predict_instance: Optional[str] = None, |
| | analysis_instance_schema_uri: Optional[str] = None, |
| | bigquery_tables_log_ttl: Optional[int] = None, |
| | stats_anomalies_base_directory: Optional[str] = None, |
| | enable_monitoring_pipeline_logs: Optional[bool] = None, |
| | labels: Optional[Dict[str, str]] = None, |
| | encryption_spec_key_name: Optional[str] = None, |
| | project: Optional[str] = None, |
| | location: Optional[str] = None, |
| | credentials: Optional[auth_credentials.Credentials] = None, |
| | create_request_timeout: Optional[float] = None, |
| | ) -> "ModelDeploymentMonitoringJob": |
| | """Creates and launches a model monitoring job. |
| | |
| | Args: |
| | endpoint (Union[str, "aiplatform.Endpoint"]): |
| | Required. Endpoint resource name or an instance of `aiplatform.Endpoint`. Format: |
| | ``projects/{project}/locations/{location}/endpoints/{endpoint}`` |
| | |
| | objective_configs (Union[model_monitoring.ObjectiveConfig, |
| | Dict[str, model_monitoring.ObjectiveConfig]]): |
| | Required. A single config if it applies to all models, or a dictionary of |
| | model_id: model_monitoring.objective.ObjectiveConfig if |
| | different model IDs have different configs. |
| | |
| | logging_sampling_strategy (model_monitoring.sampling.RandomSampleConfig): |
| | Optional. Sample Strategy for logging. |
| | |
| | schedule_config (model_monitoring.schedule.ScheduleConfig): |
| | Optional. Configures model monitoring job scheduling interval in hours. |
| | This defines how often the monitoring jobs are triggered. |
| | |
| | display_name (str): |
| | Optional. The user-defined name of the |
| | ModelDeploymentMonitoringJob. The name can be up |
| | to 128 characters long and can be consist of any |
| | UTF-8 characters. |
| | Display name of a ModelDeploymentMonitoringJob. |
| | |
| | deployed_model_ids (List[str]): |
| | Optional. Use this argument to specify which deployed models to |
| | apply the objective config to. If left unspecified, the same config |
| | will be applied to all deployed models. |
| | |
| | alert_config (model_monitoring.alert.EmailAlertConfig): |
| | Optional. Configures how alerts are sent to the user. Right now |
| | only email alert is supported. |
| | |
| | predict_instance_schema_uri (str): |
| | Optional. YAML schema file uri describing the format of |
| | a single instance, which are given to format |
| | the Endpoint's prediction (and explanation). If |
| | not set, the schema will be generated from |
| | collected predict requests. |
| | |
| | sample_predict_instance (str): |
| | Optional. Sample Predict instance, same format as PredictionRequest.instances, |
| | this can be set as a replacement of predict_instance_schema_uri |
| | If not set, the schema will be generated from collected predict requests. |
| | |
| | analysis_instance_schema_uri (str): |
| | Optional. YAML schema file uri describing the format of a single |
| | instance that you want Tensorflow Data Validation (TFDV) to |
| | analyze. If this field is empty, all the feature data types are |
| | inferred from predict_instance_schema_uri, meaning that TFDV |
| | will use the data in the exact format as prediction request/response. |
| | If there are any data type differences between predict instance |
| | and TFDV instance, this field can be used to override the schema. |
| | For models trained with Vertex AI, this field must be set as all the |
| | fields in predict instance formatted as string. |
| | |
| | bigquery_tables_log_ttl (int): |
| | Optional. The TTL(time to live) of BigQuery tables in user projects |
| | which stores logs. A day is the basic unit of |
| | the TTL and we take the ceil of TTL/86400(a |
| | day). e.g. { second: 3600} indicates ttl = 1 |
| | day. |
| | |
| | stats_anomalies_base_directory (str): |
| | Optional. Stats anomalies base folder path. |
| | |
| | enable_monitoring_pipeline_logs (bool): |
| | Optional. If true, the scheduled monitoring pipeline logs are sent to |
| | Google Cloud Logging, including pipeline status and |
| | anomalies detected. Please note the logs incur cost, which |
| | are subject to `Cloud Logging |
| | pricing <https://cloud.google.com/logging#pricing>`__. |
| | |
| | labels (Dict[str, str]): |
| | Optional. The labels with user-defined metadata to |
| | organize the ModelDeploymentMonitoringJob. |
| | Label keys and values can be no longer than 64 |
| | characters (Unicode codepoints), can only |
| | contain lowercase letters, numeric characters, |
| | underscores and dashes. International characters |
| | are allowed. See https://goo.gl/xmQnxf for more information |
| | and examples of labels. |
| | |
| | encryption_spec_key_name (str): |
| | Optional. Customer-managed encryption key spec for a |
| | ModelDeploymentMonitoringJob. If set, this |
| | ModelDeploymentMonitoringJob and all |
| | sub-resources of this |
| | ModelDeploymentMonitoringJob will be secured by |
| | this key. |
| | |
| | create_request_timeout (int): |
| | Optional. Timeout in seconds for the model monitoring job creation request. |
| | |
| | Returns: |
| | An instance of ModelDeploymentMonitoringJob. |
| | """ |
| | if not display_name: |
| | display_name = cls._generate_display_name() |
| |
|
| | utils.validate_display_name(display_name) |
| |
|
| | if labels: |
| | utils.validate_labels(labels) |
| |
|
| | if stats_anomalies_base_directory: |
| | stats_anomalies_base_directory = gca_io_compat.GcsDestination( |
| | output_uri_prefix=stats_anomalies_base_directory |
| | ) |
| |
|
| | if encryption_spec_key_name: |
| | encryption_spec_key_name = gca_encryption_spec_compat.EncryptionSpec( |
| | kms_key_name=encryption_spec_key_name |
| | ) |
| |
|
| | if credentials is None and isinstance(endpoint, aiplatform.Endpoint): |
| | credentials = endpoint.credentials |
| | self = cls._empty_constructor( |
| | project=project, location=location, credentials=credentials |
| | ) |
| |
|
| | parent = initializer.global_config.common_location_path( |
| | project=self.project, |
| | location=self.location, |
| | ) |
| |
|
| | if isinstance(endpoint, str): |
| | endpoint = aiplatform.Endpoint(endpoint, project, location, credentials) |
| |
|
| | mdm_objective_config_seq = cls._parse_configs( |
| | objective_configs, |
| | endpoint, |
| | deployed_model_ids, |
| | ) |
| |
|
| | gapic_mdm_job = ( |
| | gca_model_deployment_monitoring_job_compat.ModelDeploymentMonitoringJob( |
| | display_name=display_name, |
| | endpoint=endpoint.resource_name, |
| | model_deployment_monitoring_objective_configs=mdm_objective_config_seq, |
| | logging_sampling_strategy=logging_sampling_strategy.as_proto(), |
| | model_deployment_monitoring_schedule_config=schedule_config.as_proto(), |
| | model_monitoring_alert_config=alert_config.as_proto(), |
| | predict_instance_schema_uri=predict_instance_schema_uri, |
| | analysis_instance_schema_uri=analysis_instance_schema_uri, |
| | sample_predict_instance=sample_predict_instance, |
| | stats_anomalies_base_directory=stats_anomalies_base_directory, |
| | enable_monitoring_pipeline_logs=enable_monitoring_pipeline_logs, |
| | labels=labels, |
| | encryption_spec=encryption_spec_key_name, |
| | ) |
| | ) |
| |
|
| | _LOGGER.log_create_with_lro(cls) |
| | self._gca_resource = self.api_client.create_model_deployment_monitoring_job( |
| | parent=parent, |
| | model_deployment_monitoring_job=gapic_mdm_job, |
| | timeout=create_request_timeout, |
| | ) |
| |
|
| | _LOGGER.log_create_complete(cls, self._gca_resource, "mdm_job") |
| |
|
| | _LOGGER.info( |
| | "View Model Deployment Monitoring Job:\n%s" % self._dashboard_uri() |
| | ) |
| |
|
| | return self |
| |
|
| | @classmethod |
| | def cancel(cls): |
| | raise NotImplementedError( |
| | "Cancel method is not implemented because it is not applicable. A running model deployment monitoring job can be paused or deleted." |
| | ) |
| |
|
| | @property |
| | def end_time(self): |
| | _LOGGER.info( |
| | "Model deployment monitoring jobs do not have an end time since their inactive states are either PAUSED or PENDING." |
| | ) |
| | return None |
| |
|
| | def update( |
| | self, |
| | *, |
| | display_name: Optional[str] = None, |
| | schedule_config: Optional[model_monitoring.ScheduleConfig] = None, |
| | alert_config: Optional[model_monitoring.EmailAlertConfig] = None, |
| | logging_sampling_strategy: Optional[model_monitoring.RandomSampleConfig] = None, |
| | labels: Optional[Dict[str, str]] = None, |
| | bigquery_tables_log_ttl: Optional[int] = None, |
| | enable_monitoring_pipeline_logs: Optional[bool] = None, |
| | objective_configs: Optional[ |
| | Union[ |
| | model_monitoring.ObjectiveConfig, |
| | Dict[str, model_monitoring.ObjectiveConfig], |
| | ] |
| | ] = None, |
| | deployed_model_ids: Optional[List[str]] = None, |
| | update_request_timeout: Optional[float] = None, |
| | ) -> "ModelDeploymentMonitoringJob": |
| | """Updates an existing ModelDeploymentMonitoringJob. |
| | |
| | Args: |
| | |
| | display_name (str): |
| | Optional. The user-defined name of the |
| | ModelDeploymentMonitoringJob. The name can be up |
| | to 128 characters long and can be consist of any |
| | UTF-8 characters. |
| | Display name of a ModelDeploymentMonitoringJob. |
| | |
| | schedule_config (model_monitoring.schedule.ScheduleConfig): |
| | Required. Configures model monitoring job scheduling interval in hours. |
| | This defines how often the monitoring jobs are triggered. |
| | alert_config (model_monitoring.alert.EmailAlertConfig): |
| | Optional. Configures how alerts are sent to the user. Right now |
| | only email alert is supported. |
| | logging_sampling_strategy (model_monitoring.sampling.RandomSampleConfig): |
| | Required. Sample Strategy for logging. |
| | |
| | labels (Dict[str, str]): |
| | Optional. The labels with user-defined metadata to |
| | organize the ModelDeploymentMonitoringJob. |
| | Label keys and values can be no longer than 64 |
| | characters (Unicode codepoints), can only |
| | contain lowercase letters, numeric characters, |
| | underscores and dashes. International characters |
| | are allowed. See https://goo.gl/xmQnxf for more information |
| | and examples of labels. |
| | bigquery_tables_log_ttl (int): |
| | Optional. The number of days for which the logs are stored. |
| | The TTL(time to live) of BigQuery tables in user projects |
| | which stores logs. A day is the basic unit of |
| | the TTL and we take the ceil of TTL/86400(a |
| | day). e.g. { second: 3600} indicates ttl = 1 |
| | day. |
| | |
| | enable_monitoring_pipeline_logs (bool): |
| | Optional. If true, the scheduled monitoring pipeline logs are sent to |
| | Google Cloud Logging, including pipeline status and |
| | anomalies detected. Please note the logs incur cost, which |
| | are subject to `Cloud Logging |
| | pricing <https://cloud.google.com/logging#pricing>`__. |
| | |
| | objective_configs (Union[ |
| | Required. model_monitoring.objective.ObjectiveConfig, |
| | Dict[str, model_monitoring.objective.ObjectiveConfig]): |
| | A single config if it applies to all models, or a dictionary of |
| | model_id: model_monitoring.objective.ObjectiveConfig if |
| | different model IDs have different configs. |
| | |
| | deployed_model_ids (List[str]): |
| | Optional. Use this argument to specify which deployed models to |
| | apply the updated objective config to. If left unspecified, the same config |
| | will be applied to all deployed models. |
| | upate_request_timeout (float): |
| | Optional. Timeout in seconds for the model monitoring job update request. |
| | """ |
| | self._sync_gca_resource() |
| | current_job = copy.deepcopy(self._gca_resource) |
| | update_mask: List[str] = [] |
| | if display_name is not None: |
| | update_mask.append("display_name") |
| | current_job.display_name = display_name |
| | if schedule_config is not None: |
| | update_mask.append("model_deployment_monitoring_schedule_config") |
| | current_job.model_deployment_monitoring_schedule_config = ( |
| | schedule_config.as_proto() |
| | ) |
| | if alert_config is not None: |
| | update_mask.append("model_monitoring_alert_config") |
| | current_job.model_monitoring_alert_config = alert_config.as_proto() |
| | if logging_sampling_strategy is not None: |
| | update_mask.append("logging_sampling_strategy") |
| | current_job.logging_sampling_strategy = logging_sampling_strategy.as_proto() |
| | if labels is not None: |
| | update_mask.append("labels") |
| | current_job.labels = labels |
| | if bigquery_tables_log_ttl is not None: |
| | update_mask.append("log_ttl") |
| | current_job.log_ttl = duration_pb2.Duration( |
| | seconds=bigquery_tables_log_ttl * 86400 |
| | ) |
| | if enable_monitoring_pipeline_logs is not None: |
| | update_mask.append("enable_monitoring_pipeline_logs") |
| | current_job.enable_monitoring_pipeline_logs = ( |
| | enable_monitoring_pipeline_logs |
| | ) |
| | if objective_configs is not None: |
| | update_mask.append("model_deployment_monitoring_objective_configs") |
| | current_job.model_deployment_monitoring_objective_configs = ( |
| | ModelDeploymentMonitoringJob._parse_configs( |
| | objective_configs=objective_configs, |
| | endpoint=aiplatform.Endpoint( |
| | current_job.endpoint, credentials=self.credentials |
| | ), |
| | deployed_model_ids=deployed_model_ids, |
| | ) |
| | ) |
| | |
| | lro = self.api_client.update_model_deployment_monitoring_job( |
| | model_deployment_monitoring_job=current_job, |
| | update_mask=field_mask_pb2.FieldMask(paths=update_mask), |
| | timeout=update_request_timeout, |
| | ) |
| | self._gca_resource = lro.result(timeout=None) |
| | return self |
| |
|
| | def pause(self) -> "ModelDeploymentMonitoringJob": |
| | """Pause a running MDM job.""" |
| | self.api_client.pause_model_deployment_monitoring_job( |
| | name=self._gca_resource.name |
| | ) |
| | return self |
| |
|
| | def resume(self) -> "ModelDeploymentMonitoringJob": |
| | """Resumes a paused MDM job.""" |
| | self.api_client.resume_model_deployment_monitoring_job( |
| | name=self._gca_resource.name |
| | ) |
| | return self |
| |
|
| | def delete(self) -> None: |
| | """Deletes an MDM job.""" |
| | self.api_client.delete_model_deployment_monitoring_job( |
| | name=self._gca_resource.name |
| | ) |
| |
|