| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """A class for AutoML Job's Candidate.""" |
| | from __future__ import absolute_import |
| |
|
| | from six import string_types |
| |
|
| | from sagemaker import Session |
| | from sagemaker.job import _Job |
| | from sagemaker.utils import name_from_base |
| |
|
| |
|
| | class CandidateEstimator(object): |
| | """A class for SageMaker AutoML Job Candidate""" |
| |
|
| | def __init__(self, candidate, sagemaker_session=None): |
| | """Constructor of CandidateEstimator. |
| | |
| | Args: |
| | candidate (dict): a dictionary of candidate returned by AutoML.list_candidates() |
| | or AutoML.best_candidate(). |
| | sagemaker_session (sagemaker.session.Session): A SageMaker Session |
| | object, used for SageMaker interactions (default: None). If not |
| | specified, one is created using the default AWS configuration |
| | chain. |
| | """ |
| | self.name = candidate["CandidateName"] |
| | self.containers = candidate["InferenceContainers"] |
| | self.steps = self._process_steps(candidate["CandidateSteps"]) |
| | self.sagemaker_session = sagemaker_session or Session() |
| |
|
| | def get_steps(self): |
| | """Get the step job of a candidate so that users can construct estimators/transformers |
| | |
| | Returns: |
| | list: a list of dictionaries that provide information about each step job's name, |
| | type, inputs and description |
| | """ |
| | candidate_steps = [] |
| | for step in self.steps: |
| | step_type = step["type"] |
| | step_name = step["name"] |
| | if step_type == "TrainingJob": |
| | training_job = self.sagemaker_session.sagemaker_client.describe_training_job( |
| | TrainingJobName=step_name |
| | ) |
| |
|
| | inputs = training_job["InputDataConfig"] |
| | candidate_step = CandidateStep(step_name, inputs, step_type, training_job) |
| | candidate_steps.append(candidate_step) |
| | elif step_type == "TransformJob": |
| | transform_job = self.sagemaker_session.sagemaker_client.describe_transform_job( |
| | TransformJobName=step_name |
| | ) |
| | inputs = transform_job["TransformInput"] |
| | candidate_step = CandidateStep(step_name, inputs, step_type, transform_job) |
| | candidate_steps.append(candidate_step) |
| | return candidate_steps |
| |
|
| | def fit( |
| | self, |
| | inputs, |
| | candidate_name=None, |
| | volume_kms_key=None, |
| | encrypt_inter_container_traffic=False, |
| | vpc_config=None, |
| | wait=True, |
| | logs=True, |
| | ): |
| | """Rerun a candidate's step jobs with new input datasets or security config. |
| | |
| | Args: |
| | inputs (str or list[str]): Local path or S3 Uri where the training data is stored. If a |
| | local path is provided, the dataset will be uploaded to an S3 location. |
| | candidate_name (str): name of the candidate to be rerun, if None, candidate's original |
| | name will be used. |
| | volume_kms_key (str): The KMS key id to encrypt data on the storage volume attached to |
| | the ML compute instance(s). |
| | encrypt_inter_container_traffic (bool): To encrypt all communications between ML compute |
| | instances in distributed training. Default: False. |
| | vpc_config (dict): Specifies a VPC that jobs and hosted models have access to. |
| | Control access to and from training and model containers by configuring the VPC |
| | wait (bool): Whether the call should wait until all jobs completes (default: True). |
| | logs (bool): Whether to show the logs produced by the job. |
| | Only meaningful when wait is True (default: True). |
| | """ |
| | if logs and not wait: |
| | raise ValueError( |
| | """Logs can only be shown if wait is set to True. |
| | Please either set wait to True or set logs to False.""" |
| | ) |
| |
|
| | self.name = candidate_name or self.name |
| | running_jobs = {} |
| |
|
| | |
| | if isinstance(inputs, string_types): |
| | if not inputs.startswith("s3://"): |
| | inputs = self.sagemaker_session.upload_data(inputs, key_prefix="auto-ml-input-data") |
| |
|
| | for step in self.steps: |
| | step_type = step["type"] |
| | step_name = step["name"] |
| | if step_type == "TrainingJob": |
| | |
| | input_dict = {} |
| | if isinstance(inputs, string_types): |
| | input_dict["train"] = _Job._format_string_uri_input(inputs) |
| | else: |
| | msg = "Cannot format input {}. Expecting a string." |
| | raise ValueError(msg.format(inputs)) |
| |
|
| | channels = [ |
| | _Job._convert_input_to_channel(name, input) |
| | for name, input in input_dict.items() |
| | ] |
| |
|
| | desc = self.sagemaker_session.sagemaker_client.describe_training_job( |
| | TrainingJobName=step_name |
| | ) |
| | base_name = "sagemaker-automl-training-rerun" |
| | step_name = name_from_base(base_name) |
| | step["name"] = step_name |
| | train_args = self._get_train_args( |
| | desc, |
| | channels, |
| | step_name, |
| | volume_kms_key, |
| | encrypt_inter_container_traffic, |
| | vpc_config, |
| | ) |
| | self.sagemaker_session.train(**train_args) |
| | running_jobs[step_name] = True |
| |
|
| | elif step_type == "TransformJob": |
| | |
| | if not isinstance(inputs, string_types) or not inputs.startswith("s3://"): |
| | msg = "Cannot format input {}. Expecting a string starts with file:// or s3://" |
| | raise ValueError(msg.format(inputs)) |
| |
|
| | desc = self.sagemaker_session.sagemaker_client.describe_transform_job( |
| | TransformJobName=step_name |
| | ) |
| | base_name = "sagemaker-automl-transform-rerun" |
| | step_name = name_from_base(base_name) |
| | step["name"] = step_name |
| | transform_args = self._get_transform_args(desc, inputs, step_name, volume_kms_key) |
| | self.sagemaker_session.transform(**transform_args) |
| | running_jobs[step_name] = True |
| |
|
| | if wait: |
| | while True: |
| | for step in self.steps: |
| | status = None |
| | step_type = step["type"] |
| | step_name = step["name"] |
| | if step_type == "TrainingJob": |
| | status = self.sagemaker_session.sagemaker_client.describe_training_job( |
| | TrainingJobName=step_name |
| | )["TrainingJobStatus"] |
| | elif step_type == "TransformJob": |
| | status = self.sagemaker_session.sagemaker_client.describe_transform_job( |
| | TransformJobName=step_name |
| | )["TransformJobStatus"] |
| | if status in ("Completed", "Failed", "Stopped"): |
| | running_jobs[step_name] = False |
| | if self._check_all_job_finished(running_jobs): |
| | break |
| |
|
| | def _check_all_job_finished(self, running_jobs): |
| | """Check if all step jobs are finished. |
| | |
| | Args: |
| | running_jobs (dict): a dictionary that keeps track of the status |
| | of each step job. |
| | |
| | Returns (bool): True if all step jobs are finished. False if one or |
| | more step jobs are still running. |
| | """ |
| | for _, v in running_jobs.items(): |
| | if v: |
| | return False |
| | return True |
| |
|
| | def _get_train_args( |
| | self, desc, inputs, name, volume_kms_key, encrypt_inter_container_traffic, vpc_config |
| | ): |
| | """Format training args to pass in sagemaker_session.train. |
| | |
| | Args: |
| | desc (dict): the response from DescribeTrainingJob API. |
| | inputs (list): a list of input data channels. |
| | name (str): the name of the step job. |
| | volume_kms_key (str): The KMS key id to encrypt data on the storage volume attached to |
| | the ML compute instance(s). |
| | encrypt_inter_container_traffic (bool): To encrypt all communications between ML compute |
| | instances in distributed training. |
| | vpc_config (dict): Specifies a VPC that jobs and hosted models have access to. |
| | Control access to and from training and model containers by configuring the VPC |
| | |
| | Returns (dcit): a dictionary that can be used as args of |
| | sagemaker_session.train method. |
| | """ |
| | train_args = { |
| | "input_config": inputs, |
| | "job_name": name, |
| | "input_mode": desc["AlgorithmSpecification"]["TrainingInputMode"], |
| | "role": desc["RoleArn"], |
| | "output_config": desc["OutputDataConfig"], |
| | "resource_config": desc["ResourceConfig"], |
| | "image_uri": desc["AlgorithmSpecification"]["TrainingImage"], |
| | "enable_network_isolation": desc["EnableNetworkIsolation"], |
| | "encrypt_inter_container_traffic": encrypt_inter_container_traffic, |
| | "use_spot_instances": desc["EnableManagedSpotTraining"], |
| | "hyperparameters": {}, |
| | "stop_condition": {}, |
| | "metric_definitions": None, |
| | "checkpoint_s3_uri": None, |
| | "checkpoint_local_path": None, |
| | "tags": [], |
| | "vpc_config": None, |
| | } |
| |
|
| | if volume_kms_key is not None: |
| | train_args["resource_config"]["VolumeKmsKeyId"] = volume_kms_key |
| | if "VpcConfig" in desc: |
| | train_args["vpc_config"] = desc["VpcConfig"] |
| | elif vpc_config is not None: |
| | train_args["vpc_config"] = vpc_config |
| | if "Hyperparameters" in desc: |
| | train_args["hyperparameters"] = desc["Hyperparameters"] |
| | if "CheckpointConfig" in desc: |
| | train_args["checkpoint_s3_uri"] = desc["CheckpointConfig"]["S3Uri"] |
| | train_args["checkpoint_local_path"] = desc["CheckpointConfig"]["LocalPath"] |
| | if "StoppingCondition" in desc: |
| | train_args["stop_condition"] = desc["StoppingCondition"] |
| | return train_args |
| |
|
| | def _get_transform_args(self, desc, inputs, name, volume_kms_key): |
| | """Format training args to pass in sagemaker_session.train. |
| | |
| | Args: |
| | desc (dict): the response from DescribeTrainingJob API. |
| | inputs (str): an S3 uri where new input dataset is stored. |
| | name (str): the name of the step job. |
| | volume_kms_key (str): The KMS key id to encrypt data on the storage volume attached to |
| | the ML compute instance(s). |
| | |
| | Returns (dcit): a dictionary that can be used as args of |
| | sagemaker_session.transform method. |
| | """ |
| | transform_args = {} |
| | transform_args["job_name"] = name |
| | transform_args["model_name"] = desc["ModelName"] |
| | transform_args["output_config"] = desc["TransformOutput"] |
| | transform_args["resource_config"] = desc["TransformResources"] |
| | transform_args["data_processing"] = desc["DataProcessing"] |
| | transform_args["tags"] = [] |
| | transform_args["strategy"] = None |
| | transform_args["max_concurrent_transforms"] = None |
| | transform_args["max_payload"] = None |
| | transform_args["env"] = None |
| | transform_args["experiment_config"] = None |
| |
|
| | input_config = desc["TransformInput"] |
| | input_config["DataSource"]["S3DataSource"]["S3Uri"] = inputs |
| | transform_args["input_config"] = input_config |
| |
|
| | if volume_kms_key is not None: |
| | transform_args["resource_config"]["VolumeKmsKeyId"] = volume_kms_key |
| | if "BatchStrategy" in desc: |
| | transform_args["strategy"] = desc["BatchStrategy"] |
| | if "MaxConcurrentTransforms" in desc: |
| | transform_args["max_concurrent_transforms"] = desc["MaxConcurrentTransforms"] |
| | if "MaxPayloadInMB" in desc: |
| | transform_args["max_payload"] = desc["MaxPayloadInMB"] |
| | if "Environment" in desc: |
| | transform_args["env"] = desc["Environment"] |
| |
|
| | return transform_args |
| |
|
| | def _process_steps(self, steps): |
| | """Extract candidate's step jobs name and type. |
| | |
| | Args: |
| | steps (list): a list of a candidate's step jobs. |
| | |
| | Returns (list): a list of extracted information about step jobs' |
| | name and type. |
| | """ |
| | processed_steps = [] |
| | for step in steps: |
| | step_name = step["CandidateStepName"] |
| | step_type = step["CandidateStepType"].split("::")[2] |
| | processed_steps.append({"name": step_name, "type": step_type}) |
| | return processed_steps |
| |
|
| |
|
| | class CandidateStep(object): |
| | """A class that maintains an AutoML Candidate step's name, inputs, type, and description.""" |
| |
|
| | def __init__(self, name, inputs, step_type, description): |
| | self._name = name |
| | self._inputs = inputs |
| | self._type = step_type |
| | self._description = description |
| |
|
| | @property |
| | def name(self): |
| | """Name of the candidate step -> (str)""" |
| | return self._name |
| |
|
| | @property |
| | def inputs(self): |
| | """Inputs of the candidate step -> (dict)""" |
| | return self._inputs |
| |
|
| | @property |
| | def type(self): |
| | """Type of the candidate step, Training or Transform -> (str)""" |
| | return self._type |
| |
|
| | @property |
| | def description(self): |
| | """Description of candidate step job -> (dict)""" |
| | return self._description |
| |
|