| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | from __future__ import absolute_import |
| |
|
| | import os |
| |
|
| | import pytest |
| | import numpy as np |
| | from configparser import ParsingError |
| | from sagemaker.utils import retries |
| |
|
| | from six.moves.urllib.parse import urlparse |
| |
|
| | import tests.integ |
| | from sagemaker import ( |
| | KMeans, |
| | FactorizationMachines, |
| | IPInsights, |
| | KNN, |
| | LDA, |
| | LinearLearner, |
| | NTM, |
| | PCA, |
| | RandomCutForest, |
| | image_uris, |
| | ) |
| | from sagemaker.amazon.common import read_records |
| | from sagemaker.chainer import Chainer |
| | from sagemaker.estimator import Estimator |
| | from sagemaker.mxnet import MXNet |
| | from sagemaker.pytorch.estimator import PyTorch |
| | from sagemaker.sklearn import SKLearn |
| | from sagemaker.tensorflow import TensorFlow |
| | from sagemaker.utils import sagemaker_timestamp |
| | from sagemaker.workflow import airflow as sm_airflow |
| | from sagemaker.xgboost import XGBoost |
| | from tests.integ import datasets, DATA_DIR |
| | from tests.integ.record_set import prepare_record_set_from_local_files |
| | from tests.integ.timeout import timeout |
| |
|
| | for _ in retries( |
| | max_retry_count=10, |
| | exception_message_prefix="airflow import ", |
| | seconds_to_sleep=6, |
| | ): |
| | try: |
| | from airflow import utils |
| | from airflow import DAG |
| | from airflow.providers.amazon.aws.operators.sagemaker import SageMakerTrainingOperator |
| | from airflow.providers.amazon.aws.operators.sagemaker_transform import ( |
| | SageMakerTransformOperator, |
| | ) |
| |
|
| | break |
| | except ParsingError: |
| | pass |
| |
|
| | PYTORCH_MNIST_DIR = os.path.join(DATA_DIR, "pytorch_mnist") |
| | PYTORCH_MNIST_SCRIPT = os.path.join(PYTORCH_MNIST_DIR, "mnist.py") |
| | AIRFLOW_CONFIG_TIMEOUT_IN_SECONDS = 10 |
| |
|
| | RESOURCE_PATH = os.path.join(os.path.dirname(__file__), "..", "data") |
| | TF_MNIST_RESOURCE_PATH = os.path.join(RESOURCE_PATH, "tensorflow_mnist") |
| | SCRIPT = os.path.join(TF_MNIST_RESOURCE_PATH, "mnist.py") |
| | ROLE = "SageMakerRole" |
| | SINGLE_INSTANCE_COUNT = 1 |
| |
|
| |
|
| | def test_byo_airflow_config_uploads_data_source_to_s3_when_inputs_provided( |
| | sagemaker_session, cpu_instance_type |
| | ): |
| | with timeout(seconds=AIRFLOW_CONFIG_TIMEOUT_IN_SECONDS): |
| | training_data_path = os.path.join(DATA_DIR, "dummy_tensor") |
| |
|
| | data_source_location = "test-airflow-config-{}".format(sagemaker_timestamp()) |
| | inputs = sagemaker_session.upload_data( |
| | path=training_data_path, key_prefix=os.path.join(data_source_location, "train") |
| | ) |
| |
|
| | estimator = Estimator( |
| | image_uri=image_uris.retrieve( |
| | "factorization-machines", sagemaker_session.boto_region_name |
| | ), |
| | role=ROLE, |
| | instance_count=SINGLE_INSTANCE_COUNT, |
| | instance_type=cpu_instance_type, |
| | sagemaker_session=sagemaker_session, |
| | ) |
| |
|
| | training_config = _build_airflow_workflow( |
| | estimator=estimator, instance_type=cpu_instance_type, inputs=inputs |
| | ) |
| |
|
| | _assert_that_s3_url_contains_data( |
| | sagemaker_session, |
| | training_config["InputDataConfig"][0]["DataSource"]["S3DataSource"]["S3Uri"], |
| | ) |
| |
|
| |
|
| | def test_kmeans_airflow_config_uploads_data_source_to_s3(sagemaker_session, cpu_instance_type): |
| | with timeout(seconds=AIRFLOW_CONFIG_TIMEOUT_IN_SECONDS): |
| | kmeans = KMeans( |
| | role=ROLE, |
| | instance_count=SINGLE_INSTANCE_COUNT, |
| | instance_type=cpu_instance_type, |
| | k=10, |
| | sagemaker_session=sagemaker_session, |
| | ) |
| |
|
| | kmeans.init_method = "random" |
| | kmeans.max_iterations = 1 |
| | kmeans.tol = 1 |
| | kmeans.num_trials = 1 |
| | kmeans.local_init_method = "kmeans++" |
| | kmeans.half_life_time_size = 1 |
| | kmeans.epochs = 1 |
| | kmeans.center_factor = 1 |
| | kmeans.eval_metrics = ["ssd", "msd"] |
| |
|
| | records = kmeans.record_set(datasets.one_p_mnist()[0][:100]) |
| |
|
| | training_config = _build_airflow_workflow( |
| | estimator=kmeans, instance_type=cpu_instance_type, inputs=records |
| | ) |
| |
|
| | _assert_that_s3_url_contains_data( |
| | sagemaker_session, |
| | training_config["InputDataConfig"][0]["DataSource"]["S3DataSource"]["S3Uri"], |
| | ) |
| |
|
| |
|
| | def test_fm_airflow_config_uploads_data_source_to_s3(sagemaker_session, cpu_instance_type): |
| | with timeout(seconds=AIRFLOW_CONFIG_TIMEOUT_IN_SECONDS): |
| | fm = FactorizationMachines( |
| | role=ROLE, |
| | instance_count=SINGLE_INSTANCE_COUNT, |
| | instance_type=cpu_instance_type, |
| | num_factors=10, |
| | predictor_type="regressor", |
| | epochs=2, |
| | clip_gradient=1e2, |
| | eps=0.001, |
| | rescale_grad=1.0 / 100, |
| | sagemaker_session=sagemaker_session, |
| | ) |
| |
|
| | training_set = datasets.one_p_mnist() |
| | records = fm.record_set(training_set[0][:200], training_set[1][:200].astype("float32")) |
| |
|
| | training_config = _build_airflow_workflow( |
| | estimator=fm, instance_type=cpu_instance_type, inputs=records |
| | ) |
| |
|
| | _assert_that_s3_url_contains_data( |
| | sagemaker_session, |
| | training_config["InputDataConfig"][0]["DataSource"]["S3DataSource"]["S3Uri"], |
| | ) |
| |
|
| |
|
| | def test_ipinsights_airflow_config_uploads_data_source_to_s3(sagemaker_session, cpu_instance_type): |
| | with timeout(seconds=AIRFLOW_CONFIG_TIMEOUT_IN_SECONDS): |
| | data_path = os.path.join(DATA_DIR, "ipinsights") |
| | data_filename = "train.csv" |
| |
|
| | with open(os.path.join(data_path, data_filename), "rb") as f: |
| | num_records = len(f.readlines()) |
| |
|
| | ipinsights = IPInsights( |
| | role=ROLE, |
| | instance_count=SINGLE_INSTANCE_COUNT, |
| | instance_type=cpu_instance_type, |
| | num_entity_vectors=10, |
| | vector_dim=100, |
| | sagemaker_session=sagemaker_session, |
| | ) |
| |
|
| | records = prepare_record_set_from_local_files( |
| | data_path, ipinsights.data_location, num_records, None, sagemaker_session |
| | ) |
| |
|
| | training_config = _build_airflow_workflow( |
| | estimator=ipinsights, instance_type=cpu_instance_type, inputs=records |
| | ) |
| |
|
| | _assert_that_s3_url_contains_data( |
| | sagemaker_session, |
| | training_config["InputDataConfig"][0]["DataSource"]["S3DataSource"]["S3Uri"], |
| | ) |
| |
|
| |
|
| | def test_knn_airflow_config_uploads_data_source_to_s3(sagemaker_session, cpu_instance_type): |
| | with timeout(seconds=AIRFLOW_CONFIG_TIMEOUT_IN_SECONDS): |
| | knn = KNN( |
| | role=ROLE, |
| | instance_count=SINGLE_INSTANCE_COUNT, |
| | instance_type=cpu_instance_type, |
| | k=10, |
| | predictor_type="regressor", |
| | sample_size=500, |
| | sagemaker_session=sagemaker_session, |
| | ) |
| |
|
| | training_set = datasets.one_p_mnist() |
| | records = knn.record_set(training_set[0][:200], training_set[1][:200].astype("float32")) |
| |
|
| | training_config = _build_airflow_workflow( |
| | estimator=knn, instance_type=cpu_instance_type, inputs=records |
| | ) |
| |
|
| | _assert_that_s3_url_contains_data( |
| | sagemaker_session, |
| | training_config["InputDataConfig"][0]["DataSource"]["S3DataSource"]["S3Uri"], |
| | ) |
| |
|
| |
|
| | @pytest.mark.skipif( |
| | tests.integ.test_region() in tests.integ.NO_LDA_REGIONS, |
| | reason="LDA image is not supported in certain regions", |
| | ) |
| | def test_lda_airflow_config_uploads_data_source_to_s3(sagemaker_session, cpu_instance_type): |
| | with timeout(seconds=AIRFLOW_CONFIG_TIMEOUT_IN_SECONDS): |
| | data_path = os.path.join(DATA_DIR, "lda") |
| | data_filename = "nips-train_1.pbr" |
| |
|
| | with open(os.path.join(data_path, data_filename), "rb") as f: |
| | all_records = read_records(f) |
| |
|
| | |
| | feature_num = int(all_records[0].features["values"].float32_tensor.shape[0]) |
| |
|
| | lda = LDA( |
| | role=ROLE, |
| | instance_type=cpu_instance_type, |
| | num_topics=10, |
| | sagemaker_session=sagemaker_session, |
| | ) |
| |
|
| | records = prepare_record_set_from_local_files( |
| | data_path, lda.data_location, len(all_records), feature_num, sagemaker_session |
| | ) |
| |
|
| | training_config = _build_airflow_workflow( |
| | estimator=lda, instance_type=cpu_instance_type, inputs=records, mini_batch_size=100 |
| | ) |
| |
|
| | _assert_that_s3_url_contains_data( |
| | sagemaker_session, |
| | training_config["InputDataConfig"][0]["DataSource"]["S3DataSource"]["S3Uri"], |
| | ) |
| |
|
| |
|
| | def test_linearlearner_airflow_config_uploads_data_source_to_s3( |
| | sagemaker_session, cpu_instance_type |
| | ): |
| | with timeout(seconds=AIRFLOW_CONFIG_TIMEOUT_IN_SECONDS): |
| | training_set = datasets.one_p_mnist() |
| | training_set[1][:100] = 1 |
| | training_set[1][100:200] = 0 |
| | training_set = training_set[0], training_set[1].astype(np.dtype("float32")) |
| |
|
| | ll = LinearLearner( |
| | ROLE, |
| | 1, |
| | cpu_instance_type, |
| | predictor_type="binary_classifier", |
| | sagemaker_session=sagemaker_session, |
| | ) |
| | ll.binary_classifier_model_selection_criteria = "accuracy" |
| | ll.target_recall = 0.5 |
| | ll.target_precision = 0.5 |
| | ll.positive_example_weight_mult = 0.1 |
| | ll.epochs = 1 |
| | ll.use_bias = True |
| | ll.num_models = 1 |
| | ll.num_calibration_samples = 1 |
| | ll.init_method = "uniform" |
| | ll.init_scale = 0.5 |
| | ll.init_sigma = 0.2 |
| | ll.init_bias = 5 |
| | ll.optimizer = "adam" |
| | ll.loss = "logistic" |
| | ll.wd = 0.5 |
| | ll.l1 = 0.5 |
| | ll.momentum = 0.5 |
| | ll.learning_rate = 0.1 |
| | ll.beta_1 = 0.1 |
| | ll.beta_2 = 0.1 |
| | ll.use_lr_scheduler = True |
| | ll.lr_scheduler_step = 2 |
| | ll.lr_scheduler_factor = 0.5 |
| | ll.lr_scheduler_minimum_lr = 0.1 |
| | ll.normalize_data = False |
| | ll.normalize_label = False |
| | ll.unbias_data = True |
| | ll.unbias_label = False |
| | ll.num_point_for_scaler = 10000 |
| | ll.margin = 1.0 |
| | ll.quantile = 0.5 |
| | ll.loss_insensitivity = 0.1 |
| | ll.huber_delta = 0.1 |
| | ll.early_stopping_tolerance = 0.0001 |
| | ll.early_stopping_patience = 3 |
| |
|
| | records = ll.record_set(training_set[0][:200], training_set[1][:200]) |
| |
|
| | training_config = _build_airflow_workflow( |
| | estimator=ll, instance_type=cpu_instance_type, inputs=records |
| | ) |
| |
|
| | _assert_that_s3_url_contains_data( |
| | sagemaker_session, |
| | training_config["InputDataConfig"][0]["DataSource"]["S3DataSource"]["S3Uri"], |
| | ) |
| |
|
| |
|
| | def test_ntm_airflow_config_uploads_data_source_to_s3(sagemaker_session, cpu_instance_type): |
| | with timeout(seconds=AIRFLOW_CONFIG_TIMEOUT_IN_SECONDS): |
| | data_path = os.path.join(DATA_DIR, "ntm") |
| | data_filename = "nips-train_1.pbr" |
| |
|
| | with open(os.path.join(data_path, data_filename), "rb") as f: |
| | all_records = read_records(f) |
| |
|
| | |
| | feature_num = int(all_records[0].features["values"].float32_tensor.shape[0]) |
| |
|
| | ntm = NTM( |
| | role=ROLE, |
| | instance_count=SINGLE_INSTANCE_COUNT, |
| | instance_type=cpu_instance_type, |
| | num_topics=10, |
| | sagemaker_session=sagemaker_session, |
| | ) |
| |
|
| | records = prepare_record_set_from_local_files( |
| | data_path, ntm.data_location, len(all_records), feature_num, sagemaker_session |
| | ) |
| |
|
| | training_config = _build_airflow_workflow( |
| | estimator=ntm, instance_type=cpu_instance_type, inputs=records |
| | ) |
| |
|
| | _assert_that_s3_url_contains_data( |
| | sagemaker_session, |
| | training_config["InputDataConfig"][0]["DataSource"]["S3DataSource"]["S3Uri"], |
| | ) |
| |
|
| |
|
| | def test_pca_airflow_config_uploads_data_source_to_s3(sagemaker_session, cpu_instance_type): |
| | with timeout(seconds=AIRFLOW_CONFIG_TIMEOUT_IN_SECONDS): |
| | pca = PCA( |
| | role=ROLE, |
| | instance_count=SINGLE_INSTANCE_COUNT, |
| | instance_type=cpu_instance_type, |
| | num_components=48, |
| | sagemaker_session=sagemaker_session, |
| | ) |
| |
|
| | pca.algorithm_mode = "randomized" |
| | pca.subtract_mean = True |
| | pca.extra_components = 5 |
| |
|
| | records = pca.record_set(datasets.one_p_mnist()[0][:100]) |
| |
|
| | training_config = _build_airflow_workflow( |
| | estimator=pca, instance_type=cpu_instance_type, inputs=records |
| | ) |
| |
|
| | _assert_that_s3_url_contains_data( |
| | sagemaker_session, |
| | training_config["InputDataConfig"][0]["DataSource"]["S3DataSource"]["S3Uri"], |
| | ) |
| |
|
| |
|
| | def test_rcf_airflow_config_uploads_data_source_to_s3(sagemaker_session, cpu_instance_type): |
| | with timeout(seconds=AIRFLOW_CONFIG_TIMEOUT_IN_SECONDS): |
| | |
| | feature_num = 14 |
| | train_input = np.random.rand(1000, feature_num) |
| |
|
| | rcf = RandomCutForest( |
| | role=ROLE, |
| | instance_count=SINGLE_INSTANCE_COUNT, |
| | instance_type=cpu_instance_type, |
| | num_trees=50, |
| | num_samples_per_tree=20, |
| | eval_metrics=["accuracy", "precision_recall_fscore"], |
| | sagemaker_session=sagemaker_session, |
| | ) |
| |
|
| | records = rcf.record_set(train_input) |
| |
|
| | training_config = _build_airflow_workflow( |
| | estimator=rcf, instance_type=cpu_instance_type, inputs=records |
| | ) |
| |
|
| | _assert_that_s3_url_contains_data( |
| | sagemaker_session, |
| | training_config["InputDataConfig"][0]["DataSource"]["S3DataSource"]["S3Uri"], |
| | ) |
| |
|
| |
|
| | def test_chainer_airflow_config_uploads_data_source_to_s3( |
| | sagemaker_local_session, cpu_instance_type, chainer_latest_version, chainer_latest_py_version |
| | ): |
| | with timeout(seconds=AIRFLOW_CONFIG_TIMEOUT_IN_SECONDS): |
| | script_path = os.path.join(DATA_DIR, "chainer_mnist", "mnist.py") |
| | data_path = os.path.join(DATA_DIR, "chainer_mnist") |
| |
|
| | chainer = Chainer( |
| | entry_point=script_path, |
| | role=ROLE, |
| | instance_count=SINGLE_INSTANCE_COUNT, |
| | instance_type="local", |
| | framework_version=chainer_latest_version, |
| | py_version=chainer_latest_py_version, |
| | sagemaker_session=sagemaker_local_session, |
| | hyperparameters={"epochs": 1}, |
| | use_mpi=True, |
| | num_processes=2, |
| | process_slots_per_host=2, |
| | additional_mpi_options="-x NCCL_DEBUG=INFO", |
| | ) |
| |
|
| | train_input = "file://" + os.path.join(data_path, "train") |
| | test_input = "file://" + os.path.join(data_path, "test") |
| |
|
| | training_config = _build_airflow_workflow( |
| | estimator=chainer, |
| | instance_type=cpu_instance_type, |
| | inputs={"train": train_input, "test": test_input}, |
| | ) |
| |
|
| | _assert_that_s3_url_contains_data( |
| | sagemaker_local_session, |
| | training_config["HyperParameters"]["sagemaker_submit_directory"].strip('"'), |
| | ) |
| |
|
| |
|
| | def test_mxnet_airflow_config_uploads_data_source_to_s3( |
| | sagemaker_session, |
| | cpu_instance_type, |
| | mxnet_training_latest_version, |
| | mxnet_training_latest_py_version, |
| | ): |
| | with timeout(seconds=AIRFLOW_CONFIG_TIMEOUT_IN_SECONDS): |
| | script_path = os.path.join(DATA_DIR, "chainer_mnist", "mnist.py") |
| | data_path = os.path.join(DATA_DIR, "chainer_mnist") |
| |
|
| | mx = MXNet( |
| | entry_point=script_path, |
| | role=ROLE, |
| | framework_version=mxnet_training_latest_version, |
| | py_version=mxnet_training_latest_py_version, |
| | instance_count=SINGLE_INSTANCE_COUNT, |
| | instance_type=cpu_instance_type, |
| | sagemaker_session=sagemaker_session, |
| | ) |
| |
|
| | train_input = "file://" + os.path.join(data_path, "train") |
| | test_input = "file://" + os.path.join(data_path, "test") |
| |
|
| | training_config = _build_airflow_workflow( |
| | estimator=mx, |
| | instance_type=cpu_instance_type, |
| | inputs={"train": train_input, "test": test_input}, |
| | ) |
| |
|
| | _assert_that_s3_url_contains_data( |
| | sagemaker_session, |
| | training_config["HyperParameters"]["sagemaker_submit_directory"].strip('"'), |
| | ) |
| |
|
| |
|
| | def test_sklearn_airflow_config_uploads_data_source_to_s3( |
| | sagemaker_session, |
| | cpu_instance_type, |
| | sklearn_latest_version, |
| | sklearn_latest_py_version, |
| | ): |
| | with timeout(seconds=AIRFLOW_CONFIG_TIMEOUT_IN_SECONDS): |
| | script_path = os.path.join(DATA_DIR, "sklearn_mnist", "mnist.py") |
| | data_path = os.path.join(DATA_DIR, "sklearn_mnist") |
| |
|
| | sklearn = SKLearn( |
| | entry_point=script_path, |
| | role=ROLE, |
| | instance_type=cpu_instance_type, |
| | framework_version=sklearn_latest_version, |
| | py_version=sklearn_latest_py_version, |
| | sagemaker_session=sagemaker_session, |
| | hyperparameters={"epochs": 1}, |
| | ) |
| |
|
| | train_input = sklearn.sagemaker_session.upload_data( |
| | path=os.path.join(data_path, "train"), key_prefix="integ-test-data/sklearn_mnist/train" |
| | ) |
| | test_input = sklearn.sagemaker_session.upload_data( |
| | path=os.path.join(data_path, "test"), key_prefix="integ-test-data/sklearn_mnist/test" |
| | ) |
| |
|
| | training_config = _build_airflow_workflow( |
| | estimator=sklearn, |
| | instance_type=cpu_instance_type, |
| | inputs={"train": train_input, "test": test_input}, |
| | ) |
| |
|
| | _assert_that_s3_url_contains_data( |
| | sagemaker_session, |
| | training_config["HyperParameters"]["sagemaker_submit_directory"].strip('"'), |
| | ) |
| |
|
| |
|
| | def test_tf_airflow_config_uploads_data_source_to_s3( |
| | sagemaker_session, |
| | cpu_instance_type, |
| | tf_full_version, |
| | tf_full_py_version, |
| | ): |
| | with timeout(seconds=AIRFLOW_CONFIG_TIMEOUT_IN_SECONDS): |
| | tf = TensorFlow( |
| | entry_point=SCRIPT, |
| | role=ROLE, |
| | instance_count=SINGLE_INSTANCE_COUNT, |
| | instance_type=cpu_instance_type, |
| | sagemaker_session=sagemaker_session, |
| | framework_version=tf_full_version, |
| | py_version=tf_full_py_version, |
| | metric_definitions=[ |
| | {"Name": "train:global_steps", "Regex": r"global_step\/sec:\s(.*)"} |
| | ], |
| | ) |
| | inputs = tf.sagemaker_session.upload_data( |
| | path=os.path.join(TF_MNIST_RESOURCE_PATH, "data"), key_prefix="scriptmode/mnist" |
| | ) |
| |
|
| | training_config = _build_airflow_workflow( |
| | estimator=tf, instance_type=cpu_instance_type, inputs=inputs |
| | ) |
| |
|
| | _assert_that_s3_url_contains_data( |
| | sagemaker_session, |
| | training_config["HyperParameters"]["sagemaker_submit_directory"].strip('"'), |
| | ) |
| |
|
| |
|
| | def test_xgboost_airflow_config_uploads_data_source_to_s3( |
| | sagemaker_session, cpu_instance_type, xgboost_latest_version |
| | ): |
| | with timeout(seconds=AIRFLOW_CONFIG_TIMEOUT_IN_SECONDS): |
| | xgboost = XGBoost( |
| | entry_point=os.path.join(DATA_DIR, "dummy_script.py"), |
| | framework_version=xgboost_latest_version, |
| | py_version="py3", |
| | role=ROLE, |
| | sagemaker_session=sagemaker_session, |
| | instance_type=cpu_instance_type, |
| | instance_count=SINGLE_INSTANCE_COUNT, |
| | base_job_name="XGBoost job", |
| | ) |
| |
|
| | training_config = _build_airflow_workflow( |
| | estimator=xgboost, instance_type=cpu_instance_type |
| | ) |
| |
|
| | _assert_that_s3_url_contains_data( |
| | sagemaker_session, |
| | training_config["HyperParameters"]["sagemaker_submit_directory"].strip('"'), |
| | ) |
| |
|
| |
|
| | def test_pytorch_airflow_config_uploads_data_source_to_s3_when_inputs_not_provided( |
| | sagemaker_session, |
| | cpu_instance_type, |
| | pytorch_inference_latest_version, |
| | pytorch_inference_latest_py_version, |
| | ): |
| | with timeout(seconds=AIRFLOW_CONFIG_TIMEOUT_IN_SECONDS): |
| | estimator = PyTorch( |
| | entry_point=PYTORCH_MNIST_SCRIPT, |
| | role=ROLE, |
| | framework_version=pytorch_inference_latest_version, |
| | py_version=pytorch_inference_latest_py_version, |
| | instance_count=2, |
| | instance_type=cpu_instance_type, |
| | hyperparameters={"epochs": 6, "backend": "gloo"}, |
| | sagemaker_session=sagemaker_session, |
| | ) |
| |
|
| | training_config = _build_airflow_workflow( |
| | estimator=estimator, instance_type=cpu_instance_type |
| | ) |
| |
|
| | _assert_that_s3_url_contains_data( |
| | sagemaker_session, |
| | training_config["HyperParameters"]["sagemaker_submit_directory"].strip('"'), |
| | ) |
| |
|
| |
|
| | def _assert_that_s3_url_contains_data(sagemaker_session, s3_url): |
| | parsed_s3_url = urlparse(s3_url) |
| | s3_request = sagemaker_session.boto_session.client("s3").list_objects_v2( |
| | Bucket=parsed_s3_url.netloc, Prefix=parsed_s3_url.path.lstrip("/") |
| | ) |
| | assert s3_request["KeyCount"] > 0 |
| |
|
| |
|
| | def _build_airflow_workflow(estimator, instance_type, inputs=None, mini_batch_size=None): |
| | training_config = sm_airflow.training_config( |
| | estimator=estimator, inputs=inputs, mini_batch_size=mini_batch_size |
| | ) |
| |
|
| | model = estimator.create_model() |
| | assert model is not None |
| |
|
| | model_config = sm_airflow.model_config(model, instance_type) |
| | assert model_config is not None |
| |
|
| | transform_config = sm_airflow.transform_config_from_estimator( |
| | estimator=estimator, |
| | task_id="transform_config", |
| | task_type="training", |
| | instance_count=SINGLE_INSTANCE_COUNT, |
| | instance_type=estimator.instance_type, |
| | data=inputs, |
| | content_type="text/csv", |
| | input_filter="$", |
| | output_filter="$", |
| | ) |
| |
|
| | default_args = { |
| | "owner": "airflow", |
| | "start_date": utils.dates.days_ago(2), |
| | "provide_context": True, |
| | } |
| |
|
| | dag = DAG("tensorflow_example", default_args=default_args, schedule_interval="@once") |
| |
|
| | train_op = SageMakerTrainingOperator( |
| | task_id="tf_training", config=training_config, wait_for_completion=True, dag=dag |
| | ) |
| |
|
| | transform_op = SageMakerTransformOperator( |
| | task_id="transform_operator", config=transform_config, wait_for_completion=True, dag=dag |
| | ) |
| |
|
| | transform_op.set_upstream(train_op) |
| |
|
| | return training_config |
| |
|