| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """Data wrangler helpers for data ingestion.""" |
| | from __future__ import absolute_import |
| |
|
| | from typing import Dict |
| | from uuid import uuid4 |
| | from sagemaker.dataset_definition.inputs import ( |
| | RedshiftDatasetDefinition, |
| | AthenaDatasetDefinition, |
| | ) |
| |
|
| |
|
| | def generate_data_ingestion_flow_from_s3_input( |
| | input_name: str, |
| | s3_uri: str, |
| | s3_content_type: str = "csv", |
| | s3_has_header: bool = False, |
| | operator_version: str = "0.1", |
| | schema: Dict = None, |
| | ): |
| | """Generate the data ingestion only flow from s3 input |
| | |
| | Args: |
| | input_name (str): the name of the input to flow source node |
| | s3_uri (str): uri for the s3 input to flow source node |
| | s3_content_type (str): s3 input content type |
| | s3_has_header (bool): flag indicating the input has header or not |
| | operator_version: (str): the version of the operator |
| | schema: (typing.Dict): the schema for the data to be ingested |
| | Returns: |
| | dict (typing.Dict): A flow only conduct data ingestion with 1-1 mapping |
| | output_name (str): The output name used to configure |
| | `sagemaker.processing.FeatureStoreOutput` |
| | """ |
| | source_node = { |
| | "node_id": str(uuid4()), |
| | "type": "SOURCE", |
| | "inputs": [], |
| | "outputs": [{"name": "default"}], |
| | "operator": f"sagemaker.s3_source_{operator_version}", |
| | "parameters": { |
| | "dataset_definition": { |
| | "datasetSourceType": "S3", |
| | "name": input_name, |
| | "s3ExecutionContext": { |
| | "s3Uri": s3_uri, |
| | "s3ContentType": s3_content_type, |
| | "s3HasHeader": s3_has_header, |
| | }, |
| | } |
| | }, |
| | } |
| |
|
| | output_node = _get_output_node(source_node["node_id"], operator_version, schema) |
| |
|
| | flow = { |
| | "metadata": {"version": 1, "disable_limits": False}, |
| | "nodes": [source_node, output_node], |
| | } |
| |
|
| | return flow, f'{output_node["node_id"]}.default' |
| |
|
| |
|
| | def generate_data_ingestion_flow_from_athena_dataset_definition( |
| | input_name: str, |
| | athena_dataset_definition: AthenaDatasetDefinition, |
| | operator_version: str = "0.1", |
| | schema: Dict = None, |
| | ): |
| | """Generate the data ingestion only flow from athena input |
| | |
| | Args: |
| | input_name (str): the name of the input to flow source node |
| | athena_dataset_definition (AthenaDatasetDefinition): athena input to flow source node |
| | operator_version: (str): the version of the operator |
| | schema: (typing.Dict): the schema for the data to be ingested |
| | Returns: |
| | dict (typing.Dict): A flow only conduct data ingestion with 1-1 mapping |
| | output_name (str): The output name used to configure |
| | `sagemaker.processing.FeatureStoreOutput` |
| | """ |
| | source_node = { |
| | "node_id": str(uuid4()), |
| | "type": "SOURCE", |
| | "inputs": [], |
| | "outputs": [{"name": "default"}], |
| | "operator": f"sagemaker.athena_source_{operator_version}", |
| | "parameters": { |
| | "dataset_definition": { |
| | "datasetSourceType": "Athena", |
| | "name": input_name, |
| | "catalogName": athena_dataset_definition.catalog, |
| | "databaseName": athena_dataset_definition.database, |
| | "queryString": athena_dataset_definition.query_string, |
| | "s3OutputLocation": athena_dataset_definition.output_s3_uri, |
| | "outputFormat": athena_dataset_definition.output_format, |
| | } |
| | }, |
| | } |
| |
|
| | output_node = _get_output_node(source_node["node_id"], operator_version, schema) |
| |
|
| | flow = { |
| | "metadata": {"version": 1, "disable_limits": False}, |
| | "nodes": [source_node, output_node], |
| | } |
| |
|
| | return flow, f'{output_node["node_id"]}.default' |
| |
|
| |
|
| | def generate_data_ingestion_flow_from_redshift_dataset_definition( |
| | input_name: str, |
| | redshift_dataset_definition: RedshiftDatasetDefinition, |
| | operator_version: str = "0.1", |
| | schema: Dict = None, |
| | ): |
| | """Generate the data ingestion only flow from redshift input |
| | |
| | Args: |
| | input_name (str): the name of the input to flow source node |
| | redshift_dataset_definition (RedshiftDatasetDefinition): redshift input to flow source node |
| | operator_version: (str): the version of the operator |
| | schema: (typing.Dict): the schema for the data to be ingested |
| | Returns: |
| | dict (typing.Dict): A flow only conduct data ingestion with 1-1 mapping |
| | output_name (str): The output name used to configure |
| | `sagemaker.processing.FeatureStoreOutput` |
| | """ |
| | source_node = { |
| | "node_id": str(uuid4()), |
| | "type": "SOURCE", |
| | "inputs": [], |
| | "outputs": [{"name": "default"}], |
| | "operator": f"sagemaker.redshift_source_{operator_version}", |
| | "parameters": { |
| | "dataset_definition": { |
| | "datasetSourceType": "Redshift", |
| | "name": input_name, |
| | "clusterIdentifier": redshift_dataset_definition.cluster_id, |
| | "database": redshift_dataset_definition.database, |
| | "dbUser": redshift_dataset_definition.db_user, |
| | "queryString": redshift_dataset_definition.query_string, |
| | "unloadIamRole": redshift_dataset_definition.cluster_role_arn, |
| | "s3OutputLocation": redshift_dataset_definition.output_s3_uri, |
| | "outputFormat": redshift_dataset_definition.output_format, |
| | } |
| | }, |
| | } |
| |
|
| | output_node = _get_output_node(source_node["node_id"], operator_version, schema) |
| |
|
| | flow = { |
| | "metadata": {"version": 1, "disable_limits": False}, |
| | "nodes": [source_node, output_node], |
| | } |
| |
|
| | return flow, f'{output_node["node_id"]}.default' |
| |
|
| |
|
| | def _get_output_node(source_node_id: str, operator_version: str, schema: Dict): |
| | """A helper function to generate output node, for internal use only |
| | |
| | Args: |
| | source_node_id (str): source node id |
| | operator_version: (str): the version of the operator |
| | schema: (typing.Dict): the schema for the data to be ingested |
| | Returns: |
| | dict (typing.Dict): output node |
| | """ |
| | return { |
| | "node_id": str(uuid4()), |
| | "type": "TRANSFORM", |
| | "operator": f"sagemaker.spark.infer_and_cast_type_{operator_version}", |
| | "trained_parameters": {} if schema is None else schema, |
| | "parameters": {}, |
| | "inputs": [{"name": "default", "node_id": source_node_id, "output_name": "default"}], |
| | "outputs": [{"name": "default"}], |
| | } |
| |
|