# # Copyright (c) 2012-2023 Snowflake Computing Inc. All rights reserved. # import pytest from sqlalchemy import Column, Integer, MetaData, Sequence, String, Table from sqlalchemy.sql import select, text from snowflake.sqlalchemy import ( AWSBucket, AzureContainer, CopyFormatter, CopyIntoStorage, CSVFormatter, ExternalStage, JSONFormatter, PARQUETFormatter, ) def test_external_stage(sql_compiler): assert ExternalStage.prepare_namespace("something") == "something." assert ExternalStage.prepare_path("prefix") == "/prefix" # All arguments are handled assert ( sql_compiler( ExternalStage(name="name", path="prefix/path", namespace="namespace") ) == "@namespace.name/prefix/path" ) # defaults don't ruin things assert ( sql_compiler(ExternalStage(name="name", path=None, namespace=None)) == "@name" ) def test_copy_into_location(engine_testaccount, sql_compiler): meta = MetaData() food_items = Table( "python_tests_foods", meta, Column("id", Integer, Sequence("new_user_id_seq"), primary_key=True), Column("name", String), Column("quantity", Integer), ) meta.create_all(engine_testaccount) copy_stmt_1 = CopyIntoStorage( from_=food_items, into=AWSBucket.from_uri("s3://backup").encryption_aws_sse_kms( "1234abcd-12ab-34cd-56ef-1234567890ab" ), formatter=CSVFormatter() .record_delimiter("|") .escape(None) .null_if(["null", "Null"]), ) assert ( sql_compiler(copy_stmt_1) == "COPY INTO 's3://backup' FROM python_tests_foods FILE_FORMAT=(TYPE=csv " "ESCAPE=None NULL_IF=('null', 'Null') RECORD_DELIMITER='|') ENCRYPTION=" "(KMS_KEY_ID='1234abcd-12ab-34cd-56ef-1234567890ab' TYPE='AWS_SSE_KMS')" ) copy_stmt_2 = CopyIntoStorage( from_=select(food_items).where(food_items.c.id == 1), # Test sub-query into=AWSBucket.from_uri("s3://backup") .credentials(aws_role="some_iam_role") .encryption_aws_sse_s3(), formatter=JSONFormatter().file_extension("json").compression("zstd"), ) assert ( sql_compiler(copy_stmt_2) == "COPY INTO 's3://backup' FROM (SELECT python_tests_foods.id, " "python_tests_foods.name, python_tests_foods.quantity FROM python_tests_foods " "WHERE python_tests_foods.id = 1) FILE_FORMAT=(TYPE=json COMPRESSION='zstd' " "FILE_EXTENSION='json') CREDENTIALS=(AWS_ROLE='some_iam_role') " "ENCRYPTION=(TYPE='AWS_SSE_S3')" ) copy_stmt_3 = CopyIntoStorage( from_=food_items, into=AzureContainer.from_uri( "azure://snowflake.blob.core.windows.net/snowpile/backup" ).credentials("token"), formatter=PARQUETFormatter().snappy_compression(True), ) assert ( sql_compiler(copy_stmt_3) == "COPY INTO 'azure://snowflake.blob.core.windows.net/snowpile/backup' " "FROM python_tests_foods FILE_FORMAT=(TYPE=parquet SNAPPY_COMPRESSION=true) " "CREDENTIALS=(AZURE_SAS_TOKEN='token')" ) copy_stmt_3.maxfilesize(50000000) assert ( sql_compiler(copy_stmt_3) == "COPY INTO 'azure://snowflake.blob.core.windows.net/snowpile/backup' " "FROM python_tests_foods FILE_FORMAT=(TYPE=parquet SNAPPY_COMPRESSION=true) " "MAX_FILE_SIZE = 50000000 " "CREDENTIALS=(AZURE_SAS_TOKEN='token')" ) copy_stmt_4 = CopyIntoStorage( from_=AWSBucket.from_uri("s3://backup").encryption_aws_sse_kms( "1234abcd-12ab-34cd-56ef-1234567890ab" ), into=food_items, formatter=CSVFormatter() .record_delimiter("|") .escape(None) .null_if(["null", "Null"]), ) assert ( sql_compiler(copy_stmt_4) == "COPY INTO python_tests_foods FROM 's3://backup' FILE_FORMAT=(TYPE=csv " "ESCAPE=None NULL_IF=('null', 'Null') RECORD_DELIMITER='|') ENCRYPTION=" "(KMS_KEY_ID='1234abcd-12ab-34cd-56ef-1234567890ab' TYPE='AWS_SSE_KMS')" ) copy_stmt_5 = CopyIntoStorage( from_=AWSBucket.from_uri("s3://backup").encryption_aws_sse_kms( "1234abcd-12ab-34cd-56ef-1234567890ab" ), into=food_items, formatter=CSVFormatter().field_delimiter(","), ) assert ( sql_compiler(copy_stmt_5) == "COPY INTO python_tests_foods FROM 's3://backup' FILE_FORMAT=(TYPE=csv " "FIELD_DELIMITER=',') ENCRYPTION=" "(KMS_KEY_ID='1234abcd-12ab-34cd-56ef-1234567890ab' TYPE='AWS_SSE_KMS')" ) copy_stmt_6 = CopyIntoStorage( from_=food_items, into=ExternalStage(name="stage_name"), formatter=CSVFormatter(), ) assert ( sql_compiler(copy_stmt_6) == "COPY INTO @stage_name FROM python_tests_foods FILE_FORMAT=(TYPE=csv)" ) copy_stmt_7 = CopyIntoStorage( from_=food_items, into=ExternalStage(name="stage_name", path="prefix/file", namespace="name"), formatter=CSVFormatter(), ) assert ( sql_compiler(copy_stmt_7) == "COPY INTO @name.stage_name/prefix/file FROM python_tests_foods FILE_FORMAT=(TYPE=csv)" ) # NOTE Other than expect known compiled text, submit it to RegressionTests environment and expect them to fail, but # because of the right reasons acceptable_exc_reasons = { "Failure using stage area", "AWS_ROLE credentials are not allowed for this account.", "AWS_ROLE credentials are invalid", } try: with engine_testaccount.connect() as conn: for stmnt in (copy_stmt_1, copy_stmt_2, copy_stmt_3, copy_stmt_4): with pytest.raises(Exception) as exc: conn.execute(stmnt) if not any( map( lambda reason: reason in str(exc) or reason in str(exc.value), acceptable_exc_reasons, ) ): raise Exception( f"Not acceptable exception: {str(exc)} {str(exc.value)}" ) finally: food_items.drop(engine_testaccount) def test_copy_into_storage_csv_extended(sql_compiler): """ This test compiles the SQL to read CSV data from a stage and insert it into a table. The CSV formatting statements are inserted inline, i.e. no explicit SQL definition of that format is necessary. The Stage is a named stage, i.e. we assume that a CREATE STAGE statement was executed before. This way, the COPY INTO statement does not need to know any security details (credentials or tokens) """ # target table definition (NB: this could be omitted for the test, since the # SQL statement copies the whole CSV and assumes the target structure matches) metadata = MetaData() target_table = Table( "TEST_IMPORT", metadata, Column("COL1", Integer, primary_key=True), Column("COL2", String), ) # define a source stage (root path) root_stage = ExternalStage( name="AZURE_STAGE", namespace="ML_POC.PUBLIC", ) # define a CSV formatter formatter = ( CSVFormatter() .compression("AUTO") .field_delimiter(",") .record_delimiter(r"\n") .field_optionally_enclosed_by(None) .escape(None) .escape_unenclosed_field(r"\134") .date_format("AUTO") .null_if([r"\N"]) .skip_header(1) .trim_space(False) .error_on_column_count_mismatch(True) ) # define CopyInto object; reads all CSV data (=> pattern) from # the sub-path "testdata" beneath the root stage copy_into = CopyIntoStorage( from_=ExternalStage.from_parent_stage(root_stage, "testdata"), into=target_table, formatter=formatter, ) copy_into.copy_options = {"pattern": "'.*csv'", "force": "TRUE"} # check that the result is as expected result = sql_compiler(copy_into) expected = ( r"COPY INTO TEST_IMPORT " r"FROM @ML_POC.PUBLIC.AZURE_STAGE/testdata " r"FILE_FORMAT=(TYPE=csv COMPRESSION='auto' DATE_FORMAT='AUTO' " r"ERROR_ON_COLUMN_COUNT_MISMATCH=True ESCAPE=None " r"ESCAPE_UNENCLOSED_FIELD='\134' FIELD_DELIMITER=',' " r"FIELD_OPTIONALLY_ENCLOSED_BY=None NULL_IF=('\N') RECORD_DELIMITER='\n' " r"SKIP_HEADER=1 TRIM_SPACE=False) force = TRUE pattern = '.*csv'" ) assert result == expected def test_copy_into_storage_parquet_named_format(sql_compiler): """ This test compiles the SQL to read Parquet data from a stage and insert it into a table. The source file is accessed using a SELECT statement. The Parquet formatting definitions are defined in a named format which was explicitly created before. The Stage is a named stage, i.e. we assume that a CREATE STAGE statement was executed before. This way, the COPY INTO statement does not need to know any security details (credentials or tokens) """ # target table definition (NB: this could be omitted for the test, as long as # the statement is not executed) metadata = MetaData() target_table = Table( "TEST_IMPORT", metadata, Column("COL1", Integer, primary_key=True), Column("COL2", String), ) # define a source stage (root path) root_stage = ExternalStage( name="AZURE_STAGE", namespace="ML_POC.PUBLIC", ) # define the SELECT statement to access the source file. # we can probably defined source table metadata and use SQLAlchemy Column objects # instead of texts, but this seems to be the easiest way. sel_statement = select( text("$1:COL1::number"), text("$1:COL2::varchar") ).select_from(ExternalStage.from_parent_stage(root_stage, "testdata/out.parquet")) # use an existing source format. formatter = CopyFormatter(format_name="parquet_file_format") # setup CopyInto object copy_into = CopyIntoStorage( from_=sel_statement, into=target_table, formatter=formatter ) copy_into.copy_options = {"force": "TRUE"} # compile and check the result result = sql_compiler(copy_into) expected = ( "COPY INTO TEST_IMPORT " "FROM (SELECT $1:COL1::number, $1:COL2::varchar " "FROM @ML_POC.PUBLIC.AZURE_STAGE/testdata/out.parquet) " "FILE_FORMAT=(format_name = parquet_file_format) force = TRUE" ) assert result == expected def test_copy_into_storage_parquet_files(sql_compiler): """ This test compiles the SQL to read Parquet data from a stage and insert it into a table. The source file is accessed using a SELECT statement. The Parquet formatting definitions are defined in a named format which was explicitly created before. The format is specified as a property of the stage, not the CopyInto object. The Stage is a named stage, i.e. we assume that a CREATE STAGE statement was executed before. This way, the COPY INTO statement does not need to know any security details (credentials or tokens). The FORCE option is set using the corresponding function in CopyInto. The FILES option is set to choose the files to upload """ # target table definition (NB: this could be omitted for the test, as long as # the statement is not executed) metadata = MetaData() target_table = Table( "TEST_IMPORT", metadata, Column("COL1", Integer, primary_key=True), Column("COL2", String), ) # define a source stage (root path) root_stage = ExternalStage( name="AZURE_STAGE", namespace="ML_POC.PUBLIC", ) # define the SELECT statement to access the source file. # we can probably defined source table metadata and use SQLAlchemy Column objects # instead of texts, but this seems to be the easiest way. sel_statement = select( text("$1:COL1::number"), text("$1:COL2::varchar") ).select_from( ExternalStage.from_parent_stage( root_stage, "testdata/out.parquet", file_format="parquet_file_format" ) ) # setup CopyInto object copy_into = ( CopyIntoStorage( from_=sel_statement, into=target_table, ) .force(True) .files(["foo.txt", "bar.txt"]) ) # compile and check the result result = sql_compiler(copy_into) expected = ( "COPY INTO TEST_IMPORT " "FROM (SELECT $1:COL1::number, $1:COL2::varchar " "FROM @ML_POC.PUBLIC.AZURE_STAGE/testdata/out.parquet " "(file_format => parquet_file_format)) FILES = ('foo.txt','bar.txt') " "FORCE = true" ) assert result == expected def test_copy_into_storage_parquet_pattern(sql_compiler): """ This test compiles the SQL to read Parquet data from a stage and insert it into a table. The source file is accessed using a SELECT statement. The Parquet formatting definitions are defined in a named format which was explicitly created before. The format is specified as a property of the stage, not the CopyInto object. The Stage is a named stage, i.e. we assume that a CREATE STAGE statement was executed before. This way, the COPY INTO statement does not need to know any security details (credentials or tokens). The FORCE option is set using the corresponding function in CopyInto. The PATTERN option is set to choose multiple files """ # target table definition (NB: this could be omitted for the test, as long as # the statement is not executed) metadata = MetaData() target_table = Table( "TEST_IMPORT", metadata, Column("COL1", Integer, primary_key=True), Column("COL2", String), ) # define a source stage (root path) root_stage = ExternalStage( name="AZURE_STAGE", namespace="ML_POC.PUBLIC", ) # define the SELECT statement to access the source file. # we can probably defined source table metadata and use SQLAlchemy Column objects # instead of texts, but this seems to be the easiest way. sel_statement = select( text("$1:COL1::number"), text("$1:COL2::varchar") ).select_from( ExternalStage.from_parent_stage( root_stage, "testdata/out.parquet", file_format="parquet_file_format" ) ) # setup CopyInto object copy_into = ( CopyIntoStorage( from_=sel_statement, into=target_table, ) .force(True) .pattern("'.*csv'") ) # compile and check the result result = sql_compiler(copy_into) expected = ( "COPY INTO TEST_IMPORT " "FROM (SELECT $1:COL1::number, $1:COL2::varchar " "FROM @ML_POC.PUBLIC.AZURE_STAGE/testdata/out.parquet " "(file_format => parquet_file_format)) FORCE = true PATTERN = '.*csv'" ) assert result == expected