id
int64
0
190k
prompt
stringlengths
21
13.4M
docstring
stringlengths
1
12k
152,675
import csv import json import logging import re import os from collections import defaultdict from datetime import datetime, timedelta from urllib import parse from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from pymongo import MongoClient import boto3 from clickhouse_driver import Client a...
null
152,676
import csv import json import logging import re import os from collections import defaultdict from datetime import datetime, timedelta from urllib import parse from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from pymongo import MongoClient import boto3 from clickhouse_driver import Client a...
null
152,677
import csv import json import logging import re import os from collections import defaultdict from datetime import datetime, timedelta from urllib import parse from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from pymongo import MongoClient import boto3 from clickhouse_driver import Client a...
null
152,678
import csv import json import logging import re import os from collections import defaultdict from datetime import datetime, timedelta from urllib import parse from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from pymongo import MongoClient import boto3 from clickhouse_driver import Client a...
null
152,679
import csv import json import logging import re import os from collections import defaultdict from datetime import datetime, timedelta from urllib import parse from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from pymongo import MongoClient import boto3 from clickhouse_driver import Client a...
null
152,680
import csv import json import logging import re import os from collections import defaultdict from datetime import datetime, timedelta from urllib import parse from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from pymongo import MongoClient import boto3 from clickhouse_driver import Client a...
null
152,681
import csv import json import logging import re import os from collections import defaultdict from datetime import datetime, timedelta from urllib import parse from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from pymongo import MongoClient import boto3 from clickhouse_driver import Client a...
null
152,682
import csv import json import logging import re import os from collections import defaultdict from datetime import datetime, timedelta from urllib import parse from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from pymongo import MongoClient import boto3 from clickhouse_driver import Client a...
null
152,683
import csv import json import logging import re import os from collections import defaultdict from datetime import datetime, timedelta from urllib import parse from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from pymongo import MongoClient import boto3 from clickhouse_driver import Client a...
null
152,684
import csv import json import logging import re import os from collections import defaultdict from datetime import datetime, timedelta from urllib import parse from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from pymongo import MongoClient import boto3 from clickhouse_driver import Client a...
null
152,685
import csv import json import logging import re import os from collections import defaultdict from datetime import datetime, timedelta from urllib import parse from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from pymongo import MongoClient import boto3 from clickhouse_driver import Client a...
null
152,686
import csv import json import logging import re import os from collections import defaultdict from datetime import datetime, timedelta from urllib import parse from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from pymongo import MongoClient import boto3 from clickhouse_driver import Client a...
null
152,687
import csv import json import logging import re import os from collections import defaultdict from datetime import datetime, timedelta from urllib import parse from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from pymongo import MongoClient import boto3 from clickhouse_driver import Client a...
null
152,688
import csv import json import logging import re import os from collections import defaultdict from datetime import datetime, timedelta from urllib import parse from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker from pymongo import MongoClient import boto3 from clickhouse_driver import Client a...
null
152,689
import pytz from datetime import datetime def _fix_time(timestamp, l_timezone: pytz.tzfile): result = None if timestamp: # convert UTC timestamp to datetime and replace timezone from UTC to # schedule timezone date_utc = datetime.fromtimestamp(timestamp, tz=pytz.utc) result = l_t...
null
152,690
import logging import os import requests from datetime import datetime from kombu import Connection as QConnection, Exchange from kombu.pools import producers from collections import defaultdict from tools.cloud_adapter.cloud import Cloud from tools.cloud_adapter.model import RES_MODEL_MAP from optscale_client.config_c...
null
152,691
import logging import os import requests from datetime import datetime from kombu import Connection as QConnection, Exchange from kombu.pools import producers from collections import defaultdict from tools.cloud_adapter.cloud import Cloud from tools.cloud_adapter.model import RES_MODEL_MAP from optscale_client.config_c...
null
152,692
import logging import os from datetime import datetime, timedelta from kombu import Connection as QConnection, Exchange from kombu.pools import producers from pymongo import MongoClient from optscale_client.config_client.client import Client as ConfigClient RETRY_POLICY = {'max_retries': 15, 'interval_start': 0, ...
null
152,693
import logging import os from datetime import datetime from kombu import Connection as QConnection, Exchange from kombu.pools import producers from optscale_client.config_client.client import Client as ConfigClient from optscale_client.rest_api_client.client_v2 import Client as RestClient LOG = logging.getLogger(__name...
null
152,694
import logging import os from datetime import datetime from kombu import Connection as QConnection, Exchange from kombu.pools import producers from optscale_client.config_client.client import Client as ConfigClient from optscale_client.rest_api_client.client_v2 import Client as RestClient def get_org_ids(config_cl): ...
null
152,695
import logging import os import tornado from optscale_client.config_client.client import Client as ConfigClient import docker_images.ohsu.handlers.v2 as handlers from docker_images.ohsu.urls import urls_v2 def get_handlers(handler_kwargs): # pylint: disable=E1101 return [ (urls_v2.links, handle...
null
152,696
import logging import os from datetime import datetime from kombu import Connection as QConnection, Exchange from kombu.pools import producers from optscale_client.config_client.client import Client as ConfigClient from optscale_client.rest_api_client.client_v2 import Client as RestClient LOG = logging.getLogger(__name...
null
152,697
import logging import os from datetime import datetime from kombu import Connection as QConnection, Exchange from kombu.pools import producers from optscale_client.config_client.client import Client as ConfigClient from optscale_client.rest_api_client.client_v2 import Client as RestClient def get_org_ids(config_client...
null
152,698
from setuptools import find_packages, setup def sqlglotrs_version(): with open("sqlglotrs/Cargo.toml") as fd: for line in fd.readlines(): if line.strip().startswith("version"): return line.split("=")[1].strip().strip('"') raise ValueError("Could not find version in Cargo.tom...
null
152,699
from importlib import import_module from pathlib import Path from unittest import mock from pdoc.__main__ import cli, parser import sqlglot The provided code snippet includes necessary dependencies for implementing the `mocked_import` function. Write a Python function `def mocked_import(*args, **kwargs)` to solve the ...
Return a MagicMock if import fails for any reason
152,700
import collections.abc import timeit import numpy as np import sqltree import sqlglot def sqlglot_parse(sql): sqlglot.tokens.USE_RS_TOKENIZER = False sqlglot.parse_one(sql, error_level=sqlglot.ErrorLevel.IGNORE)
null
152,701
import collections.abc import timeit import numpy as np import sqltree import sqlglot def sqlglotrs_parse(sql): sqlglot.tokens.USE_RS_TOKENIZER = True sqlglot.parse_one(sql, error_level=sqlglot.ErrorLevel.IGNORE)
null
152,702
import collections.abc import timeit import numpy as np import sqltree import sqlglot def sqltree_parse(sql): sqltree.api.sqltree(sql.replace('"', "`").replace("''", '"'))
null
152,703
import collections.abc import timeit import numpy as np import sqltree import sqlglot def sqlparse_parse(sql): sqlparse.parse(sql)
null
152,704
import collections.abc import timeit import numpy as np import sqltree import sqlglot def moz_sql_parser_parse(sql): moz_sql_parser.parse(sql)
null
152,705
import collections.abc import timeit import numpy as np import sqltree import sqlglot def sqloxide_parse(sql): sqloxide.parse_sql(sql, dialect="ansi")
null
152,706
import collections.abc import timeit import numpy as np import sqltree import sqlglot def sqlfluff_parse(sql): sqlfluff.parse(sql)
null
152,707
import collections.abc import timeit import numpy as np import sqltree import sqlglot def border(columns): columns = " | ".join(columns) return f"| {columns} |"
null
152,708
import collections.abc import timeit import numpy as np import sqltree import sqlglot def diff(row, column): if column == "Query": return "" column = row[column] if isinstance(column, str): return " (N/A)" return f" ({str(column / row['sqlglot'])[0:5]})"
null
152,709
from __future__ import annotations import typing as t from collections import defaultdict from dataclasses import dataclass from heapq import heappop, heappush from sqlglot import Dialect, expressions as exp from sqlglot.helper import ensure_list if t.TYPE_CHECKING: from sqlglot._typing import T Edit = t.Union[...
Returns the list of changes between the source and the target expressions. Examples: >>> diff(parse_one("a + b"), parse_one("a + c")) [ Remove(expression=(COLUMN this: (IDENTIFIER this: b, quoted: False))), Insert(expression=(COLUMN this: (IDENTIFIER this: c, quoted: False))), Keep( source=(ADD this: ...), target=(ADD ...
152,710
from __future__ import annotations import typing as t from collections import defaultdict from dataclasses import dataclass from heapq import heappop, heappush from sqlglot import Dialect, expressions as exp from sqlglot.helper import ensure_list if t.TYPE_CHECKING: from sqlglot._typing import T Edit = t.Union[...
null
152,711
from __future__ import annotations import typing as t from collections import defaultdict from dataclasses import dataclass from heapq import heappop, heappush from sqlglot import Dialect, expressions as exp from sqlglot.helper import ensure_list def _is_same_type(source: exp.Expression, target: exp.Expression) -> boo...
null
152,712
from __future__ import annotations import typing as t from collections import defaultdict from dataclasses import dataclass from heapq import heappop, heappush from sqlglot import Dialect, expressions as exp from sqlglot.helper import ensure_list if t.TYPE_CHECKING: from sqlglot._typing import T Edit = t.Union[...
null
152,713
from __future__ import annotations import typing as t from collections import defaultdict from dataclasses import dataclass from heapq import heappop, heappush from sqlglot import Dialect, expressions as exp from sqlglot.helper import ensure_list if t.TYPE_CHECKING: from sqlglot._typing import T Edit = t.Union[...
null
152,714
from __future__ import annotations import typing as t from collections import defaultdict from dataclasses import dataclass from heapq import heappop, heappush from sqlglot import Dialect, expressions as exp from sqlglot.helper import ensure_list if t.TYPE_CHECKING: from sqlglot._typing import T Edit = t.Union[...
Calculates the longest common subsequence
152,715
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, datestrtodate_sql, build_formatted_time, no_trycast_sql, rename_func, str_position_sql, timestrtotime_sql, ) from sqlglot.trans...
null
152,716
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, datestrtodate_sql, build_formatted_time, no_trycast_sql, rename_func, str_position_sql, timestrtotime_sql, ) from sqlglot.trans...
null
152,717
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, any_value_to_max_sql, arrow_json_extract_sql, concat_to_dpipe_sql, count_if_to_sum, no_ilike_sql, no...
null
152,718
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, any_value_to_max_sql, arrow_json_extract_sql, concat_to_dpipe_sql, count_if_to_sum, no_ilike_sql, no...
null
152,719
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, any_value_to_max_sql, arrow_json_extract_sql, concat_to_dpipe_sql, count_if_to_sum, no_ilike_sql, no...
Move primary key to a column and enforce auto_increment on primary keys.
152,720
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, date_delta_sql, date_trunc_to_time, datestrtodate_sql, build_formatted_time, i...
null
152,721
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, date_delta_sql, date_trunc_to_time, datestrtodate_sql, build_formatted_time, i...
null
152,722
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, date_delta_sql, date_trunc_to_time, datestrtodate_sql, build_formatted_time, i...
null
152,723
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, date_delta_sql, date_trunc_to_time, datestrtodate_sql, build_formatted_time, i...
null
152,724
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, date_delta_sql, date_trunc_to_time, datestrtodate_sql, build_formatted_time, i...
null
152,725
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, date_delta_sql, date_trunc_to_time, datestrtodate_sql, build_formatted_time, i...
null
152,726
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, date_delta_sql, date_trunc_to_time, datestrtodate_sql, build_formatted_time, i...
null
152,727
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, date_delta_sql, date_trunc_to_time, datestrtodate_sql, build_formatted_time, i...
null
152,728
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, date_delta_sql, date_trunc_to_time, datestrtodate_sql, build_formatted_time, i...
null
152,729
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, date_delta_sql, date_trunc_to_time, datestrtodate_sql, build_formatted_time, i...
null
152,730
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, date_delta_sql, date_trunc_to_time, datestrtodate_sql, build_formatted_time, i...
null
152,731
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, date_delta_sql, date_trunc_to_time, datestrtodate_sql, build_formatted_time, i...
null
152,732
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, date_delta_sql, date_trunc_to_time, datestrtodate_sql, build_formatted_time, i...
Snowflake doesn't allow columns referenced in UNPIVOT to be qualified, so we need to unqualify them. Example: >>> from sqlglot import parse_one >>> expr = parse_one("SELECT * FROM m_sales UNPIVOT(sales FOR month IN (m_sales.jan, feb, mar, april))") >>> print(_unqualify_unpivot_columns(expr).sql(dialect="snowflake")) SE...
152,733
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, date_delta_sql, date_trunc_to_time, datestrtodate_sql, build_formatted_time, i...
null
152,734
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, bool_xor_sql, date_trunc_to_time, datestrtodate_sql, encode_decode_sql, build_...
null
152,735
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, bool_xor_sql, date_trunc_to_time, datestrtodate_sql, encode_decode_sql, build_...
null
152,736
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, bool_xor_sql, date_trunc_to_time, datestrtodate_sql, encode_decode_sql, build_...
null
152,737
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, bool_xor_sql, date_trunc_to_time, datestrtodate_sql, encode_decode_sql, build_...
null
152,738
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, bool_xor_sql, date_trunc_to_time, datestrtodate_sql, encode_decode_sql, build_...
null
152,739
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, bool_xor_sql, date_trunc_to_time, datestrtodate_sql, encode_decode_sql, build_...
null
152,740
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, bool_xor_sql, date_trunc_to_time, datestrtodate_sql, encode_decode_sql, build_...
null
152,741
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, bool_xor_sql, date_trunc_to_time, datestrtodate_sql, encode_decode_sql, build_...
null
152,742
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, bool_xor_sql, date_trunc_to_time, datestrtodate_sql, encode_decode_sql, build_...
null
152,743
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, bool_xor_sql, date_trunc_to_time, datestrtodate_sql, encode_decode_sql, build_...
null
152,744
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, bool_xor_sql, date_trunc_to_time, datestrtodate_sql, encode_decode_sql, build_...
null
152,745
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, bool_xor_sql, date_trunc_to_time, datestrtodate_sql, encode_decode_sql, build_...
Trino doesn't support FIRST / LAST as functions, but they're valid in the context of MATCH_RECOGNIZE, so we need to preserve them in that case. In all other cases they're converted into an ARBITRARY call. Reference: https://trino.io/docs/current/sql/match-recognize.html#logical-navigation-functions
152,746
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, bool_xor_sql, date_trunc_to_time, datestrtodate_sql, encode_decode_sql, build_...
null
152,747
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, bool_xor_sql, date_trunc_to_time, datestrtodate_sql, encode_decode_sql, build_...
null
152,748
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, binary_from_function, bool_xor_sql, date_trunc_to_time, datestrtodate_sql, encode_decode_sql, build_...
null
152,749
from __future__ import annotations import typing as t from sqlglot import exp from sqlglot.dialects.dialect import rename_func from sqlglot.dialects.hive import _build_with_ignore_nulls from sqlglot.dialects.spark2 import Spark2, temporary_storage_provider from sqlglot.helper import seq_get from sqlglot.transforms impo...
Although Spark docs don't mention the "unit" argument, Spark3 added support for it at some point. Databricks also supports this variant (see below). For example, in spark-sql (v3.3.1): - SELECT DATEDIFF('2020-01-01', '2020-01-05') results in -4 - SELECT DATEDIFF(day, '2020-01-01', '2020-01-05') results in 4 See also: -...
152,750
from __future__ import annotations import typing as t from sqlglot import exp from sqlglot.dialects.dialect import rename_func from sqlglot.dialects.hive import _build_with_ignore_nulls from sqlglot.dialects.spark2 import Spark2, temporary_storage_provider from sqlglot.helper import seq_get from sqlglot.transforms impo...
Normalize the expressions in PARTITION BY (<expression>, <expression>, ...)
152,751
from __future__ import annotations import typing as t from sqlglot import exp, transforms from sqlglot.dialects.dialect import ( NormalizationStrategy, concat_to_dpipe_sql, concat_ws_to_dpipe_sql, date_delta_sql, generatedasidentitycolumnconstraint_sql, json_extract_segments, no_tablesample_...
null
152,752
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, arrow_json_extract_sql, date_add_interval_sql, datestrtodate_sql, build_formatted_time, isnull_to_is_nul...
null
152,753
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, arrow_json_extract_sql, date_add_interval_sql, datestrtodate_sql, build_formatted_time, isnull_to_is_nul...
null
152,754
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, arrow_json_extract_sql, date_add_interval_sql, datestrtodate_sql, build_formatted_time, isnull_to_is_nul...
null
152,755
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, arrow_json_extract_sql, date_add_interval_sql, datestrtodate_sql, build_formatted_time, isnull_to_is_nul...
null
152,756
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, arrow_json_extract_sql, date_add_interval_sql, datestrtodate_sql, build_formatted_time, isnull_to_is_nul...
null
152,757
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, arrow_json_extract_sql, date_add_interval_sql, datestrtodate_sql, build_formatted_time, isnull_to_is_nul...
null
152,758
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, arrow_json_extract_sql, date_add_interval_sql, datestrtodate_sql, build_formatted_time, isnull_to_is_nul...
null
152,759
from __future__ import annotations import typing as t from sqlglot import exp, transforms from sqlglot.dialects.dialect import ( binary_from_function, build_formatted_time, is_parse_json, pivot_column_names, rename_func, trim_sql, ) from sqlglot.dialects.hive import Hive from sqlglot.helper impo...
null
152,760
from __future__ import annotations import typing as t from sqlglot import exp, transforms from sqlglot.dialects.dialect import ( binary_from_function, build_formatted_time, is_parse_json, pivot_column_names, rename_func, trim_sql, ) from sqlglot.dialects.hive import Hive from sqlglot.helper impo...
null
152,761
from __future__ import annotations import typing as t from sqlglot import exp, transforms from sqlglot.dialects.dialect import ( binary_from_function, build_formatted_time, is_parse_json, pivot_column_names, rename_func, trim_sql, ) from sqlglot.dialects.hive import Hive from sqlglot.helper impo...
null
152,762
from __future__ import annotations import typing as t from sqlglot import exp, transforms from sqlglot.dialects.dialect import ( binary_from_function, build_formatted_time, is_parse_json, pivot_column_names, rename_func, trim_sql, ) from sqlglot.dialects.hive import Hive from sqlglot.helper impo...
null
152,763
from __future__ import annotations import typing as t from sqlglot import exp, transforms from sqlglot.dialects.dialect import ( binary_from_function, build_formatted_time, is_parse_json, pivot_column_names, rename_func, trim_sql, ) from sqlglot.dialects.hive import Hive from sqlglot.helper impo...
Spark doesn't allow PIVOT aliases, so we need to remove them and possibly wrap a pivoted source in a subquery with the same alias to preserve the query's semantics. Example: >>> from sqlglot import parse_one >>> expr = parse_one("SELECT piv.x FROM tbl PIVOT (SUM(a) FOR b IN ('x')) piv") >>> print(_unalias_pivot(expr).s...
152,764
from __future__ import annotations import typing as t from sqlglot import exp, transforms from sqlglot.dialects.dialect import ( binary_from_function, build_formatted_time, is_parse_json, pivot_column_names, rename_func, trim_sql, ) from sqlglot.dialects.hive import Hive from sqlglot.helper impo...
Spark doesn't allow the column referenced in the PIVOT's field to be qualified, so we need to unqualify it. Example: >>> from sqlglot import parse_one >>> expr = parse_one("SELECT * FROM tbl PIVOT (SUM(tbl.sales) FOR tbl.quarter IN ('Q1', 'Q2'))") >>> print(_unqualify_pivot_columns(expr).sql(dialect="spark")) SELECT * ...
152,765
from __future__ import annotations import typing as t from sqlglot import exp, transforms from sqlglot.dialects.dialect import ( binary_from_function, build_formatted_time, is_parse_json, pivot_column_names, rename_func, trim_sql, ) from sqlglot.dialects.hive import Hive from sqlglot.helper impo...
null
152,766
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, arg_max_or_min_no_count, date_delta_sql, inline_array_sql, json_extract_segments, json_path_key_only_name, no_pivot_sql, bu...
null
152,767
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, arg_max_or_min_no_count, date_delta_sql, inline_array_sql, json_extract_segments, json_path_key_only_name, no_pivot_sql, bu...
null
152,768
from __future__ import annotations import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, arg_max_or_min_no_count, date_delta_sql, inline_array_sql, json_extract_segments, json_path_key_only_name, no_pivot_sql, bu...
null
152,769
from __future__ import annotations import datetime import re import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, any_value_to_max_sql, date_delta_sql, generatedasidentitycolumnconstraint_sql, max...
null
152,770
from __future__ import annotations import datetime import re import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, any_value_to_max_sql, date_delta_sql, generatedasidentitycolumnconstraint_sql, max...
null
152,771
from __future__ import annotations import datetime import re import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, any_value_to_max_sql, date_delta_sql, generatedasidentitycolumnconstraint_sql, max...
null
152,772
from __future__ import annotations import datetime import re import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, any_value_to_max_sql, date_delta_sql, generatedasidentitycolumnconstraint_sql, max...
null
152,773
from __future__ import annotations import datetime import re import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, any_value_to_max_sql, date_delta_sql, generatedasidentitycolumnconstraint_sql, max...
null
152,774
from __future__ import annotations import datetime import re import typing as t from sqlglot import exp, generator, parser, tokens, transforms from sqlglot.dialects.dialect import ( Dialect, NormalizationStrategy, any_value_to_max_sql, date_delta_sql, generatedasidentitycolumnconstraint_sql, max...
null