| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | """Scripts for benchmarking BigQuery queries performance.""" |
| |
|
| | import argparse |
| | from datetime import datetime |
| | import json |
| | import os |
| |
|
| | from google.api_core import exceptions |
| |
|
| | from google.cloud import bigquery |
| |
|
| | _run_schema = [ |
| | bigquery.SchemaField("groupname", "STRING", mode="NULLABLE"), |
| | bigquery.SchemaField("name", "STRING", mode="NULLABLE"), |
| | bigquery.SchemaField( |
| | "tags", |
| | "RECORD", |
| | mode="REPEATED", |
| | fields=[ |
| | bigquery.SchemaField("key", "STRING", mode="NULLABLE"), |
| | bigquery.SchemaField("value", "STRING", mode="NULLABLE"), |
| | ], |
| | ), |
| | bigquery.SchemaField("SQL", "STRING", mode="NULLABLE"), |
| | bigquery.SchemaField( |
| | "runs", |
| | "RECORD", |
| | mode="REPEATED", |
| | fields=[ |
| | bigquery.SchemaField("errorstring", "STRING", mode="NULLABLE"), |
| | bigquery.SchemaField("start_time", "TIMESTAMP", mode="NULLABLE"), |
| | bigquery.SchemaField("query_end_time", "TIMESTAMP", mode="NULLABLE"), |
| | bigquery.SchemaField( |
| | "first_row_returned_time", "TIMESTAMP", mode="NULLABLE" |
| | ), |
| | bigquery.SchemaField( |
| | "all_rows_returned_time", "TIMESTAMP", mode="NULLABLE" |
| | ), |
| | bigquery.SchemaField("total_rows", "INTEGER", mode="NULLABLE"), |
| | ], |
| | ), |
| | bigquery.SchemaField("event_time", "TIMESTAMP", mode="NULLABLE"), |
| | ] |
| |
|
| |
|
| | def _check_pos_int(value): |
| | """Verifies the value is a positive integer.""" |
| | ivalue = int(value) |
| | if ivalue <= 0: |
| | raise argparse.ArgumentTypeError( |
| | f"Argument rerun should be positive int. Actual value: {value}" |
| | ) |
| | return ivalue |
| |
|
| |
|
| | def _parse_tag(tag): |
| | """Parses input tag into key value pair as a dict.""" |
| | tagstring = str(tag) |
| | key, value = tagstring.split(":") |
| | if not key or not value: |
| | raise argparse.ArgumentTypeError( |
| | "key and value in tag need to be non-empty. Actual value: " |
| | + f"key={key}, value={value}" |
| | ) |
| | return {"key": key, "value": value} |
| |
|
| |
|
| | def _parse_args() -> dict: |
| | """Parses input flags.""" |
| | parser = argparse.ArgumentParser(description="Benchmark for BigQuery.") |
| |
|
| | parser.add_argument( |
| | "--reruns", |
| | action="store", |
| | type=_check_pos_int, |
| | default=3, |
| | metavar="", |
| | help="how many times each query is run. Must be a positive integer." |
| | + "Default 3 times", |
| | ) |
| |
|
| | parser.add_argument( |
| | "--projectid", |
| | action="store", |
| | type=str, |
| | metavar="", |
| | help="run benchmarks in a different project. If unset, the " |
| | + "GOOGLE_CLOUD_PROJECT environment variable is used", |
| | ) |
| |
|
| | parser.add_argument( |
| | "--queryfile", |
| | action="store", |
| | type=str, |
| | metavar="", |
| | default="queries.json", |
| | help="override the default file which contains queries to be instrumented", |
| | ) |
| |
|
| | parser.add_argument( |
| | "--table", |
| | action="store", |
| | type=str, |
| | metavar="", |
| | help="specify a table to which benchmarking results should be " |
| | + "streamed. The format for this string is in BigQuery standard SQL " |
| | + "notation without escapes, e.g. projectid.datasetid.tableid", |
| | ) |
| |
|
| | parser.add_argument( |
| | "--create_table", |
| | action="store_true", |
| | help="let the benchmarking tool create the destination table prior to" |
| | + " streaming; if set, also need to set --table to specify table name", |
| | ) |
| |
|
| | parser.add_argument( |
| | "--tag", |
| | action="append", |
| | type=_parse_tag, |
| | metavar="", |
| | help="set arbitrary key:value pairs, can be set multiple times", |
| | ) |
| |
|
| | args = parser.parse_args() |
| | args_dict = vars(args) |
| |
|
| | |
| | if not args_dict.get("projectid"): |
| | if projectid_env := os.environ["GOOGLE_CLOUD_PROJECT"]: |
| | args_dict["projectid"] = projectid_env |
| | else: |
| | raise ValueError( |
| | "Must provide --projectid or set " |
| | "GOOGLE_CLOUD_PROJECT environment variable" |
| | ) |
| |
|
| | |
| | if args_dict.get("create_table") and not args_dict.get("table"): |
| | raise ValueError( |
| | "When --create_table is present, must specify table name with --table" |
| | ) |
| |
|
| | return args_dict |
| |
|
| |
|
| | def _prepare_table(client, create_table: bool, table_name: str) -> str: |
| | """Ensures a table exists, and optionally creates it if directed.""" |
| |
|
| | |
| | parts = table_name.split(".") |
| | if len(parts) != 3: |
| | raise ValueError(f"Expected table in p.d.t format, got: {table_name}") |
| |
|
| | table = bigquery.Table(table_name, schema=_run_schema) |
| |
|
| | |
| | if create_table: |
| | table = client.create_table(table) |
| | print(f"Created table {table.project}.{table.dataset_id}." f"{table.table_id}") |
| |
|
| | |
| | client.get_table(table_name) |
| | return table_name |
| |
|
| |
|
| | def _run_query(client, query: str, rerun: int) -> list: |
| | """Runs individual query for `rerun` times, and returns run results.""" |
| | runs = [] |
| |
|
| | for _ in range(rerun): |
| | print(".", end="", flush=True) |
| | run = {} |
| | num_rows = 0 |
| | num_cols = 0 |
| | start_time = datetime.now() |
| | first_row_time = datetime.min |
| | end_time = datetime.min |
| |
|
| | job = client.query(query) |
| | query_end_time = datetime.now() |
| |
|
| | try: |
| | rows = job.result() |
| | for row in rows: |
| | if num_rows == 0: |
| | num_cols = len(row) |
| | first_row_time = datetime.now() |
| | elif num_cols != len(row): |
| | raise RuntimeError(f"found {len(row)} columns, expected {num_cols}") |
| | num_rows += 1 |
| | end_time = datetime.now() |
| | except exceptions.BadRequest as exc: |
| | run["errorstring"] = repr(exc) |
| |
|
| | run["start_time"] = start_time.isoformat() |
| | run["query_end_time"] = query_end_time.isoformat() |
| | run["first_row_returned_time"] = first_row_time.isoformat() |
| | run["all_rows_returned_time"] = end_time.isoformat() |
| | run["total_rows"] = num_rows |
| | runs.append(run) |
| |
|
| | print("") |
| | return runs |
| |
|
| |
|
| | def _get_delta(time_str_1: str, time_str_2: str) -> str: |
| | """Calculates delta of two ISO format time string, and return as a string.""" |
| | time_1 = datetime.fromisoformat(time_str_1) |
| | time_2 = datetime.fromisoformat(time_str_2) |
| | delta = time_1 - time_2 |
| | return str(delta) |
| |
|
| |
|
| | def _is_datetime_min(time_str: str) -> bool: |
| | return datetime.fromisoformat(time_str) == datetime.min |
| |
|
| |
|
| | def _summary(run: dict) -> str: |
| | """Converts run dict to run summary string.""" |
| | no_val = "NODATA" |
| | output = ["QUERYTIME "] |
| |
|
| | if not _is_datetime_min(run.get("query_end_time")): |
| | output.append(f"{_get_delta(run.get('query_end_time'), run.get('start_time'))}") |
| | else: |
| | output.append(no_val) |
| | output.append(" FIRSTROW ") |
| |
|
| | if not _is_datetime_min(run.get("first_row_returned_time")): |
| | output.append( |
| | f"{_get_delta(run.get('first_row_returned_time'), run.get('start_time'))}" |
| | ) |
| | else: |
| | output.append(no_val) |
| | output += " ALLROWS " |
| |
|
| | if not _is_datetime_min(run.get("all_rows_returned_time")): |
| | output.append( |
| | f"{_get_delta(run.get('all_rows_returned_time'), run.get('start_time'))}" |
| | ) |
| | else: |
| | output.append(no_val) |
| |
|
| | if run.get("total_rows"): |
| | output.append(f" ROWS {run.get('total_rows')}") |
| | if run.get("errorstring"): |
| | output.append(f" ERRORED {run.get('errorstring')}") |
| |
|
| | return "".join(output) |
| |
|
| |
|
| | def _print_results(profiles: list): |
| | for i, prof in enumerate(profiles): |
| | print(f"{i+1}: ({prof['groupname']}:{prof['name']})") |
| | print(f"SQL: {prof['SQL']}") |
| | print("MEASUREMENTS") |
| | for j, run in enumerate(prof["runs"]): |
| | print(f"\t\t({j}) {_summary(run)}") |
| |
|
| |
|
| | def _run_benchmarks(args: dict) -> list: |
| | client = bigquery.Client() |
| |
|
| | |
| | |
| | table_id = "" |
| | if args.get("create_table") or args.get("table"): |
| | table_id = _prepare_table(client, args.get("create_table"), args.get("table")) |
| |
|
| | queries_file = args.get("queryfile") |
| | with open(queries_file, "r") as f: |
| | groups = json.loads(f.read()) |
| |
|
| | measure_start = datetime.now() |
| | profiles = [] |
| | for group_name, group in groups.items(): |
| | for name, query in group.items(): |
| | print(f"Measuring {group_name} : {name}", end="", flush=True) |
| | event_time = datetime.now() |
| | runs = _run_query(client, query, args.get("reruns")) |
| |
|
| | profile = {} |
| | profile["groupname"] = group_name |
| | profile["name"] = name |
| | profile["tags"] = args.get("tag") or [] |
| | profile["SQL"] = query |
| | profile["runs"] = runs |
| | profile["event_time"] = event_time.isoformat() |
| | profiles.append(profile) |
| |
|
| | measure_end = datetime.now() |
| | print(f"Measurement time: {str(measure_end-measure_start)}") |
| |
|
| | |
| | if table_id: |
| | print(f"Streaming test results to table {table_id}...") |
| | errors = client.insert_rows_json(table_id, profiles) |
| | if errors: |
| | raise RuntimeError(f"Cannot upload queries profiles: {errors}") |
| | print("Streaming complete.") |
| |
|
| | return profiles |
| |
|
| |
|
| | if __name__ == "__main__": |
| | args = _parse_args() |
| | profiles = _run_benchmarks(args) |
| | _print_results(profiles) |
| |
|