| | from sqlglot.optimizer.qualify import qualify |
| | from sqlglot.optimizer.normalize import normalize |
| | from sqlglot.optimizer import optimize |
| | from sqlglot import exp |
| | from sqlglot import parse_one, diff |
| | from sqlglot.errors import ParseError, OptimizeError |
| | import sqlite3 |
| | import threading |
| | from ortools.linear_solver import pywraplp |
| | from sqlglot.diff import Insert, Remove, Update, Move, Keep |
| | from sqlglot.expressions import Alias, TableAlias |
| | from utils import SQLITE_DB_DIR |
| |
|
| |
|
| | def get_bird_db(db_name): |
| | return (SQLITE_DB_DIR / db_name / f"{db_name}.sqlite").resolve() |
| |
|
| | class QueryThread(threading.Thread): |
| | def __init__(self, query, db_str): |
| | threading.Thread.__init__(self) |
| | self.db_str = db_str |
| | self.query = query |
| | self.results = 'Failed to run query' |
| | self.daemon = True |
| |
|
| | def run(self): |
| | conn = sqlite3.connect(self.db_str) |
| | cursor = conn.cursor() |
| | try: |
| | cursor.execute(self.query) |
| | self.results = cursor.fetchall() |
| | except (sqlite3.OperationalError, sqlite3.ProgrammingError) as e: |
| | self.results = "Failed to run query" |
| | conn.close() |
| |
|
| | def execute_sql(query_str, db_str): |
| | TIMEOUT_SECONDS = 45.0 |
| | try: |
| | db_location = get_bird_db(db_str) |
| | |
| | thread = QueryThread(query_str, db_location) |
| | thread.start() |
| | thread.join(TIMEOUT_SECONDS) |
| | if not thread.is_alive(): |
| | generated_result = set([tuple(str(c) for c in r) for r in thread.results]) |
| | else: |
| | generated_result = "Failed to run query" |
| | except (sqlite3.OperationalError, sqlite3.ProgrammingError) as e: |
| | generated_result = "Failed to run query" |
| | return generated_result |
| |
|
| |
|
| | def normalize_commutative_ordering(expression): |
| | normalize_types = {exp.Add, exp.SafeMultiply, exp.Select} |
| |
|
| | for node in expression.walk(bfs=True): |
| | if type(node) in normalize_types: |
| |
|
| | operands = [node.this, node.args.get('expression')] |
| | |
| | try: |
| | operands.sort(key=lambda x: x.sql()) |
| | node.set(arg_key='this', value=operands[0]) |
| | node.set(arg_key='expression', value=operands[1]) |
| | except Exception as e: |
| | |
| | |
| | pass |
| | |
| | |
| | |
| | |
| | return expression |
| |
|
| |
|
| | def remove_table_aliases(sql): |
| | |
| | tab_aliases = {} |
| | for table_exp in sql.find_all(exp.Table): |
| | if table_exp.alias: |
| | tab_aliases[table_exp.alias] = table_exp |
| | for node in sql.walk(): |
| | if isinstance(node, exp.Table) and node.alias: |
| | node.set("alias", None) |
| | elif isinstance(node, exp.Column) and node.table: |
| |
|
| | if node.table in tab_aliases: |
| | node.args['table'] = exp.Table(this=tab_aliases[node.table], quoted=node.args['table'].quoted) |
| |
|
| | |
| |
|
| | return sql |
| |
|
| |
|
| | def cleanup_query(query, schema): |
| | query = query.replace("`", "\"") |
| | try: |
| | parse = parse_one(query) |
| | except ParseError as e: |
| | return None |
| | try: |
| | try: |
| | qual = qualify(parse, dialect='sqlite', schema=schema) |
| | except OptimizeError as e: |
| | qual = qualify(parse, dialect='sqlite') |
| | except OptimizeError as e: |
| | return None |
| | opt_norm = optimize(normalize(qual, dnf=True)) |
| |
|
| | cleaner_sql = remove_table_aliases(opt_norm) |
| | cleaner_sql = normalize_commutative_ordering(cleaner_sql) |
| | return cleaner_sql |
| |
|
| |
|
| | def get_nonalias_diff(q1, q2, schema): |
| | try: |
| | clean_q1 = cleanup_query(q1, schema) |
| | clean_q2 = cleanup_query(q2, schema) |
| | except Exception as e: |
| | |
| | return None, None |
| | if clean_q1 is None or clean_q2 is None: |
| | return None, None |
| | totalcount = 0 |
| | diffcount = 0 |
| | for d in diff(clean_q1, clean_q2): |
| | totalcount += 1 |
| | if isinstance(d, Keep): |
| | continue |
| | elif isinstance(d, Remove) or isinstance(d, Insert): |
| | if not (isinstance(d.expression, Alias) or isinstance(d.expression, |
| | TableAlias)): |
| | diffcount += 1 |
| | elif isinstance(d, Update) or isinstance(d, Move): |
| | |
| | if d.source == d.target: |
| | continue |
| | if not (isinstance(d.source, Alias) or isinstance(d.source, |
| | TableAlias)): |
| | diffcount += 1 |
| | elif not (isinstance(d.target, Alias) or isinstance(d.target, |
| | TableAlias)): |
| | diffcount += 1 |
| | return diffcount, totalcount |
| |
|
| |
|
| | def sql_diff_sim(q1, q2, schema): |
| | diff, tot = get_nonalias_diff(q1, q2, schema) |
| | if diff is None or tot is None: |
| | return None |
| | return 1 - (diff / tot) |
| |
|
| | def solve_optimal_mapping(score_matrix, debug_solver: bool = False): |
| | |
| | solver = pywraplp.Solver.CreateSolver("SCIP") |
| |
|
| | var_list = [] |
| | to_gt_constraints = [] |
| | objective_terms = [] |
| | p_list = [[] for _ in range(len(score_matrix[0]))] |
| | for i in range(len(score_matrix)): |
| | c_list = [] |
| | for j, scoreval in enumerate(score_matrix[i]): |
| | new_var = solver.IntVar(0, 1, "") |
| | var_list.append((new_var, i, j)) |
| | c_list.append(new_var) |
| | p_list[j].append(new_var) |
| | objective_terms.append(new_var * scoreval) |
| | solver.Add(sum(c_list) <= 1) |
| | for pl in p_list: |
| | solver.Add(sum(pl) <= 1) |
| | solver.Maximize(sum(objective_terms)) |
| |
|
| | status = solver.Solve() |
| |
|
| | if debug_solver: |
| | if status == pywraplp.Solver.INFEASIBLE: |
| | print("INFEASIBLE") |
| | elif status == pywraplp.Solver.FEASIBLE: |
| | print("FEASIBLE") |
| |
|
| | soln_gt_to_pred = {} |
| | for var, gt_ind, pred_ind in var_list: |
| | if var.solution_value() == 1: |
| | soln_gt_to_pred[gt_ind] = pred_ind |
| | soln_total_score = solver.Objective().Value() |
| | return soln_gt_to_pred |