| | |
| | |
| |
|
| | import os |
| |
|
| | from fairlearn.postprocessing import ThresholdOptimizer |
| | from fairlearn.reductions import ExponentiatedGradient, GridSearch |
| |
|
| | from timed_execution import TimedExecution |
| |
|
| |
|
| | _MITIGATION = "mitigation" |
| | _ESTIMATOR_FIT = 'estimator_fit' |
| |
|
| |
|
| | def generate_script(request, perf_test_configuration, script_name, script_directory): |
| | if not os.path.exists(script_directory): |
| | os.makedirs(script_directory) |
| |
|
| | script_lines = [] |
| | add_imports(script_lines) |
| | script_lines.append("") |
| | script_lines.append("run = Run.get_context()") |
| | add_dataset_setup(script_lines, perf_test_configuration) |
| | add_unconstrained_estimator_fitting(script_lines, perf_test_configuration) |
| | add_mitigation(script_lines, perf_test_configuration) |
| | add_additional_metric_calculation(script_lines, perf_test_configuration) |
| | script_lines.append("") |
| |
|
| | print("\n\n{}\n\n".format("="*100)) |
| |
|
| | with open(os.path.join(script_directory, script_name), 'w') as script_file: |
| | script_file.write("\n".join(script_lines)) |
| |
|
| |
|
| | def add_imports(script_lines): |
| | script_lines.append('from time import time') |
| | script_lines.append('from tempeh.configurations import models, datasets') |
| | script_lines.append('from fairlearn.postprocessing import ThresholdOptimizer') |
| | script_lines.append('from fairlearn.reductions import ExponentiatedGradient, GridSearch') |
| | script_lines.append('from fairlearn.reductions import DemographicParity, EqualizedOdds') |
| | script_lines.append('from azureml.core.run import Run') |
| |
|
| |
|
| | def add_dataset_setup(script_lines, perf_test_configuration): |
| | script_lines.append('print("Downloading dataset")') |
| | script_lines.append('dataset = datasets["{}"]()'.format(perf_test_configuration.dataset)) |
| | script_lines.append('X_train, X_test = dataset.get_X()') |
| | script_lines.append('y_train, y_test = dataset.get_y()') |
| | script_lines.append('print("Done downloading dataset")') |
| |
|
| | if perf_test_configuration.dataset == "adult_uci": |
| | |
| | script_lines.append('sensitive_features_train = X_train[:, 7]') |
| | script_lines.append('sensitive_features_test = X_test[:, 7]') |
| | elif perf_test_configuration.dataset == "diabetes_sklearn": |
| | |
| | |
| | script_lines.append('sensitive_features_train = X_train[:, 1].astype(str)') |
| | script_lines.append('sensitive_features_test = X_test[:, 1].astype(str)') |
| | |
| | script_lines.append('y_train = y_train.astype(int)') |
| | script_lines.append('y_test = y_test.astype(int)') |
| | elif perf_test_configuration.dataset == "compas": |
| | |
| | |
| | script_lines.append('sensitive_features_train, sensitive_features_test = dataset.get_sensitive_features("race")') |
| | script_lines.append('y_train = y_train.astype(int)') |
| | script_lines.append('y_test = y_test.astype(int)') |
| | else: |
| | raise ValueError("Sensitive features unknown for dataset {}" |
| | .format(perf_test_configuration.dataset)) |
| |
|
| |
|
| | def add_unconstrained_estimator_fitting(script_lines, perf_test_configuration): |
| | with TimedExecution(_ESTIMATOR_FIT, script_lines): |
| | script_lines.append('estimator = models["{}"]()'.format(perf_test_configuration.predictor)) |
| | script_lines.append('unconstrained_predictor = models["{}"]()'.format(perf_test_configuration.predictor)) |
| | script_lines.append('unconstrained_predictor.fit(X_train, y_train)') |
| |
|
| |
|
| | def add_mitigation(script_lines, perf_test_configuration): |
| | with TimedExecution(_MITIGATION, script_lines): |
| | if perf_test_configuration.mitigator == ThresholdOptimizer.__name__: |
| | script_lines.append('mitigator = ThresholdOptimizer(' |
| | 'unconstrained_predictor=unconstrained_predictor, ' |
| | 'constraints="{}")'.format(perf_test_configuration.disparity_metric)) |
| | elif perf_test_configuration.mitigator == ExponentiatedGradient.__name__: |
| | script_lines.append('mitigator = ExponentiatedGradient(' |
| | 'estimator=estimator, ' |
| | 'constraints={}())'.format(perf_test_configuration.disparity_metric)) |
| | elif perf_test_configuration.mitigator == GridSearch.__name__: |
| | script_lines.append('mitigator = GridSearch(estimator=estimator, ' |
| | 'constraints={}())'.format(perf_test_configuration.disparity_metric)) |
| | else: |
| | raise Exception("Unknown mitigation technique.") |
| |
|
| | script_lines.append('mitigator.fit(X_train, y_train, sensitive_features=sensitive_features_train)') |
| |
|
| | if perf_test_configuration.mitigator == ThresholdOptimizer.__name__: |
| | script_lines.append('mitigator.predict(' |
| | 'X_test, ' |
| | 'sensitive_features=sensitive_features_test, ' |
| | 'random_state=1)') |
| | else: |
| | script_lines.append('predictions = mitigator.predict(X_test)') |
| |
|
| |
|
| | def add_additional_metric_calculation(script_lines, perf_test_configuration): |
| | |
| | |
| | |
| | |
| | if perf_test_configuration.mitigator == ExponentiatedGradient.__name__: |
| | script_lines.append("n_oracle_calls = mitigator._expgrad_result.n_oracle_calls") |
| | script_lines.append("oracle_calls_execution_time = mitigator._expgrad_result.oracle_calls_execution_time") |
| | elif perf_test_configuration.mitigator == GridSearch.__name__: |
| | script_lines.append("n_oracle_calls = len(mitigator._all_results)") |
| | script_lines.append("oracle_calls_execution_time = [result._oracle_call_execution_time for result in mitigator._all_results]") |
| |
|
| | if perf_test_configuration.mitigator in [ExponentiatedGradient.__name__, GridSearch.__name__]: |
| | add_metric_logging_script(script_lines, "metric_logging_script_expgrad_gridsearch.txt") |
| | elif perf_test_configuration.mitigator in [ThresholdOptimizer.__name__]: |
| | add_metric_logging_script(script_lines, "metric_logging_script_postprocessing.txt") |
| |
|
| |
|
| | def add_metric_logging_script(script_lines, metric_logging_script_file_name): |
| | skip_lines = [ |
| | "# Copyright (c) Microsoft Corporation. All rights reserved." |
| | "# Licensed under the MIT License." |
| | ] |
| | script_directory = os.path.dirname(__file__) |
| | with open(os.path.join(script_directory, metric_logging_script_file_name), 'r') as metric_logging_script_file: |
| | for line in metric_logging_script_file.readlines(): |
| | if line not in skip_lines: |
| | script_lines.append(line.replace("\n", "")) |
| |
|