File size: 7,155 Bytes
fc0f7bd | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 | # Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License.
import os
from fairlearn.postprocessing import ThresholdOptimizer
from fairlearn.reductions import ExponentiatedGradient, GridSearch
from timed_execution import TimedExecution
_MITIGATION = "mitigation"
_ESTIMATOR_FIT = 'estimator_fit'
def generate_script(request, perf_test_configuration, script_name, script_directory):
if not os.path.exists(script_directory):
os.makedirs(script_directory)
script_lines = []
add_imports(script_lines)
script_lines.append("")
script_lines.append("run = Run.get_context()")
add_dataset_setup(script_lines, perf_test_configuration)
add_unconstrained_estimator_fitting(script_lines, perf_test_configuration)
add_mitigation(script_lines, perf_test_configuration)
add_additional_metric_calculation(script_lines, perf_test_configuration)
script_lines.append("")
print("\n\n{}\n\n".format("="*100))
with open(os.path.join(script_directory, script_name), 'w') as script_file: # noqa: E501
script_file.write("\n".join(script_lines))
def add_imports(script_lines):
script_lines.append('from time import time')
script_lines.append('from tempeh.configurations import models, datasets')
script_lines.append('from fairlearn.postprocessing import ThresholdOptimizer')
script_lines.append('from fairlearn.reductions import ExponentiatedGradient, GridSearch')
script_lines.append('from fairlearn.reductions import DemographicParity, EqualizedOdds')
script_lines.append('from azureml.core.run import Run')
def add_dataset_setup(script_lines, perf_test_configuration):
script_lines.append('print("Downloading dataset")')
script_lines.append('dataset = datasets["{}"]()'.format(perf_test_configuration.dataset))
script_lines.append('X_train, X_test = dataset.get_X()')
script_lines.append('y_train, y_test = dataset.get_y()')
script_lines.append('print("Done downloading dataset")')
if perf_test_configuration.dataset == "adult_uci":
# sensitive feature is 8th column (sex)
script_lines.append('sensitive_features_train = X_train[:, 7]')
script_lines.append('sensitive_features_test = X_test[:, 7]')
elif perf_test_configuration.dataset == "diabetes_sklearn":
# sensitive feature is 2nd column (sex)
# features have been scaled, but sensitive feature needs to be str or int
script_lines.append('sensitive_features_train = X_train[:, 1].astype(str)')
script_lines.append('sensitive_features_test = X_test[:, 1].astype(str)')
# labels can't be floats as of now
script_lines.append('y_train = y_train.astype(int)')
script_lines.append('y_test = y_test.astype(int)')
elif perf_test_configuration.dataset == "compas":
# sensitive feature is either race or sex
# TODO add another case where we use sex as well, or both (?)
script_lines.append('sensitive_features_train, sensitive_features_test = dataset.get_sensitive_features("race")')
script_lines.append('y_train = y_train.astype(int)')
script_lines.append('y_test = y_test.astype(int)')
else:
raise ValueError("Sensitive features unknown for dataset {}"
.format(perf_test_configuration.dataset))
def add_unconstrained_estimator_fitting(script_lines, perf_test_configuration):
with TimedExecution(_ESTIMATOR_FIT, script_lines):
script_lines.append('estimator = models["{}"]()'.format(perf_test_configuration.predictor))
script_lines.append('unconstrained_predictor = models["{}"]()'.format(perf_test_configuration.predictor))
script_lines.append('unconstrained_predictor.fit(X_train, y_train)')
def add_mitigation(script_lines, perf_test_configuration):
with TimedExecution(_MITIGATION, script_lines):
if perf_test_configuration.mitigator == ThresholdOptimizer.__name__:
script_lines.append('mitigator = ThresholdOptimizer('
'unconstrained_predictor=unconstrained_predictor, '
'constraints="{}")'.format(perf_test_configuration.disparity_metric))
elif perf_test_configuration.mitigator == ExponentiatedGradient.__name__:
script_lines.append('mitigator = ExponentiatedGradient('
'estimator=estimator, '
'constraints={}())'.format(perf_test_configuration.disparity_metric))
elif perf_test_configuration.mitigator == GridSearch.__name__:
script_lines.append('mitigator = GridSearch(estimator=estimator, '
'constraints={}())'.format(perf_test_configuration.disparity_metric))
else:
raise Exception("Unknown mitigation technique.")
script_lines.append('mitigator.fit(X_train, y_train, sensitive_features=sensitive_features_train)')
if perf_test_configuration.mitigator == ThresholdOptimizer.__name__:
script_lines.append('mitigator.predict('
'X_test, '
'sensitive_features=sensitive_features_test, '
'random_state=1)')
else:
script_lines.append('predictions = mitigator.predict(X_test)')
def add_additional_metric_calculation(script_lines, perf_test_configuration):
# In certain mitigation methods we re-run the estimators many times.
# For that reason we need metrics to compare the mitigation time with the time that the
# estimators took since fairlearn only controls the mitigation overhead and not the estimator
# training time.
if perf_test_configuration.mitigator == ExponentiatedGradient.__name__:
script_lines.append("n_oracle_calls = mitigator._expgrad_result.n_oracle_calls")
script_lines.append("oracle_calls_execution_time = mitigator._expgrad_result.oracle_calls_execution_time")
elif perf_test_configuration.mitigator == GridSearch.__name__:
script_lines.append("n_oracle_calls = len(mitigator._all_results)")
script_lines.append("oracle_calls_execution_time = [result._oracle_call_execution_time for result in mitigator._all_results]")
if perf_test_configuration.mitigator in [ExponentiatedGradient.__name__, GridSearch.__name__]:
add_metric_logging_script(script_lines, "metric_logging_script_expgrad_gridsearch.txt")
elif perf_test_configuration.mitigator in [ThresholdOptimizer.__name__]:
add_metric_logging_script(script_lines, "metric_logging_script_postprocessing.txt")
def add_metric_logging_script(script_lines, metric_logging_script_file_name):
skip_lines = [
"# Copyright (c) Microsoft Corporation. All rights reserved."
"# Licensed under the MIT License."
]
script_directory = os.path.dirname(__file__)
with open(os.path.join(script_directory, metric_logging_script_file_name), 'r') as metric_logging_script_file:
for line in metric_logging_script_file.readlines():
if line not in skip_lines:
script_lines.append(line.replace("\n", ""))
|