File size: 7,155 Bytes
fc0f7bd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License.

import os

from fairlearn.postprocessing import ThresholdOptimizer
from fairlearn.reductions import ExponentiatedGradient, GridSearch

from timed_execution import TimedExecution


_MITIGATION = "mitigation"
_ESTIMATOR_FIT = 'estimator_fit'


def generate_script(request, perf_test_configuration, script_name, script_directory):
    if not os.path.exists(script_directory):
        os.makedirs(script_directory)

    script_lines = []
    add_imports(script_lines)
    script_lines.append("")
    script_lines.append("run = Run.get_context()")
    add_dataset_setup(script_lines, perf_test_configuration)
    add_unconstrained_estimator_fitting(script_lines, perf_test_configuration)
    add_mitigation(script_lines, perf_test_configuration)
    add_additional_metric_calculation(script_lines, perf_test_configuration)
    script_lines.append("")

    print("\n\n{}\n\n".format("="*100))

    with open(os.path.join(script_directory, script_name), 'w') as script_file:  # noqa: E501
        script_file.write("\n".join(script_lines))


def add_imports(script_lines):
    script_lines.append('from time import time')
    script_lines.append('from tempeh.configurations import models, datasets')
    script_lines.append('from fairlearn.postprocessing import ThresholdOptimizer')
    script_lines.append('from fairlearn.reductions import ExponentiatedGradient, GridSearch')
    script_lines.append('from fairlearn.reductions import DemographicParity, EqualizedOdds')
    script_lines.append('from azureml.core.run import Run')


def add_dataset_setup(script_lines, perf_test_configuration):
    script_lines.append('print("Downloading dataset")')
    script_lines.append('dataset = datasets["{}"]()'.format(perf_test_configuration.dataset))
    script_lines.append('X_train, X_test = dataset.get_X()')
    script_lines.append('y_train, y_test = dataset.get_y()')
    script_lines.append('print("Done downloading dataset")')

    if perf_test_configuration.dataset == "adult_uci":
        # sensitive feature is 8th column (sex)
        script_lines.append('sensitive_features_train = X_train[:, 7]')
        script_lines.append('sensitive_features_test = X_test[:, 7]')
    elif perf_test_configuration.dataset == "diabetes_sklearn":
        # sensitive feature is 2nd column (sex)
        # features have been scaled, but sensitive feature needs to be str or int
        script_lines.append('sensitive_features_train = X_train[:, 1].astype(str)')
        script_lines.append('sensitive_features_test = X_test[:, 1].astype(str)')
        # labels can't be floats as of now
        script_lines.append('y_train = y_train.astype(int)')
        script_lines.append('y_test = y_test.astype(int)')
    elif perf_test_configuration.dataset == "compas":
        # sensitive feature is either race or sex
        # TODO add another case where we use sex as well, or both (?)
        script_lines.append('sensitive_features_train, sensitive_features_test = dataset.get_sensitive_features("race")')
        script_lines.append('y_train = y_train.astype(int)')
        script_lines.append('y_test = y_test.astype(int)')
    else:
        raise ValueError("Sensitive features unknown for dataset {}"
                         .format(perf_test_configuration.dataset))


def add_unconstrained_estimator_fitting(script_lines, perf_test_configuration):
    with TimedExecution(_ESTIMATOR_FIT, script_lines):
        script_lines.append('estimator = models["{}"]()'.format(perf_test_configuration.predictor))
        script_lines.append('unconstrained_predictor = models["{}"]()'.format(perf_test_configuration.predictor))
        script_lines.append('unconstrained_predictor.fit(X_train, y_train)')


def add_mitigation(script_lines, perf_test_configuration):
    with TimedExecution(_MITIGATION, script_lines):
        if perf_test_configuration.mitigator == ThresholdOptimizer.__name__:
            script_lines.append('mitigator = ThresholdOptimizer('
                                'unconstrained_predictor=unconstrained_predictor, '
                                'constraints="{}")'.format(perf_test_configuration.disparity_metric))
        elif perf_test_configuration.mitigator == ExponentiatedGradient.__name__:
            script_lines.append('mitigator = ExponentiatedGradient('
                                'estimator=estimator, '
                                'constraints={}())'.format(perf_test_configuration.disparity_metric))
        elif perf_test_configuration.mitigator == GridSearch.__name__:
            script_lines.append('mitigator = GridSearch(estimator=estimator, '
                                'constraints={}())'.format(perf_test_configuration.disparity_metric))
        else:
            raise Exception("Unknown mitigation technique.")

        script_lines.append('mitigator.fit(X_train, y_train, sensitive_features=sensitive_features_train)')

    if perf_test_configuration.mitigator == ThresholdOptimizer.__name__:
        script_lines.append('mitigator.predict('
                            'X_test, '
                            'sensitive_features=sensitive_features_test, '
                            'random_state=1)')
    else:
        script_lines.append('predictions = mitigator.predict(X_test)')


def add_additional_metric_calculation(script_lines, perf_test_configuration):
    # In certain mitigation methods we re-run the estimators many times.
    # For that reason we need metrics to compare the mitigation time with the time that the
    # estimators took since fairlearn only controls the mitigation overhead and not the estimator
    # training time.
    if perf_test_configuration.mitigator == ExponentiatedGradient.__name__:
        script_lines.append("n_oracle_calls = mitigator._expgrad_result.n_oracle_calls")
        script_lines.append("oracle_calls_execution_time = mitigator._expgrad_result.oracle_calls_execution_time")
    elif perf_test_configuration.mitigator == GridSearch.__name__:
        script_lines.append("n_oracle_calls = len(mitigator._all_results)")
        script_lines.append("oracle_calls_execution_time = [result._oracle_call_execution_time for result in mitigator._all_results]")

    if perf_test_configuration.mitigator in [ExponentiatedGradient.__name__, GridSearch.__name__]:
        add_metric_logging_script(script_lines, "metric_logging_script_expgrad_gridsearch.txt")
    elif perf_test_configuration.mitigator in [ThresholdOptimizer.__name__]:
        add_metric_logging_script(script_lines, "metric_logging_script_postprocessing.txt")


def add_metric_logging_script(script_lines, metric_logging_script_file_name):
    skip_lines = [
        "# Copyright (c) Microsoft Corporation. All rights reserved."
        "# Licensed under the MIT License."
    ]
    script_directory = os.path.dirname(__file__)
    with open(os.path.join(script_directory, metric_logging_script_file_name), 'r') as metric_logging_script_file:
        for line in metric_logging_script_file.readlines():
            if line not in skip_lines:
                script_lines.append(line.replace("\n", ""))