hc99's picture
Add files using upload-large-folder tool
fc0f7bd verified
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License.
import pytest
from fairlearn.postprocessing import ThresholdOptimizer
from fairlearn.reductions import ExponentiatedGradient, GridSearch, EqualizedOdds, \
DemographicParity
try:
from tempeh.execution.azureml.workspace import get_workspace
except ImportError:
raise Exception("fairlearn performance tests require azureml-sdk to be installed.")
from environment_setup import build_package
THRESHOLD_OPTIMIZER = ThresholdOptimizer.__name__
EXPONENTIATED_GRADIENT = ExponentiatedGradient.__name__
GRID_SEARCH = GridSearch.__name__
MEMORY = "memory"
TIME = "time"
ADULT_UCI = 'adult_uci'
COMPAS = 'compas'
RBM_SVM = 'rbm_svm'
DECISION_TREE_CLASSIFIER = 'decision_tree_classifier'
DATASETS = [ADULT_UCI, COMPAS]
PREDICTORS = [RBM_SVM, DECISION_TREE_CLASSIFIER]
MITIGATORS = [THRESHOLD_OPTIMIZER, EXPONENTIATED_GRADIENT, GRID_SEARCH]
class PerfTestConfiguration:
def __init__(self, dataset, predictor, mitigator, disparity_metric):
self.dataset = dataset
self.predictor = predictor
self.mitigator = mitigator
self.disparity_metric = disparity_metric
def __repr__(self):
return "[dataset: {}, predictor: {}, mitigator: {}, disparity_metric: {}]" \
.format(self.dataset, self.predictor, self.mitigator, self.disparity_metric)
def get_all_perf_test_configurations():
perf_test_configurations = []
for dataset in DATASETS:
for predictor in PREDICTORS:
for mitigator in MITIGATORS:
if mitigator == THRESHOLD_OPTIMIZER:
disparity_metrics = ["equalized_odds", "demographic_parity"]
elif mitigator == EXPONENTIATED_GRADIENT:
disparity_metrics = [EqualizedOdds.__name__, DemographicParity.__name__]
elif mitigator == GRID_SEARCH:
disparity_metrics = [EqualizedOdds.__name__, DemographicParity.__name__]
else:
raise Exception("Unknown mitigator {}".format(mitigator))
for disparity_metric in disparity_metrics:
perf_test_configurations.append(
PerfTestConfiguration(dataset, predictor, mitigator, disparity_metric))
return perf_test_configurations
@pytest.fixture(scope="session")
def workspace():
return get_workspace()
@pytest.fixture(scope="session")
def wheel_file():
return build_package()