code stringlengths 66 870k | docstring stringlengths 19 26.7k | func_name stringlengths 1 138 | language stringclasses 1
value | repo stringlengths 7 68 | path stringlengths 5 324 | url stringlengths 46 389 | license stringclasses 7
values |
|---|---|---|---|---|---|---|---|
def threshold_(self):
"""Threshold value used for feature selection."""
scores = _get_feature_importances(
estimator=self.estimator_,
getter=self.importance_getter,
transform_func="norm",
norm_order=self.norm_order,
)
return _calculate_thre... | Threshold value used for feature selection. | threshold_ | python | scikit-learn/scikit-learn | sklearn/feature_selection/_from_model.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_from_model.py | BSD-3-Clause |
def partial_fit(self, X, y=None, **partial_fit_params):
"""Fit the SelectFromModel meta-transformer only once.
Parameters
----------
X : array-like of shape (n_samples, n_features)
The training input samples.
y : array-like of shape (n_samples,), default=None
... | Fit the SelectFromModel meta-transformer only once.
Parameters
----------
X : array-like of shape (n_samples, n_features)
The training input samples.
y : array-like of shape (n_samples,), default=None
The target values (integers that correspond to classes in
... | partial_fit | python | scikit-learn/scikit-learn | sklearn/feature_selection/_from_model.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_from_model.py | BSD-3-Clause |
def n_features_in_(self):
"""Number of features seen during `fit`."""
# For consistency with other estimators we raise a AttributeError so
# that hasattr() fails if the estimator isn't fitted.
try:
check_is_fitted(self)
except NotFittedError as nfe:
raise ... | Number of features seen during `fit`. | n_features_in_ | python | scikit-learn/scikit-learn | sklearn/feature_selection/_from_model.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_from_model.py | BSD-3-Clause |
def _compute_mi_cc(x, y, n_neighbors):
"""Compute mutual information between two continuous variables.
Parameters
----------
x, y : ndarray, shape (n_samples,)
Samples of two continuous random variables, must have an identical
shape.
n_neighbors : int
Number of nearest neig... | Compute mutual information between two continuous variables.
Parameters
----------
x, y : ndarray, shape (n_samples,)
Samples of two continuous random variables, must have an identical
shape.
n_neighbors : int
Number of nearest neighbors to search for each point, see [1]_.
... | _compute_mi_cc | python | scikit-learn/scikit-learn | sklearn/feature_selection/_mutual_info.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_mutual_info.py | BSD-3-Clause |
def _compute_mi_cd(c, d, n_neighbors):
"""Compute mutual information between continuous and discrete variables.
Parameters
----------
c : ndarray, shape (n_samples,)
Samples of a continuous random variable.
d : ndarray, shape (n_samples,)
Samples of a discrete random variable.
... | Compute mutual information between continuous and discrete variables.
Parameters
----------
c : ndarray, shape (n_samples,)
Samples of a continuous random variable.
d : ndarray, shape (n_samples,)
Samples of a discrete random variable.
n_neighbors : int
Number of nearest n... | _compute_mi_cd | python | scikit-learn/scikit-learn | sklearn/feature_selection/_mutual_info.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_mutual_info.py | BSD-3-Clause |
def _compute_mi(x, y, x_discrete, y_discrete, n_neighbors=3):
"""Compute mutual information between two variables.
This is a simple wrapper which selects a proper function to call based on
whether `x` and `y` are discrete or not.
"""
if x_discrete and y_discrete:
return mutual_info_score(x,... | Compute mutual information between two variables.
This is a simple wrapper which selects a proper function to call based on
whether `x` and `y` are discrete or not.
| _compute_mi | python | scikit-learn/scikit-learn | sklearn/feature_selection/_mutual_info.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_mutual_info.py | BSD-3-Clause |
def _iterate_columns(X, columns=None):
"""Iterate over columns of a matrix.
Parameters
----------
X : ndarray or csc_matrix, shape (n_samples, n_features)
Matrix over which to iterate.
columns : iterable or None, default=None
Indices of columns to iterate over. If None, iterate ove... | Iterate over columns of a matrix.
Parameters
----------
X : ndarray or csc_matrix, shape (n_samples, n_features)
Matrix over which to iterate.
columns : iterable or None, default=None
Indices of columns to iterate over. If None, iterate over all columns.
Yields
------
x : ... | _iterate_columns | python | scikit-learn/scikit-learn | sklearn/feature_selection/_mutual_info.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_mutual_info.py | BSD-3-Clause |
def _estimate_mi(
X,
y,
*,
discrete_features="auto",
discrete_target=False,
n_neighbors=3,
copy=True,
random_state=None,
n_jobs=None,
):
"""Estimate mutual information between the features and the target.
Parameters
----------
X : array-like or sparse matrix, shape (... | Estimate mutual information between the features and the target.
Parameters
----------
X : array-like or sparse matrix, shape (n_samples, n_features)
Feature matrix.
y : array-like of shape (n_samples,)
Target vector.
discrete_features : {'auto', bool, array-like}, default='auto'
... | _estimate_mi | python | scikit-learn/scikit-learn | sklearn/feature_selection/_mutual_info.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_mutual_info.py | BSD-3-Clause |
def mutual_info_regression(
X,
y,
*,
discrete_features="auto",
n_neighbors=3,
copy=True,
random_state=None,
n_jobs=None,
):
"""Estimate mutual information for a continuous target variable.
Mutual information (MI) [1]_ between two random variables is a non-negative
value, whi... | Estimate mutual information for a continuous target variable.
Mutual information (MI) [1]_ between two random variables is a non-negative
value, which measures the dependency between the variables. It is equal
to zero if and only if two random variables are independent, and higher
values mean higher de... | mutual_info_regression | python | scikit-learn/scikit-learn | sklearn/feature_selection/_mutual_info.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_mutual_info.py | BSD-3-Clause |
def mutual_info_classif(
X,
y,
*,
discrete_features="auto",
n_neighbors=3,
copy=True,
random_state=None,
n_jobs=None,
):
"""Estimate mutual information for a discrete target variable.
Mutual information (MI) [1]_ between two random variables is a non-negative
value, which me... | Estimate mutual information for a discrete target variable.
Mutual information (MI) [1]_ between two random variables is a non-negative
value, which measures the dependency between the variables. It is equal
to zero if and only if two random variables are independent, and higher
values mean higher depe... | mutual_info_classif | python | scikit-learn/scikit-learn | sklearn/feature_selection/_mutual_info.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_mutual_info.py | BSD-3-Clause |
def _rfe_single_fit(rfe, estimator, X, y, train, test, scorer, routed_params):
"""
Return the score and n_features per step for a fit across one fold.
"""
X_train, y_train = _safe_split(estimator, X, y, train)
X_test, y_test = _safe_split(estimator, X, y, test, train)
fit_params = _check_method_... |
Return the score and n_features per step for a fit across one fold.
| _rfe_single_fit | python | scikit-learn/scikit-learn | sklearn/feature_selection/_rfe.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_rfe.py | BSD-3-Clause |
def fit(self, X, y, **fit_params):
"""Fit the RFE model and then the underlying estimator on the selected features.
Parameters
----------
X : {array-like, sparse matrix} of shape (n_samples, n_features)
The training input samples.
y : array-like of shape (n_samples,... | Fit the RFE model and then the underlying estimator on the selected features.
Parameters
----------
X : {array-like, sparse matrix} of shape (n_samples, n_features)
The training input samples.
y : array-like of shape (n_samples,)
The target values.
**fi... | fit | python | scikit-learn/scikit-learn | sklearn/feature_selection/_rfe.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_rfe.py | BSD-3-Clause |
def predict(self, X, **predict_params):
"""Reduce X to the selected features and predict using the estimator.
Parameters
----------
X : array of shape [n_samples, n_features]
The input samples.
**predict_params : dict
Parameters to route to the ``predict... | Reduce X to the selected features and predict using the estimator.
Parameters
----------
X : array of shape [n_samples, n_features]
The input samples.
**predict_params : dict
Parameters to route to the ``predict`` method of the
underlying estimator.
... | predict | python | scikit-learn/scikit-learn | sklearn/feature_selection/_rfe.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_rfe.py | BSD-3-Clause |
def score(self, X, y, **score_params):
"""Reduce X to the selected features and return the score of the estimator.
Parameters
----------
X : array of shape [n_samples, n_features]
The input samples.
y : array of shape [n_samples]
The target values.
... | Reduce X to the selected features and return the score of the estimator.
Parameters
----------
X : array of shape [n_samples, n_features]
The input samples.
y : array of shape [n_samples]
The target values.
**score_params : dict
- If `enable... | score | python | scikit-learn/scikit-learn | sklearn/feature_selection/_rfe.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_rfe.py | BSD-3-Clause |
def fit(self, X, y, *, groups=None, **params):
"""Fit the RFE model and automatically tune the number of selected features.
Parameters
----------
X : {array-like, sparse matrix} of shape (n_samples, n_features)
Training vector, where `n_samples` is the number of samples and
... | Fit the RFE model and automatically tune the number of selected features.
Parameters
----------
X : {array-like, sparse matrix} of shape (n_samples, n_features)
Training vector, where `n_samples` is the number of samples and
`n_features` is the total number of features.
... | fit | python | scikit-learn/scikit-learn | sklearn/feature_selection/_rfe.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_rfe.py | BSD-3-Clause |
def score(self, X, y, **score_params):
"""Score using the `scoring` option on the given test data and labels.
Parameters
----------
X : array-like of shape (n_samples, n_features)
Test samples.
y : array-like of shape (n_samples,)
True labels for X.
... | Score using the `scoring` option on the given test data and labels.
Parameters
----------
X : array-like of shape (n_samples, n_features)
Test samples.
y : array-like of shape (n_samples,)
True labels for X.
**score_params : dict
Parameters ... | score | python | scikit-learn/scikit-learn | sklearn/feature_selection/_rfe.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_rfe.py | BSD-3-Clause |
def fit(self, X, y=None, **params):
"""Learn the features to select from X.
Parameters
----------
X : array-like of shape (n_samples, n_features)
Training vectors, where `n_samples` is the number of samples and
`n_features` is the number of predictors.
y... | Learn the features to select from X.
Parameters
----------
X : array-like of shape (n_samples, n_features)
Training vectors, where `n_samples` is the number of samples and
`n_features` is the number of predictors.
y : array-like of shape (n_samples,), default=No... | fit | python | scikit-learn/scikit-learn | sklearn/feature_selection/_sequential.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_sequential.py | BSD-3-Clause |
def _clean_nans(scores):
"""
Fixes Issue #1240: NaNs can't be properly compared, so change them to the
smallest value of scores's dtype. -inf seems to be unreliable.
"""
# XXX where should this function be called? fit? scoring functions
# themselves?
scores = as_float_array(scores, copy=True... |
Fixes Issue #1240: NaNs can't be properly compared, so change them to the
smallest value of scores's dtype. -inf seems to be unreliable.
| _clean_nans | python | scikit-learn/scikit-learn | sklearn/feature_selection/_univariate_selection.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_univariate_selection.py | BSD-3-Clause |
def f_oneway(*args):
"""Perform a 1-way ANOVA.
The one-way ANOVA tests the null hypothesis that 2 or more groups have
the same population mean. The test is applied to samples from two or
more groups, possibly with differing sizes.
Read more in the :ref:`User Guide <univariate_feature_selection>`.
... | Perform a 1-way ANOVA.
The one-way ANOVA tests the null hypothesis that 2 or more groups have
the same population mean. The test is applied to samples from two or
more groups, possibly with differing sizes.
Read more in the :ref:`User Guide <univariate_feature_selection>`.
Parameters
--------... | f_oneway | python | scikit-learn/scikit-learn | sklearn/feature_selection/_univariate_selection.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_univariate_selection.py | BSD-3-Clause |
def f_classif(X, y):
"""Compute the ANOVA F-value for the provided sample.
Read more in the :ref:`User Guide <univariate_feature_selection>`.
Parameters
----------
X : {array-like, sparse matrix} of shape (n_samples, n_features)
The set of regressors that will be tested sequentially.
... | Compute the ANOVA F-value for the provided sample.
Read more in the :ref:`User Guide <univariate_feature_selection>`.
Parameters
----------
X : {array-like, sparse matrix} of shape (n_samples, n_features)
The set of regressors that will be tested sequentially.
y : array-like of shape (n_s... | f_classif | python | scikit-learn/scikit-learn | sklearn/feature_selection/_univariate_selection.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_univariate_selection.py | BSD-3-Clause |
def _chisquare(f_obs, f_exp):
"""Fast replacement for scipy.stats.chisquare.
Version from https://github.com/scipy/scipy/pull/2525 with additional
optimizations.
"""
f_obs = np.asarray(f_obs, dtype=np.float64)
k = len(f_obs)
# Reuse f_obs for chi-squared statistics
chisq = f_obs
ch... | Fast replacement for scipy.stats.chisquare.
Version from https://github.com/scipy/scipy/pull/2525 with additional
optimizations.
| _chisquare | python | scikit-learn/scikit-learn | sklearn/feature_selection/_univariate_selection.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_univariate_selection.py | BSD-3-Clause |
def chi2(X, y):
"""Compute chi-squared stats between each non-negative feature and class.
This score can be used to select the `n_features` features with the
highest values for the test chi-squared statistic from X, which must
contain only **non-negative integer feature values** such as booleans or fre... | Compute chi-squared stats between each non-negative feature and class.
This score can be used to select the `n_features` features with the
highest values for the test chi-squared statistic from X, which must
contain only **non-negative integer feature values** such as booleans or frequencies
(e.g., ter... | chi2 | python | scikit-learn/scikit-learn | sklearn/feature_selection/_univariate_selection.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_univariate_selection.py | BSD-3-Clause |
def r_regression(X, y, *, center=True, force_finite=True):
"""Compute Pearson's r for each features and the target.
Pearson's r is also known as the Pearson correlation coefficient.
Linear model for testing the individual effect of each of many regressors.
This is a scoring function to be used in a fe... | Compute Pearson's r for each features and the target.
Pearson's r is also known as the Pearson correlation coefficient.
Linear model for testing the individual effect of each of many regressors.
This is a scoring function to be used in a feature selection procedure, not
a free standing feature selecti... | r_regression | python | scikit-learn/scikit-learn | sklearn/feature_selection/_univariate_selection.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_univariate_selection.py | BSD-3-Clause |
def f_regression(X, y, *, center=True, force_finite=True):
"""Univariate linear regression tests returning F-statistic and p-values.
Quick linear model for testing the effect of a single regressor,
sequentially for many regressors.
This is done in 2 steps:
1. The cross correlation between each re... | Univariate linear regression tests returning F-statistic and p-values.
Quick linear model for testing the effect of a single regressor,
sequentially for many regressors.
This is done in 2 steps:
1. The cross correlation between each regressor and the target is computed
using :func:`r_regressio... | f_regression | python | scikit-learn/scikit-learn | sklearn/feature_selection/_univariate_selection.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_univariate_selection.py | BSD-3-Clause |
def fit(self, X, y=None):
"""Run score function on (X, y) and get the appropriate features.
Parameters
----------
X : array-like of shape (n_samples, n_features)
The training input samples.
y : array-like of shape (n_samples,) or None
The target values (... | Run score function on (X, y) and get the appropriate features.
Parameters
----------
X : array-like of shape (n_samples, n_features)
The training input samples.
y : array-like of shape (n_samples,) or None
The target values (class labels in classification, real ... | fit | python | scikit-learn/scikit-learn | sklearn/feature_selection/_univariate_selection.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_univariate_selection.py | BSD-3-Clause |
def fit(self, X, y=None):
"""Learn empirical variances from X.
Parameters
----------
X : {array-like, sparse matrix}, shape (n_samples, n_features)
Data from which to compute variances, where `n_samples` is
the number of samples and `n_features` is the number of ... | Learn empirical variances from X.
Parameters
----------
X : {array-like, sparse matrix}, shape (n_samples, n_features)
Data from which to compute variances, where `n_samples` is
the number of samples and `n_features` is the number of features.
y : any, default=N... | fit | python | scikit-learn/scikit-learn | sklearn/feature_selection/_variance_threshold.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/_variance_threshold.py | BSD-3-Clause |
def test_output_dataframe():
"""Check output dtypes for dataframes is consistent with the input dtypes."""
pd = pytest.importorskip("pandas")
X = pd.DataFrame(
{
"a": pd.Series([1.0, 2.4, 4.5], dtype=np.float32),
"b": pd.Series(["a", "b", "a"], dtype="category"),
... | Check output dtypes for dataframes is consistent with the input dtypes. | test_output_dataframe | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_base.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_base.py | BSD-3-Clause |
def test_r_regression_force_finite(X, y, expected_corr_coef, force_finite):
"""Check the behaviour of `force_finite` for some corner cases with `r_regression`.
Non-regression test for:
https://github.com/scikit-learn/scikit-learn/issues/15672
"""
with warnings.catch_warnings():
warnings.sim... | Check the behaviour of `force_finite` for some corner cases with `r_regression`.
Non-regression test for:
https://github.com/scikit-learn/scikit-learn/issues/15672
| test_r_regression_force_finite | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_feature_select.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_feature_select.py | BSD-3-Clause |
def test_f_regression_corner_case(
X, y, expected_f_statistic, expected_p_values, force_finite
):
"""Check the behaviour of `force_finite` for some corner cases with `f_regression`.
Non-regression test for:
https://github.com/scikit-learn/scikit-learn/issues/15672
"""
with warnings.catch_warnin... | Check the behaviour of `force_finite` for some corner cases with `f_regression`.
Non-regression test for:
https://github.com/scikit-learn/scikit-learn/issues/15672
| test_f_regression_corner_case | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_feature_select.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_feature_select.py | BSD-3-Clause |
def test_dataframe_output_dtypes():
"""Check that the output datafarme dtypes are the same as the input.
Non-regression test for gh-24860.
"""
pd = pytest.importorskip("pandas")
X, y = load_iris(return_X_y=True, as_frame=True)
X = X.astype(
{
"petal length (cm)": np.float32... | Check that the output datafarme dtypes are the same as the input.
Non-regression test for gh-24860.
| test_dataframe_output_dtypes | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_feature_select.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_feature_select.py | BSD-3-Clause |
def test_unsupervised_filter(selector):
"""Check support for unsupervised feature selection for the filter that could
require only `X`.
"""
rng = np.random.RandomState(0)
X = rng.randn(10, 5)
def score_func(X, y=None):
return np.array([1, 1, 1, 1, 0])
selector.set_params(score_func... | Check support for unsupervised feature selection for the filter that could
require only `X`.
| test_unsupervised_filter | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_feature_select.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_feature_select.py | BSD-3-Clause |
def test_inferred_max_features_integer(max_features):
"""Check max_features_ and output shape for integer max_features."""
clf = RandomForestClassifier(n_estimators=5, random_state=0)
transformer = SelectFromModel(
estimator=clf, max_features=max_features, threshold=-np.inf
)
X_trans = trans... | Check max_features_ and output shape for integer max_features. | test_inferred_max_features_integer | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_from_model.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_from_model.py | BSD-3-Clause |
def test_inferred_max_features_callable(max_features):
"""Check max_features_ and output shape for callable max_features."""
clf = RandomForestClassifier(n_estimators=5, random_state=0)
transformer = SelectFromModel(
estimator=clf, max_features=max_features, threshold=-np.inf
)
X_trans = tra... | Check max_features_ and output shape for callable max_features. | test_inferred_max_features_callable | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_from_model.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_from_model.py | BSD-3-Clause |
def test_max_features_callable_data(max_features):
"""Tests that the callable passed to `fit` is called on X."""
clf = RandomForestClassifier(n_estimators=50, random_state=0)
m = Mock(side_effect=max_features)
transformer = SelectFromModel(estimator=clf, max_features=m, threshold=-np.inf)
transforme... | Tests that the callable passed to `fit` is called on X. | test_max_features_callable_data | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_from_model.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_from_model.py | BSD-3-Clause |
def test_prefit_max_features():
"""Check the interaction between `prefit` and `max_features`."""
# case 1: an error should be raised at `transform` if `fit` was not called to
# validate the attributes
estimator = RandomForestClassifier(n_estimators=5, random_state=0)
estimator.fit(data, y)
model... | Check the interaction between `prefit` and `max_features`. | test_prefit_max_features | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_from_model.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_from_model.py | BSD-3-Clause |
def test_get_feature_names_out_elasticnetcv():
"""Check if ElasticNetCV works with a list of floats.
Non-regression test for #30936."""
X, y = make_regression(n_features=5, n_informative=3, random_state=0)
estimator = ElasticNetCV(l1_ratio=[0.25, 0.5, 0.75], random_state=0)
selector = SelectFromMod... | Check if ElasticNetCV works with a list of floats.
Non-regression test for #30936. | test_get_feature_names_out_elasticnetcv | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_from_model.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_from_model.py | BSD-3-Clause |
def test_prefit_get_feature_names_out():
"""Check the interaction between prefit and the feature names."""
clf = RandomForestClassifier(n_estimators=2, random_state=0)
clf.fit(data, y)
model = SelectFromModel(clf, prefit=True, max_features=1)
name = type(model).__name__
err_msg = (
f"Th... | Check the interaction between prefit and the feature names. | test_prefit_get_feature_names_out | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_from_model.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_from_model.py | BSD-3-Clause |
def test_select_from_model_pls(PLSEstimator):
"""Check the behaviour of SelectFromModel with PLS estimators.
Non-regression test for:
https://github.com/scikit-learn/scikit-learn/issues/12410
"""
X, y = make_friedman1(n_samples=50, n_features=10, random_state=0)
estimator = PLSEstimator(n_compo... | Check the behaviour of SelectFromModel with PLS estimators.
Non-regression test for:
https://github.com/scikit-learn/scikit-learn/issues/12410
| test_select_from_model_pls | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_from_model.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_from_model.py | BSD-3-Clause |
def test_estimator_does_not_support_feature_names():
"""SelectFromModel works with estimators that do not support feature_names_in_.
Non-regression test for #21949.
"""
pytest.importorskip("pandas")
X, y = datasets.load_iris(as_frame=True, return_X_y=True)
all_feature_names = set(X.columns)
... | SelectFromModel works with estimators that do not support feature_names_in_.
Non-regression test for #21949.
| test_estimator_does_not_support_feature_names | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_from_model.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_from_model.py | BSD-3-Clause |
def test_from_model_estimator_attribute_error():
"""Check that we raise the proper AttributeError when the estimator
does not implement the `partial_fit` method, which is decorated with
`available_if`.
Non-regression test for:
https://github.com/scikit-learn/scikit-learn/issues/28108
"""
# ... | Check that we raise the proper AttributeError when the estimator
does not implement the `partial_fit` method, which is decorated with
`available_if`.
Non-regression test for:
https://github.com/scikit-learn/scikit-learn/issues/28108
| test_from_model_estimator_attribute_error | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_from_model.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_from_model.py | BSD-3-Clause |
def test_mutual_information_symmetry_classif_regression(correlated, global_random_seed):
"""Check that `mutual_info_classif` and `mutual_info_regression` are
symmetric by switching the target `y` as `feature` in `X` and vice
versa.
Non-regression test for:
https://github.com/scikit-learn/scikit-lea... | Check that `mutual_info_classif` and `mutual_info_regression` are
symmetric by switching the target `y` as `feature` in `X` and vice
versa.
Non-regression test for:
https://github.com/scikit-learn/scikit-learn/issues/23720
| test_mutual_information_symmetry_classif_regression | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_mutual_info.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_mutual_info.py | BSD-3-Clause |
def test_mutual_info_regression_X_int_dtype(global_random_seed):
"""Check that results agree when X is integer dtype and float dtype.
Non-regression test for Issue #26696.
"""
rng = np.random.RandomState(global_random_seed)
X = rng.randint(100, size=(100, 10))
X_float = X.astype(np.float64, cop... | Check that results agree when X is integer dtype and float dtype.
Non-regression test for Issue #26696.
| test_mutual_info_regression_X_int_dtype | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_mutual_info.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_mutual_info.py | BSD-3-Clause |
def test_mutual_info_n_jobs(global_random_seed, mutual_info_func, data_generator):
"""Check that results are consistent with different `n_jobs`."""
X, y = data_generator(random_state=global_random_seed)
single_job = mutual_info_func(X, y, random_state=global_random_seed, n_jobs=1)
multi_job = mutual_inf... | Check that results are consistent with different `n_jobs`. | test_mutual_info_n_jobs | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_mutual_info.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_mutual_info.py | BSD-3-Clause |
def test_pipeline_with_nans(ClsRFE):
"""Check that RFE works with pipeline that accept nans.
Non-regression test for gh-21743.
"""
X, y = load_iris(return_X_y=True)
X[0, 0] = np.nan
pipe = make_pipeline(
SimpleImputer(),
StandardScaler(),
LogisticRegression(),
)
... | Check that RFE works with pipeline that accept nans.
Non-regression test for gh-21743.
| test_pipeline_with_nans | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_rfe.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_rfe.py | BSD-3-Clause |
def test_rfe_pls(ClsRFE, PLSEstimator):
"""Check the behaviour of RFE with PLS estimators.
Non-regression test for:
https://github.com/scikit-learn/scikit-learn/issues/12410
"""
X, y = make_friedman1(n_samples=50, n_features=10, random_state=0)
estimator = PLSEstimator(n_components=1)
selec... | Check the behaviour of RFE with PLS estimators.
Non-regression test for:
https://github.com/scikit-learn/scikit-learn/issues/12410
| test_rfe_pls | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_rfe.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_rfe.py | BSD-3-Clause |
def test_rfe_estimator_attribute_error():
"""Check that we raise the proper AttributeError when the estimator
does not implement the `decision_function` method, which is decorated with
`available_if`.
Non-regression test for:
https://github.com/scikit-learn/scikit-learn/issues/28108
"""
iri... | Check that we raise the proper AttributeError when the estimator
does not implement the `decision_function` method, which is decorated with
`available_if`.
Non-regression test for:
https://github.com/scikit-learn/scikit-learn/issues/28108
| test_rfe_estimator_attribute_error | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_rfe.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_rfe.py | BSD-3-Clause |
def test_rfe_n_features_to_select_warning(ClsRFE, param):
"""Check if the correct warning is raised when trying to initialize a RFE
object with a n_features_to_select attribute larger than the number of
features present in the X variable that is passed to the fit method
"""
X, y = make_classificatio... | Check if the correct warning is raised when trying to initialize a RFE
object with a n_features_to_select attribute larger than the number of
features present in the X variable that is passed to the fit method
| test_rfe_n_features_to_select_warning | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_rfe.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_rfe.py | BSD-3-Clause |
def test_rfe_with_sample_weight():
"""Test that `RFE` works correctly with sample weights."""
X, y = make_classification(random_state=0)
n_samples = X.shape[0]
# Assign the first half of the samples with twice the weight
sample_weight = np.ones_like(y)
sample_weight[: n_samples // 2] = 2
#... | Test that `RFE` works correctly with sample weights. | test_rfe_with_sample_weight | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_rfe.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_rfe.py | BSD-3-Clause |
def test_results_per_cv_in_rfecv(global_random_seed):
"""
Test that the results of RFECV are consistent across the different folds
in terms of length of the arrays.
"""
X, y = make_classification(random_state=global_random_seed)
clf = LogisticRegression()
rfecv = RFECV(
estimator=cl... |
Test that the results of RFECV are consistent across the different folds
in terms of length of the arrays.
| test_results_per_cv_in_rfecv | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_rfe.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_rfe.py | BSD-3-Clause |
def test_n_features_to_select_auto(direction):
"""Check the behaviour of `n_features_to_select="auto"` with different
values for the parameter `tol`.
"""
n_features = 10
tol = 1e-3
X, y = make_regression(n_features=n_features, random_state=0)
sfs = SequentialFeatureSelector(
LinearR... | Check the behaviour of `n_features_to_select="auto"` with different
values for the parameter `tol`.
| test_n_features_to_select_auto | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_sequential.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_sequential.py | BSD-3-Clause |
def test_n_features_to_select_stopping_criterion(direction):
"""Check the behaviour stopping criterion for feature selection
depending on the values of `n_features_to_select` and `tol`.
When `direction` is `'forward'`, select a new features at random
among those not currently selected in selector.suppo... | Check the behaviour stopping criterion for feature selection
depending on the values of `n_features_to_select` and `tol`.
When `direction` is `'forward'`, select a new features at random
among those not currently selected in selector.support_,
build a new version of the data that includes all the featu... | test_n_features_to_select_stopping_criterion | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_sequential.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_sequential.py | BSD-3-Clause |
def test_forward_neg_tol_error():
"""Check that we raise an error when tol<0 and direction='forward'"""
X, y = make_regression(n_features=10, random_state=0)
sfs = SequentialFeatureSelector(
LinearRegression(),
n_features_to_select="auto",
direction="forward",
tol=-1e-3,
... | Check that we raise an error when tol<0 and direction='forward' | test_forward_neg_tol_error | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_sequential.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_sequential.py | BSD-3-Clause |
def test_backward_neg_tol():
"""Check that SequentialFeatureSelector works negative tol
non-regression test for #25525
"""
X, y = make_regression(n_features=10, random_state=0)
lr = LinearRegression()
initial_score = lr.fit(X, y).score(X, y)
sfs = SequentialFeatureSelector(
lr,
... | Check that SequentialFeatureSelector works negative tol
non-regression test for #25525
| test_backward_neg_tol | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_sequential.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_sequential.py | BSD-3-Clause |
def test_cv_generator_support():
"""Check that no exception raised when cv is generator
non-regression test for #25957
"""
X, y = make_classification(random_state=0)
groups = np.zeros_like(y, dtype=int)
groups[y.size // 2 :] = 1
cv = LeaveOneGroupOut()
splits = cv.split(X, y, groups=g... | Check that no exception raised when cv is generator
non-regression test for #25957
| test_cv_generator_support | python | scikit-learn/scikit-learn | sklearn/feature_selection/tests/test_sequential.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/feature_selection/tests/test_sequential.py | BSD-3-Clause |
def _estimator_has(attr):
"""Check that final_estimator has `attr`.
Used together with `available_if`.
"""
def check(self):
# raise original `AttributeError` if `attr` does not exist
getattr(self.estimator, attr)
return True
return check | Check that final_estimator has `attr`.
Used together with `available_if`.
| _estimator_has | python | scikit-learn/scikit-learn | sklearn/frozen/_frozen.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/frozen/_frozen.py | BSD-3-Clause |
def set_params(self, **kwargs):
"""Set the parameters of this estimator.
The only valid key here is `estimator`. You cannot set the parameters of the
inner estimator.
Parameters
----------
**kwargs : dict
Estimator parameters.
Returns
------... | Set the parameters of this estimator.
The only valid key here is `estimator`. You cannot set the parameters of the
inner estimator.
Parameters
----------
**kwargs : dict
Estimator parameters.
Returns
-------
self : FrozenEstimator
... | set_params | python | scikit-learn/scikit-learn | sklearn/frozen/_frozen.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/frozen/_frozen.py | BSD-3-Clause |
def test_frozen_methods(estimator, dataset, request, method):
"""Test that frozen.fit doesn't do anything, and that all other methods are
exposed by the frozen estimator and return the same values as the estimator.
"""
X, y = request.getfixturevalue(dataset)
set_random_state(estimator)
estimator... | Test that frozen.fit doesn't do anything, and that all other methods are
exposed by the frozen estimator and return the same values as the estimator.
| test_frozen_methods | python | scikit-learn/scikit-learn | sklearn/frozen/tests/test_frozen.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/frozen/tests/test_frozen.py | BSD-3-Clause |
def test_frozen_metadata_routing(regression_dataset):
"""Test that metadata routing works with frozen estimators."""
class ConsumesMetadata(BaseEstimator):
def __init__(self, on_fit=None, on_predict=None):
self.on_fit = on_fit
self.on_predict = on_predict
def fit(self, ... | Test that metadata routing works with frozen estimators. | test_frozen_metadata_routing | python | scikit-learn/scikit-learn | sklearn/frozen/tests/test_frozen.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/frozen/tests/test_frozen.py | BSD-3-Clause |
def test_composite_fit(classification_dataset):
"""Test that calling fit_transform and fit_predict doesn't call fit."""
class Estimator(BaseEstimator):
def fit(self, X, y):
try:
self._fit_counter += 1
except AttributeError:
self._fit_counter = 1
... | Test that calling fit_transform and fit_predict doesn't call fit. | test_composite_fit | python | scikit-learn/scikit-learn | sklearn/frozen/tests/test_frozen.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/frozen/tests/test_frozen.py | BSD-3-Clause |
def test_clone_frozen(regression_dataset):
"""Test that cloning a frozen estimator keeps the frozen state."""
X, y = regression_dataset
estimator = LinearRegression().fit(X, y)
frozen = FrozenEstimator(estimator)
cloned = clone(frozen)
assert cloned.estimator is estimator | Test that cloning a frozen estimator keeps the frozen state. | test_clone_frozen | python | scikit-learn/scikit-learn | sklearn/frozen/tests/test_frozen.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/frozen/tests/test_frozen.py | BSD-3-Clause |
def test_check_is_fitted(regression_dataset):
"""Test that check_is_fitted works on frozen estimators."""
X, y = regression_dataset
estimator = LinearRegression()
frozen = FrozenEstimator(estimator)
with pytest.raises(NotFittedError):
check_is_fitted(frozen)
estimator = LinearRegressio... | Test that check_is_fitted works on frozen estimators. | test_check_is_fitted | python | scikit-learn/scikit-learn | sklearn/frozen/tests/test_frozen.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/frozen/tests/test_frozen.py | BSD-3-Clause |
def test_frozen_tags():
"""Test that frozen estimators have the same tags as the original estimator
except for the skip_test tag."""
class Estimator(BaseEstimator):
def __sklearn_tags__(self):
tags = super().__sklearn_tags__()
tags.input_tags.categorical = True
r... | Test that frozen estimators have the same tags as the original estimator
except for the skip_test tag. | test_frozen_tags | python | scikit-learn/scikit-learn | sklearn/frozen/tests/test_frozen.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/frozen/tests/test_frozen.py | BSD-3-Clause |
def test_frozen_params():
"""Test that FrozenEstimator only exposes the estimator parameter."""
est = LogisticRegression()
frozen = FrozenEstimator(est)
with pytest.raises(ValueError, match="You cannot set parameters of the inner"):
frozen.set_params(estimator__C=1)
assert frozen.get_param... | Test that FrozenEstimator only exposes the estimator parameter. | test_frozen_params | python | scikit-learn/scikit-learn | sklearn/frozen/tests/test_frozen.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/frozen/tests/test_frozen.py | BSD-3-Clause |
def get_params(self, deep=True):
"""Get parameters of this kernel.
Parameters
----------
deep : bool, default=True
If True, will return the parameters for this estimator and
contained subobjects that are estimators.
Returns
-------
params... | Get parameters of this kernel.
Parameters
----------
deep : bool, default=True
If True, will return the parameters for this estimator and
contained subobjects that are estimators.
Returns
-------
params : dict
Parameter names mapped t... | get_params | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def set_params(self, **params):
"""Set the parameters of this kernel.
The method works on simple kernels as well as on nested kernels.
The latter have parameters of the form ``<component>__<parameter>``
so that it's possible to update each component of a nested object.
Returns
... | Set the parameters of this kernel.
The method works on simple kernels as well as on nested kernels.
The latter have parameters of the form ``<component>__<parameter>``
so that it's possible to update each component of a nested object.
Returns
-------
self
| set_params | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def clone_with_theta(self, theta):
"""Returns a clone of self with given hyperparameters theta.
Parameters
----------
theta : ndarray of shape (n_dims,)
The hyperparameters
"""
cloned = clone(self)
cloned.theta = theta
return cloned | Returns a clone of self with given hyperparameters theta.
Parameters
----------
theta : ndarray of shape (n_dims,)
The hyperparameters
| clone_with_theta | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def hyperparameters(self):
"""Returns a list of all hyperparameter specifications."""
r = [
getattr(self, attr)
for attr in dir(self)
if attr.startswith("hyperparameter_")
]
return r | Returns a list of all hyperparameter specifications. | hyperparameters | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def theta(self):
"""Returns the (flattened, log-transformed) non-fixed hyperparameters.
Note that theta are typically the log-transformed values of the
kernel's hyperparameters as this representation of the search space
is more amenable for hyperparameter search, as hyperparameters like... | Returns the (flattened, log-transformed) non-fixed hyperparameters.
Note that theta are typically the log-transformed values of the
kernel's hyperparameters as this representation of the search space
is more amenable for hyperparameter search, as hyperparameters like
length-scales natur... | theta | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def theta(self, theta):
"""Sets the (flattened, log-transformed) non-fixed hyperparameters.
Parameters
----------
theta : ndarray of shape (n_dims,)
The non-fixed, log-transformed hyperparameters of the kernel
"""
params = self.get_params()
i = 0
... | Sets the (flattened, log-transformed) non-fixed hyperparameters.
Parameters
----------
theta : ndarray of shape (n_dims,)
The non-fixed, log-transformed hyperparameters of the kernel
| theta | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def bounds(self):
"""Returns the log-transformed bounds on the theta.
Returns
-------
bounds : ndarray of shape (n_dims, 2)
The log-transformed bounds on the kernel's hyperparameters theta
"""
bounds = [
hyperparameter.bounds
for hyper... | Returns the log-transformed bounds on the theta.
Returns
-------
bounds : ndarray of shape (n_dims, 2)
The log-transformed bounds on the kernel's hyperparameters theta
| bounds | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def diag(self, X):
"""Returns the diagonal of the kernel k(X, X).
The result of this method is identical to np.diag(self(X)); however,
it can be evaluated more efficiently since only the diagonal is
evaluated.
Parameters
----------
X : array-like of shape (n_sam... | Returns the diagonal of the kernel k(X, X).
The result of this method is identical to np.diag(self(X)); however,
it can be evaluated more efficiently since only the diagonal is
evaluated.
Parameters
----------
X : array-like of shape (n_samples,)
Left argume... | diag | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def _check_bounds_params(self):
"""Called after fitting to warn if bounds may have been too tight."""
list_close = np.isclose(self.bounds, np.atleast_2d(self.theta).T)
idx = 0
for hyp in self.hyperparameters:
if hyp.fixed:
continue
for dim in range... | Called after fitting to warn if bounds may have been too tight. | _check_bounds_params | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def theta(self, theta):
"""Sets the (flattened, log-transformed) non-fixed hyperparameters.
Parameters
----------
theta : array of shape (n_dims,)
The non-fixed, log-transformed hyperparameters of the kernel
"""
k_dims = self.k1.n_dims
for i, kernel i... | Sets the (flattened, log-transformed) non-fixed hyperparameters.
Parameters
----------
theta : array of shape (n_dims,)
The non-fixed, log-transformed hyperparameters of the kernel
| theta | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def __call__(self, X, Y=None, eval_gradient=False):
"""Return the kernel k(X, Y) and optionally its gradient.
Note that this compound kernel returns the results of all simple kernel
stacked along an additional axis.
Parameters
----------
X : array-like of shape (n_sampl... | Return the kernel k(X, Y) and optionally its gradient.
Note that this compound kernel returns the results of all simple kernel
stacked along an additional axis.
Parameters
----------
X : array-like of shape (n_samples_X, n_features) or list of object, default=None
... | __call__ | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def __call__(self, X, Y=None, eval_gradient=False):
"""Return the kernel k(X, Y) and optionally its gradient.
Parameters
----------
X : array-like of shape (n_samples_X, n_features) or list of object
Left argument of the returned kernel k(X, Y)
Y : array-like of sha... | Return the kernel k(X, Y) and optionally its gradient.
Parameters
----------
X : array-like of shape (n_samples_X, n_features) or list of object
Left argument of the returned kernel k(X, Y)
Y : array-like of shape (n_samples_X, n_features) or list of object, ... | __call__ | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def __call__(self, X, Y=None, eval_gradient=False):
"""Return the kernel k(X, Y) and optionally its gradient.
Parameters
----------
X : array-like of shape (n_samples_X, n_features) or list of object
Left argument of the returned kernel k(X, Y)
Y : array-like of sha... | Return the kernel k(X, Y) and optionally its gradient.
Parameters
----------
X : array-like of shape (n_samples_X, n_features) or list of object
Left argument of the returned kernel k(X, Y)
Y : array-like of shape (n_samples_Y, n_features) or list of object, defa... | __call__ | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def __call__(self, X, Y=None, eval_gradient=False):
"""Return the kernel k(X, Y) and optionally its gradient.
Parameters
----------
X : array-like of shape (n_samples_X, n_features) or list of object
Left argument of the returned kernel k(X, Y)
Y : array-like of sha... | Return the kernel k(X, Y) and optionally its gradient.
Parameters
----------
X : array-like of shape (n_samples_X, n_features) or list of object
Left argument of the returned kernel k(X, Y)
Y : array-like of shape (n_samples_Y, n_features) or list of object, defa... | __call__ | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def __call__(self, X, Y=None, eval_gradient=False):
"""Return the kernel k(X, Y) and optionally its gradient.
Parameters
----------
X : array-like of shape (n_samples_X, n_features) or list of object
Left argument of the returned kernel k(X, Y)
Y : array-like of sha... | Return the kernel k(X, Y) and optionally its gradient.
Parameters
----------
X : array-like of shape (n_samples_X, n_features) or list of object
Left argument of the returned kernel k(X, Y)
Y : array-like of shape (n_samples_X, n_features) or list of object, def... | __call__ | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def diag(self, X):
"""Returns the diagonal of the kernel k(X, X).
The result of this method is identical to np.diag(self(X)); however,
it can be evaluated more efficiently since only the diagonal is
evaluated.
Parameters
----------
X : array-like of shape (n_sam... | Returns the diagonal of the kernel k(X, X).
The result of this method is identical to np.diag(self(X)); however,
it can be evaluated more efficiently since only the diagonal is
evaluated.
Parameters
----------
X : array-like of shape (n_samples_X, n_features) or list of... | diag | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def __call__(self, X, Y=None, eval_gradient=False):
"""Return the kernel k(X, Y) and optionally its gradient.
Parameters
----------
X : array-like of shape (n_samples_X, n_features) or list of object
Left argument of the returned kernel k(X, Y)
Y : array-like of sha... | Return the kernel k(X, Y) and optionally its gradient.
Parameters
----------
X : array-like of shape (n_samples_X, n_features) or list of object
Left argument of the returned kernel k(X, Y)
Y : array-like of shape (n_samples_X, n_features) or list of object, defa... | __call__ | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def __call__(self, X, Y=None, eval_gradient=False):
"""Return the kernel k(X, Y) and optionally its gradient.
Parameters
----------
X : ndarray of shape (n_samples_X, n_features)
Left argument of the returned kernel k(X, Y)
Y : ndarray of shape (n_samples_Y, n_featu... | Return the kernel k(X, Y) and optionally its gradient.
Parameters
----------
X : ndarray of shape (n_samples_X, n_features)
Left argument of the returned kernel k(X, Y)
Y : ndarray of shape (n_samples_Y, n_features), default=None
Right argument of the returned k... | __call__ | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def __call__(self, X, Y=None, eval_gradient=False):
"""Return the kernel k(X, Y) and optionally its gradient.
Parameters
----------
X : ndarray of shape (n_samples_X, n_features)
Left argument of the returned kernel k(X, Y)
Y : ndarray of shape (n_samples_Y, n_featu... | Return the kernel k(X, Y) and optionally its gradient.
Parameters
----------
X : ndarray of shape (n_samples_X, n_features)
Left argument of the returned kernel k(X, Y)
Y : ndarray of shape (n_samples_Y, n_features), default=None
Right argument of the returned k... | __call__ | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def __call__(self, X, Y=None, eval_gradient=False):
"""Return the kernel k(X, Y) and optionally its gradient.
Parameters
----------
X : ndarray of shape (n_samples_X, n_features)
Left argument of the returned kernel k(X, Y)
Y : ndarray of shape (n_samples_Y, n_featu... | Return the kernel k(X, Y) and optionally its gradient.
Parameters
----------
X : ndarray of shape (n_samples_X, n_features)
Left argument of the returned kernel k(X, Y)
Y : ndarray of shape (n_samples_Y, n_features), default=None
Right argument of the returned k... | __call__ | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def diag(self, X):
"""Returns the diagonal of the kernel k(X, X).
The result of this method is identical to np.diag(self(X)); however,
it can be evaluated more efficiently since only the diagonal is
evaluated.
Parameters
----------
X : ndarray of shape (n_sample... | Returns the diagonal of the kernel k(X, X).
The result of this method is identical to np.diag(self(X)); however,
it can be evaluated more efficiently since only the diagonal is
evaluated.
Parameters
----------
X : ndarray of shape (n_samples_X, n_features)
L... | diag | python | scikit-learn/scikit-learn | sklearn/gaussian_process/kernels.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/kernels.py | BSD-3-Clause |
def fit(self, X, y):
"""Fit Gaussian process classification model.
Parameters
----------
X : array-like of shape (n_samples, n_features) or list of object
Feature vectors or other representations of training data.
y : array-like of shape (n_samples,)
Tar... | Fit Gaussian process classification model.
Parameters
----------
X : array-like of shape (n_samples, n_features) or list of object
Feature vectors or other representations of training data.
y : array-like of shape (n_samples,)
Target values, must be binary.
... | fit | python | scikit-learn/scikit-learn | sklearn/gaussian_process/_gpc.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/_gpc.py | BSD-3-Clause |
def predict(self, X):
"""Perform classification on an array of test vectors X.
Parameters
----------
X : array-like of shape (n_samples, n_features) or list of object
Query points where the GP is evaluated for classification.
Returns
-------
C : ndar... | Perform classification on an array of test vectors X.
Parameters
----------
X : array-like of shape (n_samples, n_features) or list of object
Query points where the GP is evaluated for classification.
Returns
-------
C : ndarray of shape (n_samples,)
... | predict | python | scikit-learn/scikit-learn | sklearn/gaussian_process/_gpc.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/_gpc.py | BSD-3-Clause |
def predict_proba(self, X):
"""Return probability estimates for the test vector X.
Parameters
----------
X : array-like of shape (n_samples, n_features) or list of object
Query points where the GP is evaluated for classification.
Returns
-------
C : ... | Return probability estimates for the test vector X.
Parameters
----------
X : array-like of shape (n_samples, n_features) or list of object
Query points where the GP is evaluated for classification.
Returns
-------
C : array-like of shape (n_samples, n_class... | predict_proba | python | scikit-learn/scikit-learn | sklearn/gaussian_process/_gpc.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/_gpc.py | BSD-3-Clause |
def log_marginal_likelihood(
self, theta=None, eval_gradient=False, clone_kernel=True
):
"""Returns log-marginal likelihood of theta for training data.
Parameters
----------
theta : array-like of shape (n_kernel_params,), default=None
Kernel hyperparameters for w... | Returns log-marginal likelihood of theta for training data.
Parameters
----------
theta : array-like of shape (n_kernel_params,), default=None
Kernel hyperparameters for which the log-marginal likelihood is
evaluated. If None, the precomputed log_marginal_likelihood
... | log_marginal_likelihood | python | scikit-learn/scikit-learn | sklearn/gaussian_process/_gpc.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/_gpc.py | BSD-3-Clause |
def latent_mean_and_variance(self, X):
"""Compute the mean and variance of the latent function values.
Based on algorithm 3.2 of [RW2006]_, this function returns the latent
mean (Line 4) and variance (Line 6) of the Gaussian process
classification model.
Note that this function... | Compute the mean and variance of the latent function values.
Based on algorithm 3.2 of [RW2006]_, this function returns the latent
mean (Line 4) and variance (Line 6) of the Gaussian process
classification model.
Note that this function is only supported for binary classification.
... | latent_mean_and_variance | python | scikit-learn/scikit-learn | sklearn/gaussian_process/_gpc.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/_gpc.py | BSD-3-Clause |
def _posterior_mode(self, K, return_temporaries=False):
"""Mode-finding for binary Laplace GPC and fixed kernel.
This approximates the posterior of the latent function values for given
inputs and target observations with a Gaussian approximation and uses
Newton's iteration to find the m... | Mode-finding for binary Laplace GPC and fixed kernel.
This approximates the posterior of the latent function values for given
inputs and target observations with a Gaussian approximation and uses
Newton's iteration to find the mode of this approximation.
| _posterior_mode | python | scikit-learn/scikit-learn | sklearn/gaussian_process/_gpc.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/_gpc.py | BSD-3-Clause |
def fit(self, X, y):
"""Fit Gaussian process classification model.
Parameters
----------
X : array-like of shape (n_samples, n_features) or list of object
Feature vectors or other representations of training data.
y : array-like of shape (n_samples,)
Tar... | Fit Gaussian process classification model.
Parameters
----------
X : array-like of shape (n_samples, n_features) or list of object
Feature vectors or other representations of training data.
y : array-like of shape (n_samples,)
Target values, must be binary.
... | fit | python | scikit-learn/scikit-learn | sklearn/gaussian_process/_gpc.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/_gpc.py | BSD-3-Clause |
def predict(self, X):
"""Perform classification on an array of test vectors X.
Parameters
----------
X : array-like of shape (n_samples, n_features) or list of object
Query points where the GP is evaluated for classification.
Returns
-------
C : ndar... | Perform classification on an array of test vectors X.
Parameters
----------
X : array-like of shape (n_samples, n_features) or list of object
Query points where the GP is evaluated for classification.
Returns
-------
C : ndarray of shape (n_samples,)
... | predict | python | scikit-learn/scikit-learn | sklearn/gaussian_process/_gpc.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/_gpc.py | BSD-3-Clause |
def predict_proba(self, X):
"""Return probability estimates for the test vector X.
Parameters
----------
X : array-like of shape (n_samples, n_features) or list of object
Query points where the GP is evaluated for classification.
Returns
-------
C : ... | Return probability estimates for the test vector X.
Parameters
----------
X : array-like of shape (n_samples, n_features) or list of object
Query points where the GP is evaluated for classification.
Returns
-------
C : array-like of shape (n_samples, n_class... | predict_proba | python | scikit-learn/scikit-learn | sklearn/gaussian_process/_gpc.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/_gpc.py | BSD-3-Clause |
def kernel_(self):
"""Return the kernel of the base estimator."""
if self.n_classes_ == 2:
return self.base_estimator_.kernel_
else:
return CompoundKernel(
[estimator.kernel_ for estimator in self.base_estimator_.estimators_]
) | Return the kernel of the base estimator. | kernel_ | python | scikit-learn/scikit-learn | sklearn/gaussian_process/_gpc.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/_gpc.py | BSD-3-Clause |
def log_marginal_likelihood(
self, theta=None, eval_gradient=False, clone_kernel=True
):
"""Return log-marginal likelihood of theta for training data.
In the case of multi-class classification, the mean log-marginal
likelihood of the one-versus-rest classifiers are returned.
... | Return log-marginal likelihood of theta for training data.
In the case of multi-class classification, the mean log-marginal
likelihood of the one-versus-rest classifiers are returned.
Parameters
----------
theta : array-like of shape (n_kernel_params,), default=None
... | log_marginal_likelihood | python | scikit-learn/scikit-learn | sklearn/gaussian_process/_gpc.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/_gpc.py | BSD-3-Clause |
def latent_mean_and_variance(self, X):
"""Compute the mean and variance of the latent function.
Based on algorithm 3.2 of [RW2006]_, this function returns the latent
mean (Line 4) and variance (Line 6) of the Gaussian process
classification model.
Note that this function is onl... | Compute the mean and variance of the latent function.
Based on algorithm 3.2 of [RW2006]_, this function returns the latent
mean (Line 4) and variance (Line 6) of the Gaussian process
classification model.
Note that this function is only supported for binary classification.
..... | latent_mean_and_variance | python | scikit-learn/scikit-learn | sklearn/gaussian_process/_gpc.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/_gpc.py | BSD-3-Clause |
def fit(self, X, y):
"""Fit Gaussian process regression model.
Parameters
----------
X : array-like of shape (n_samples, n_features) or list of object
Feature vectors or other representations of training data.
y : array-like of shape (n_samples,) or (n_samples, n_ta... | Fit Gaussian process regression model.
Parameters
----------
X : array-like of shape (n_samples, n_features) or list of object
Feature vectors or other representations of training data.
y : array-like of shape (n_samples,) or (n_samples, n_targets)
Target values... | fit | python | scikit-learn/scikit-learn | sklearn/gaussian_process/_gpr.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/_gpr.py | BSD-3-Clause |
def predict(self, X, return_std=False, return_cov=False):
"""Predict using the Gaussian process regression model.
We can also predict based on an unfitted model by using the GP prior.
In addition to the mean of the predictive distribution, optionally also
returns its standard deviation ... | Predict using the Gaussian process regression model.
We can also predict based on an unfitted model by using the GP prior.
In addition to the mean of the predictive distribution, optionally also
returns its standard deviation (`return_std=True`) or covariance
(`return_cov=True`). Note t... | predict | python | scikit-learn/scikit-learn | sklearn/gaussian_process/_gpr.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/_gpr.py | BSD-3-Clause |
def sample_y(self, X, n_samples=1, random_state=0):
"""Draw samples from Gaussian process and evaluate at X.
Parameters
----------
X : array-like of shape (n_samples_X, n_features) or list of object
Query points where the GP is evaluated.
n_samples : int, default=1
... | Draw samples from Gaussian process and evaluate at X.
Parameters
----------
X : array-like of shape (n_samples_X, n_features) or list of object
Query points where the GP is evaluated.
n_samples : int, default=1
Number of samples drawn from the Gaussian process p... | sample_y | python | scikit-learn/scikit-learn | sklearn/gaussian_process/_gpr.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/_gpr.py | BSD-3-Clause |
def log_marginal_likelihood(
self, theta=None, eval_gradient=False, clone_kernel=True
):
"""Return log-marginal likelihood of theta for training data.
Parameters
----------
theta : array-like of shape (n_kernel_params,) default=None
Kernel hyperparameters for whi... | Return log-marginal likelihood of theta for training data.
Parameters
----------
theta : array-like of shape (n_kernel_params,) default=None
Kernel hyperparameters for which the log-marginal likelihood is
evaluated. If None, the precomputed log_marginal_likelihood
... | log_marginal_likelihood | python | scikit-learn/scikit-learn | sklearn/gaussian_process/_gpr.py | https://github.com/scikit-learn/scikit-learn/blob/master/sklearn/gaussian_process/_gpr.py | BSD-3-Clause |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.