Source code for tune_sklearn.tune_gridsearch

"""Class for doing grid search over lists of hyperparameters
    -- Anthony Yu and Michael Chau
import warnings

from sklearn.model_selection import ParameterGrid
from sklearn.base import clone

from ray import tune
from ray.tune.stopper import CombinedStopper

from tune_sklearn.list_searcher import ListSearcher
from tune_sklearn.utils import (_check_param_grid_tune_grid_search,
                                check_is_pipeline, check_error_warm_start,
                                is_tune_grid_search, MaximumIterationStopper,
from tune_sklearn.tune_basesearch import TuneBaseSearchCV
from tune_sklearn._trainable import _Trainable
from tune_sklearn._trainable import _PipelineTrainable

[docs]class TuneGridSearchCV(TuneBaseSearchCV): """Exhaustive search over specified parameter values for an estimator. Important members are fit, predict. GridSearchCV implements a "fit" and a "score" method. It also implements "predict", "predict_proba", "decision_function", "transform" and "inverse_transform" if they are implemented in the estimator used. The parameters of the estimator used to apply these methods are optimized by cross-validated grid-search over a parameter grid. Args: estimator (`estimator`): Object that implements the scikit-learn estimator interface. Either estimator needs to provide a ``score`` function, or ``scoring`` must be passed. param_grid (`dict` or `list` of `dict`): Dictionary with parameters names (string) as keys and lists or tune.grid_search outputs of parameter settings to try as values, or a list of such dictionaries, in which case the grids spanned by each dictionary in the list are explored. This enables searching over any sequence of parameter settings. early_stopping (bool, str or :class:`TrialScheduler`, optional): Option to stop fitting to a hyperparameter configuration if it performs poorly. Possible inputs are: - If True, defaults to ASHAScheduler. - A string corresponding to the name of a Tune Trial Scheduler (i.e., "ASHAScheduler"). To specify parameters of the scheduler, pass in a scheduler object instead of a string. - Scheduler for executing fit with early stopping. Only a subset of schedulers are currently supported. The scheduler will only be used if the estimator supports partial fitting - If None or False, early stopping will not be used. scoring (str, list/tuple, dict, or None): A single string or a callable to evaluate the predictions on the test set. See #scoring-parameter for all options. For evaluating multiple metrics, either give a list/tuple of (unique) strings or a dict with names as keys and callables as values. If None, the estimator's score method is used. Defaults to None. n_jobs (int): Number of jobs to run in parallel. None or -1 means using all processors. Defaults to None. If set to 1, jobs will be run using Ray's 'local mode'. This can lead to significant speedups if the model takes < 10 seconds to fit due to removing inter-process communication overheads. cv (int, `cross-validation generator` or `iterable`): Determines the cross-validation splitting strategy. Possible inputs for cv are: - None, to use the default 5-fold cross validation, - integer, to specify the number of folds in a `(Stratified)KFold`, - An iterable yielding (train, test) splits as arrays of indices. For integer/None inputs, if the estimator is a classifier and ``y`` is either binary or multiclass, :class:`StratifiedKFold` is used. In all other cases, :class:`KFold` is used. Defaults to None. refit (bool or str): Refit an estimator using the best found parameters on the whole dataset. For multiple metric evaluation, this needs to be a string denoting the scorer that would be used to find the best parameters for refitting the estimator at the end. The refitted estimator is made available at the ``best_estimator_`` attribute and permits using ``predict`` directly on this ``GridSearchCV`` instance. Also for multiple metric evaluation, the attributes ``best_index_``, ``best_score_`` and ``best_params_`` will only be available if ``refit`` is set and all of them will be determined w.r.t this specific scorer. If refit not needed, set to False. See ``scoring`` parameter to know more about multiple metric evaluation. Defaults to True. verbose (int): Controls the verbosity: 0 = silent, 1 = only status updates, 2 = status and trial results. Defaults to 0. error_score ('raise' or int or float): Value to assign to the score if an error occurs in estimator fitting. If set to 'raise', the error is raised. If a numeric value is given, FitFailedWarning is raised. This parameter does not affect the refit step, which will always raise the error. Defaults to np.nan. return_train_score (bool): If ``False``, the ``cv_results_`` attribute will not include training scores. Defaults to False. Computing training scores is used to get insights on how different parameter settings impact the overfitting/underfitting trade-off. However computing the scores on the training set can be computationally expensive and is not strictly required to select the parameters that yield the best generalization performance. local_dir (str): A string that defines where checkpoints will be stored. Defaults to "~/ray_results". name (str) – Name of experiment (for Ray Tune). max_iters (int): Indicates the maximum number of epochs to run for each hyperparameter configuration sampled. This parameter is used for early stopping. Defaults to 1. Depending on the classifier type provided, a resource parameter (`resource_param = max_iter or n_estimators`) will be detected. The value of `resource_param` will be treated as a "max resource value", and all classifiers will be initialized with `max resource value // max_iters`, where `max_iters` is this defined parameter. On each epoch, resource_param (max_iter or n_estimators) is incremented by `max resource value // max_iters`. use_gpu (bool): Indicates whether to use gpu for fitting. Defaults to False. If True, training will start processes with the proper CUDA VISIBLE DEVICE settings set. If a Ray cluster has been initialized, all available GPUs will be used. loggers (list): A list of the names of the Tune loggers as strings to be used to log results. Possible values are "tensorboard", "csv", "mlflow", and "json" pipeline_auto_early_stop (bool): Only relevant if estimator is Pipeline object and early_stopping is enabled/True. If True, early stopping will be performed on the last stage of the pipeline (which must support early stopping). If False, early stopping will be determined by 'Pipeline.warm_start' or 'Pipeline.partial_fit' capabilities, which are by default not supported by standard SKlearn. Defaults to True. stopper (ray.tune.stopper.Stopper): Stopper objects passed to ````. time_budget_s (int|float|datetime.timedelta): Global time budget in seconds after which all trials are stopped. Can also be a ``datetime.timedelta`` object. mode (str): One of {min, max}. Determines whether objective is minimizing or maximizing the metric attribute. Defaults to "max". """ def __init__(self, estimator, param_grid, early_stopping=None, scoring=None, n_jobs=None, cv=5, refit=True, verbose=0, error_score="raise", return_train_score=False, local_dir="~/ray_results", name=None, max_iters=1, use_gpu=False, loggers=None, pipeline_auto_early_stop=True, stopper=None, time_budget_s=None, sk_n_jobs=None, mode=None): if sk_n_jobs not in (None, 1): raise ValueError( "Tune-sklearn no longer supports nested parallelism " "with new versions of joblib/sklearn. Don't set 'sk_n_jobs'.") super(TuneGridSearchCV, self).__init__( estimator=estimator, early_stopping=early_stopping, scoring=scoring, n_jobs=n_jobs or -1, cv=cv, refit=refit, error_score=error_score, return_train_score=return_train_score, local_dir=local_dir, name=name, max_iters=max_iters, verbose=verbose, use_gpu=use_gpu, loggers=loggers, pipeline_auto_early_stop=pipeline_auto_early_stop, stopper=stopper, time_budget_s=time_budget_s, mode=mode) check_error_warm_start(self.early_stop_type, param_grid, estimator) _check_param_grid_tune_grid_search(param_grid) self.param_grid = param_grid def _fill_config_hyperparam(self, config): """Fill in the ``config`` dictionary with the hyperparameters. Each distribution is converted to a list, then returns a dictionary showing the values of the hyperparameters that have been grid searched over. Args: config (`dict`): dictionary to be filled in as the configuration for ``. """ if isinstance(self.param_grid, list): return for key, distribution in self.param_grid.items(): if is_tune_grid_search(distribution): config[key] = distribution else: config[key] = tune.grid_search(list(distribution)) def _list_grid_num_samples(self): """Calculate the num_samples for ``. This is used when a list of dictionaries is passed in for the `param_grid` """ return len(list(ParameterGrid(self.param_grid))) def _tune_run(self, X, y, config, resources_per_trial, tune_params=None, fit_params=None): """Wrapper to call ````. Multiple estimators are generated when early stopping is possible, whereas a single estimator is generated when early stopping is not possible. Args: X (:obj:`array-like` (shape = [n_samples, n_features])): Training vector, where n_samples is the number of samples and n_features is the number of features. y (:obj:`array-like`): Shape of array expected to be [n_samples] or [n_samples, n_output]). Target relative to X for classification or regression; None for unsupervised learning. config (dict): Configurations such as hyperparameters to run ```` on. resources_per_trial (dict): Resources to use per trial within Ray. Accepted keys are `cpu`, `gpu` and custom resources, and values are integers specifying the number of each resource to use. tune_params (dict): User defined parameters passed to ````. Parameters inside `tune_params` override preset parameters. fit_params (dict): Parameters passed to the ``fit`` method of the estimator. Returns: analysis (`ExperimentAnalysis`): Object returned by ``. """ trainable = _Trainable if self.pipeline_auto_early_stop and check_is_pipeline( self.estimator) and self.early_stopping_: trainable = _PipelineTrainable if self.early_stopping_ is not None: estimator_list = [ clone(self.estimator) for _ in range(self.n_splits) ] else: estimator_list = [clone(self.estimator)] stopper = MaximumIterationStopper(max_iter=self.max_iters) if self.stopper: stopper = CombinedStopper(stopper, self.stopper) run_args = dict( scheduler=self.early_stopping_, reuse_actors=True, verbose=self.verbose, stop=stopper, config=config, fail_fast="raise", resources_per_trial=resources_per_trial, local_dir=self.local_dir,, callbacks=resolve_logger_callbacks(self.loggers, self.defined_loggers), time_budget_s=self.time_budget_s, metric=self._metric_name, mode=self.mode) if isinstance(self.param_grid, list): run_args.update( dict( search_alg=ListSearcher(self.param_grid), num_samples=self._list_grid_num_samples())) run_args = self._override_run_args_with_tune_params( run_args, tune_params) trainable = tune.with_parameters( trainable, X=X, y=y, estimator_list=estimator_list, fit_params=fit_params) with warnings.catch_warnings(): warnings.filterwarnings( "ignore", message="fail_fast='raise' " "detected.") analysis =, **run_args) return analysis