Using gingado to understand economic growth

An illustration with Barro and Lee (1994)
Author

Douglas K. G. Araujo

This notebook showcases one possible use of gingado by estimating economic growth across countries, using the dataset studied by Barro and Lee (1994). You can run this notebook interactively, by clicking on the appropriate link above.

This dataset has been widely studied in economics. Belloni, Chernozhukov, and Hansen (2011) and Giannone, Lenza, and Primiceri (2021) are two studies of this dataset that are most related to machine learning.

This notebook will use gingado to compare quickly setup a well-performing machine learning model and use its results as evidence to support the conditional convergence hypothesis; compare different classes of models (and their combination in a single model), and use and document the best performing alternative.

Because the notebook is for pedagogical purposes only, please bear in mind some aspects of the machine learning workflow (such as carefully thinking about the cross-validation strategy) are glossed over in this notebook. Also, only the key academic references are cited; more references can be found in the papers mentioned in this example.

Setting the stage

We will import packages as the work progresses. This will help highlight the specific steps in the workflow that gingado can be helpful with.

Code
import pandas as pd

The data is available in the online annex to Giannone, Lenza, and Primiceri (2021). In that paper, this dataset corresponds to what the authors call “macro2”. The original data, along with more information on the variables, can be found in this NBER website. A very helpful codebook is found in this repo.

Code
from gingado.datasets import load_BarroLee_1994

X, y = load_BarroLee_1994()

The dataset contains explanatory variables representing per-capita growth between 1960 and 1985, for 90 countries.

Code
X.columns
Index(['Unnamed: 0', 'gdpsh465', 'bmp1l', 'freeop', 'freetar', 'h65', 'hm65',
       'hf65', 'p65', 'pm65', 'pf65', 's65', 'sm65', 'sf65', 'fert65',
       'mort65', 'lifee065', 'gpop1', 'fert1', 'mort1', 'invsh41', 'geetot1',
       'geerec1', 'gde1', 'govwb1', 'govsh41', 'gvxdxe41', 'high65', 'highm65',
       'highf65', 'highc65', 'highcm65', 'highcf65', 'human65', 'humanm65',
       'humanf65', 'hyr65', 'hyrm65', 'hyrf65', 'no65', 'nom65', 'nof65',
       'pinstab1', 'pop65', 'worker65', 'pop1565', 'pop6565', 'sec65',
       'secm65', 'secf65', 'secc65', 'seccm65', 'seccf65', 'syr65', 'syrm65',
       'syrf65', 'teapri65', 'teasec65', 'ex1', 'im1', 'xr65', 'tot1'],
      dtype='object')
Code
X.head().T
0 1 2 3 4
Unnamed: 0 0.000000 1.000000 2.000000 3.000000 4.000000
gdpsh465 6.591674 6.829794 8.895082 7.565275 7.162397
bmp1l 0.283700 0.614100 0.000000 0.199700 0.174000
freeop 0.153491 0.313509 0.204244 0.248714 0.299252
freetar 0.043888 0.061827 0.009186 0.036270 0.037367
... ... ... ... ... ...
teasec65 17.300000 18.000000 20.700000 22.700000 17.600000
ex1 0.072900 0.094000 0.174100 0.126500 0.121100
im1 0.066700 0.143800 0.175000 0.149600 0.130800
xr65 0.348000 0.525000 1.082000 6.625000 2.500000
tot1 -0.014727 0.005750 -0.010040 -0.002195 0.003283

62 rows × 5 columns

The outcome variable is represented here:

Code
y.plot.hist(bins=90, title='GDP growth')
<Axes: title={'center': 'GDP growth'}, ylabel='Frequency'>

Establishing a benchmark model

Generally speaking, it is a good idea to establish a benchmark model at the first stages of development of the machine learning model. gingado offers a class of automatic benchmarks that can be used off-the-shelf depending on the task at hand: RegressionBenchmark and ClassificationBenchmark. It is also good to keep in mind that more advanced users can create their own benchmark on top of a base class provided by gingado: ggdBenchmark.

For this application, since we are interested in running a regression task, we will use RegressionBenchmark:

Code
from gingado.benchmark import RegressionBenchmark

What this object does is the following:

  • it creates a random forest

  • three different versions of the random forest are trained on the user data

  • the version that performs better is chosen as the benchmark

  • right after it is trained, the benchmark is documented using gingado’s ModelCard documenter.

The user can easily change the parameters above. For example, instead of a random forest the user might prefer a neural network as the benchmark. Or, in lieu of the default parameters provided by gingado, users might have their own idea of what could be a reasonable parameter space to search.

Random forests are chosen as the go-to benchmark algorithm because of their reasonably good performance in a wide variety of settings, the fact that they don’t require much data transformation (ie, normalising the data to have zero mean and one standard deviation), and by virtue of their relatively transparency about the importance of each regressor.

The first step is to initialise the benchmark object. At this time, we pass some arguments about how we want it to behave. In this case, we set the verbosity level to produce output related to each alternative considered. Then we fit it to the data.

Code
#####
#####
from sklearn.ensemble import RandomForestRegressor
rfr = RandomForestRegressor()
rfr.fit(X, y)
RandomForestRegressor()
In a Jupyter environment, please rerun this cell to show the HTML representation or trust the notebook.
On GitHub, the HTML representation is unable to render, please try loading this page with nbviewer.org.
Code
benchmark = RegressionBenchmark(verbose_grid=2)
benchmark.fit(X, y)
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=250; total time=   0.9s
[CV] END ................max_features=None, n_estimators=250; total time=   0.9s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.9s
[CV] END ................max_features=None, n_estimators=250; total time=   0.9s
[CV] END ................max_features=None, n_estimators=250; total time=   0.9s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
RegressionBenchmark(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                    verbose_grid=2)
In a Jupyter environment, please rerun this cell to show the HTML representation or trust the notebook.
On GitHub, the HTML representation is unable to render, please try loading this page with nbviewer.org.

As we can see above, with a few lines we have trained a random forest on the dataset. In this case, the benchmark was the better of six versions of the random forest, according to the default hyperparameters: 100 and 250 estimators were alternated with models for which the maximum number of regressors analysed by individual trees changesd fom the maximum, a square root and a log of the number of regressors. They were each trained using a 5-fold cross-validation.

Let’s see which one was the best performing in this case, and hence our benchmark model:

Code
pd.DataFrame(benchmark.benchmark.cv_results_).T
0 1 2 3 4 5
mean_fit_time 0.170493 0.422258 0.155009 0.373679 0.387799 0.940205
std_fit_time 0.011707 0.016794 0.004909 0.00299 0.015659 0.049657
mean_score_time 0.007007 0.016624 0.006951 0.015246 0.007209 0.015802
std_score_time 0.000776 0.000502 0.000563 0.000427 0.000759 0.000915
param_max_features sqrt sqrt log2 log2 None None
param_n_estimators 100 250 100 250 100 250
params {'max_features': 'sqrt', 'n_estimators': 100} {'max_features': 'sqrt', 'n_estimators': 250} {'max_features': 'log2', 'n_estimators': 100} {'max_features': 'log2', 'n_estimators': 250} {'max_features': None, 'n_estimators': 100} {'max_features': None, 'n_estimators': 250}
split0_test_score -0.453993 -0.521908 -0.517093 -0.439997 -0.338789 -0.316319
split1_test_score 0.273267 0.205131 0.186335 0.201619 0.233346 0.29916
split2_test_score -0.157436 -0.161351 -0.110423 -0.156365 -0.483266 -0.492917
split3_test_score 0.120799 0.059299 0.039521 -0.017347 -0.390511 -0.331298
split4_test_score 0.12008 0.060117 0.076103 0.16242 -0.065341 -0.035908
split5_test_score 0.435757 0.487523 0.440871 0.403803 0.40979 0.529639
split6_test_score -0.10016 -0.040821 0.087645 -0.085637 -0.435305 -0.406628
split7_test_score -0.146782 -0.188275 -0.384896 -0.229306 -0.815013 -0.543591
split8_test_score 0.154761 0.197665 0.128997 0.144068 0.102077 0.11356
split9_test_score 0.006076 0.144667 0.013465 0.18501 0.103891 0.224073
mean_test_score 0.025237 0.024205 -0.003948 0.016827 -0.167912 -0.096023
std_test_score 0.239512 0.260109 0.262395 0.23675 0.363384 0.354376
rank_test_score 1 2 4 3 6 5

The values above are calculated with \(R^2\), the default scoring function for a random forest from the scikit-learn package. Suppose that instead we would like a benchmark model that is optimised on the maximum error, ie a benchmark that minimises the worst deviation from prediction to ground truth for all the sample. These are the steps that we would take. Note that a more complete list of ready-made scoring parameters and how to create your own function can be found here.

Code
benchmark_lower_worsterror = RegressionBenchmark(scoring='max_error', verbose_grid=2)
benchmark_lower_worsterror.fit(X, y)
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.9s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   1.1s
[CV] END ................max_features=None, n_estimators=250; total time=   1.0s
[CV] END ................max_features=None, n_estimators=250; total time=   0.9s
RegressionBenchmark(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                    scoring='max_error', verbose_grid=2)
In a Jupyter environment, please rerun this cell to show the HTML representation or trust the notebook.
On GitHub, the HTML representation is unable to render, please try loading this page with nbviewer.org.
Code
pd.DataFrame(benchmark_lower_worsterror.benchmark.cv_results_).T
0 1 2 3 4 5
mean_fit_time 0.158678 0.414091 0.162536 0.379948 0.379933 0.966317
std_fit_time 0.003112 0.025825 0.009565 0.014874 0.015383 0.083762
mean_score_time 0.006551 0.016099 0.006698 0.015734 0.006666 0.0171
std_score_time 0.000578 0.001146 0.000456 0.000967 0.003575 0.003048
param_max_features sqrt sqrt log2 log2 None None
param_n_estimators 100 250 100 250 100 250
params {'max_features': 'sqrt', 'n_estimators': 100} {'max_features': 'sqrt', 'n_estimators': 250} {'max_features': 'log2', 'n_estimators': 100} {'max_features': 'log2', 'n_estimators': 250} {'max_features': None, 'n_estimators': 100} {'max_features': None, 'n_estimators': 250}
split0_test_score -0.132541 -0.130797 -0.132639 -0.137975 -0.172946 -0.163201
split1_test_score -0.075971 -0.074046 -0.078251 -0.074139 -0.105464 -0.089156
split2_test_score -0.136658 -0.138142 -0.135362 -0.135359 -0.150068 -0.150891
split3_test_score -0.0732 -0.066847 -0.070098 -0.066397 -0.066406 -0.075587
split4_test_score -0.131595 -0.131686 -0.140142 -0.137684 -0.130717 -0.126934
split5_test_score -0.145162 -0.142632 -0.144629 -0.140714 -0.149868 -0.150065
split6_test_score -0.081683 -0.075402 -0.085084 -0.083132 -0.097191 -0.093416
split7_test_score -0.118622 -0.11768 -0.12245 -0.117731 -0.110574 -0.118458
split8_test_score -0.094402 -0.100634 -0.104239 -0.102213 -0.094715 -0.097183
split9_test_score -0.138249 -0.134564 -0.140264 -0.135003 -0.15336 -0.150245
mean_test_score -0.112808 -0.111243 -0.115316 -0.113035 -0.123131 -0.121514
std_test_score 0.026967 0.028011 0.02703 0.027745 0.031719 0.029695
rank_test_score 2 1 4 3 6 5

Now we even have two benchmark models.

We could further tweak and adjust them, but one of the ideas behind having a benchmark is that it is simple and easy to set up.

Let’s retain only the first benchmark, for simplicity, and now look at the predictions, comparing them to the original growth values.

Code
y_pred = benchmark.predict(X)

pd.DataFrame({
    'y': y,
    'y_pred': y_pred
    }).plot.scatter(
        x='y', y='y_pred',
         grid=True, 
         title='Actual and predicted outcome',
         xlabel='actual GDP growth',
         ylabel='predicted GDP growth')
<Axes: title={'center': 'Actual and predicted outcome'}, xlabel='actual GDP growth', ylabel='predicted GDP growth'>

And now a histogram of the benchmark’s errors:

Code
pd.DataFrame(y - y_pred).plot.hist(bins=30, title='Residual')
<Axes: title={'center': 'Residual'}, ylabel='Frequency'>

Since the benchmark is a random forest model, we can see what are the most important regressors, measured as the average reduction in impurity across the trees in the random forest that actually use that particular regressor. They are scaled so that the sum for all features is one. Higher importance amounts indicate that that particular regressor is a more important contributor to the final prediction.

Code
regressor_importance = pd.DataFrame(
    benchmark.benchmark.best_estimator_.feature_importances_, 
    index=X.columns, 
    columns=["Importance"]
    )

regressor_importance.sort_values(by="Importance", ascending=False) \
    .plot.bar(figsize=(20, 8), title='Regressor importance')
<Axes: title={'center': 'Regressor importance'}>

From the graph above, we can see that the regressor bmp1l (black-market premium on foreign exchange) predominates. Interestingly, Belloni, Chernozhukov, and Hansen (2011) using squared-root lasso also find this regressor to be important.

Testing the conditional converge hypothesis

Now we can leverage our automatic benchmark model to test the conditional converge hypothesis - ie, the preposition that countries with lower starting GDP tend to grow faster than other comparable countries. In other words, this hypothesis predicts that when GDP growth is regressed on the level of past GDP and on an adequate set of covariates \(X\), the coefficient on past GDP levels are negative.

Since we have the results for the importance of each regressor in separating countries by their growth result, we can compare the estimated coefficient for GDP levels in regressions that include different regressors in the vector \(X\). To maintain this example a simple exercise, the following three models are estimated:

  • \(X\) contains the five most important regressors, as estimated by the benchmark model (see the graph above)
  • \(X\) contains the five least important regressors, from the same estimation as above
  • \(X\) is the empty set - in other words, this is a simple equation on GDP growth on GDP levels

A result that would be consistent with the conditionality of the conditional convergence hypothesis is the first equation resulting in a negative coefficient for starting GDP, while the following two equations may not necessarily be successful in identifying a negative coefficient. This is because the least important regressors are not likely to have sufficient predictive power to separate countries into comparable groups.

The five more and less important regressors are:

Code
top_five = regressor_importance.sort_values(by="Importance", ascending=False).head(5)
bottom_five = regressor_importance.sort_values(by="Importance", ascending=True).head(5)

top_five, bottom_five
(         Importance
 bmp1l      0.047130
 pop6565    0.043468
 ex1        0.031683
 geerec1    0.030568
 gpop1      0.028123,
          Importance
 pf65       0.002350
 p65        0.003418
 secc65     0.003542
 human65    0.005687
 sf65       0.006447)
Code
import statsmodels.api as sm
Code
gdp_level = 'gdpsh465'
Code
X_topfive = X[[gdp_level] + list(top_five.index)]
X_topfive = sm.add_constant(X_topfive)
X_topfive.head()
const gdpsh465 bmp1l pop6565 ex1 geerec1 gpop1
0 1.0 6.591674 0.2837 0.027591 0.0729 0.0176 0.0203
1 1.0 6.829794 0.6141 0.035637 0.0940 0.0369 0.0185
2 1.0 8.895082 0.0000 0.076685 0.1741 0.0365 0.0188
3 1.0 7.565275 0.1997 0.031039 0.1265 0.0350 0.0345
4 1.0 7.162397 0.1740 0.026281 0.1211 0.0224 0.0310
Code
X_bottomfive = X[[gdp_level] + list(bottom_five.index)]
X_bottomfive = sm.add_constant(X_bottomfive)
X_bottomfive.head()
const gdpsh465 pf65 p65 secc65 human65 sf65
0 1.0 6.591674 0.21 0.29 0.13 0.301 0.02
1 1.0 6.829794 0.65 0.91 1.36 0.706 0.09
2 1.0 8.895082 1.00 1.00 15.68 8.317 0.51
3 1.0 7.565275 1.00 1.00 2.76 3.833 0.31
4 1.0 7.162397 0.81 0.82 2.17 1.900 0.13
Code
X_onlyGDPlevel = sm.add_constant(X[gdp_level])
X_onlyGDPlevel.head()
const gdpsh465
0 1.0 6.591674
1 1.0 6.829794
2 1.0 8.895082
3 1.0 7.565275
4 1.0 7.162397
Code
models = dict(
    topfive = sm.OLS(y, X_topfive).fit(),
    bottomfive = sm.OLS(y, X_bottomfive).fit(),
    onlyGDPlevel = sm.OLS(y, X_onlyGDPlevel).fit()
)
Code
coefs = pd.DataFrame({name: model.conf_int().loc[gdp_level] for name, model in models.items()})
coefs.loc[0.5] = [model.params[gdp_level] for _, model in models.items()]
coefs = coefs.sort_index().reset_index(drop=True)
coefs.index = ['[0.025', 'coef on GDP levels', '0.975]']
coefs
topfive bottomfive onlyGDPlevel
[0.025 -0.037318 -0.037676 -0.010810
coef on GDP levels -0.019321 -0.013389 0.001317
0.975] -0.001324 0.010898 0.013444

The equation using the top five regressors in explanatory power yielded a coefficient that is statistically speaking negative under the usual confidence interval levels. In contrast, the regression using the bottom five regressors failed to maintain that level of statistical significance (although the coefficient point estimate was still negative). And finally the regression on GDP level solely resulted, as in the past literature, on a point estimate that is also statistically not different than zero.

These results above offer a different way to add evidence to the conditional convergence hypothesis. In particular, with the help of gingado’s RegressionBenchmark model, it is possible to identify which covariates can meaningfully serve as covariates in a growth equation from those that cannot. This is important because if the covariate selection for some reason included only variables with little explanatory power instead of the most relevant ones, an economist might erroneously reach a different conclusion.

Model documentation

Importantly for model documentation, the benchmark already has some baseline documentation set up. If the user wishes, they can use that as a basis to document their model. Note that the output is in a raw format that is suitable for machine reading and writing. Intermediary and advanced users may wish to use that format to construct personalised forms, documents, etc.

Code
benchmark.model_documentation.show_json()
{'model_details': {'developer': 'Person or organisation developing the model',
  'datetime': '2024-02-27 15:36:04 ',
  'version': 'Model version',
  'type': 'Model type',
  'info': {'_estimator_type': 'regressor',
   'best_estimator_': RandomForestRegressor(max_features='sqrt', oob_score=True),
   'best_index_': 0,
   'best_params_': {'max_features': 'sqrt', 'n_estimators': 100},
   'best_score_': 0.02523692933917774,
   'cv_results_': {'mean_fit_time': array([0.17049298, 0.42225838, 0.15500853, 0.37367864, 0.38779874,
           0.94020476]),
    'std_fit_time': array([0.01170658, 0.01679355, 0.00490875, 0.00299033, 0.01565909,
           0.04965667]),
    'mean_score_time': array([0.00700731, 0.01662374, 0.00695083, 0.01524622, 0.00720918,
           0.0158015 ]),
    'std_score_time': array([0.00077616, 0.00050157, 0.00056257, 0.00042689, 0.00075885,
           0.00091489]),
    'param_max_features': masked_array(data=['sqrt', 'sqrt', 'log2', 'log2', None, None],
                 mask=[False, False, False, False, False, False],
           fill_value='?',
                dtype=object),
    'param_n_estimators': masked_array(data=[100, 250, 100, 250, 100, 250],
                 mask=[False, False, False, False, False, False],
           fill_value='?',
                dtype=object),
    'params': [{'max_features': 'sqrt', 'n_estimators': 100},
     {'max_features': 'sqrt', 'n_estimators': 250},
     {'max_features': 'log2', 'n_estimators': 100},
     {'max_features': 'log2', 'n_estimators': 250},
     {'max_features': None, 'n_estimators': 100},
     {'max_features': None, 'n_estimators': 250}],
    'split0_test_score': array([-0.45399254, -0.52190829, -0.51709337, -0.43999707, -0.33878882,
           -0.31631861]),
    'split1_test_score': array([0.27326693, 0.20513145, 0.18633491, 0.20161913, 0.23334643,
           0.29916025]),
    'split2_test_score': array([-0.15743575, -0.1613513 , -0.11042349, -0.15636454, -0.48326606,
           -0.49291682]),
    'split3_test_score': array([ 0.12079877,  0.05929888,  0.0395209 , -0.01734706, -0.39051119,
           -0.33129809]),
    'split4_test_score': array([ 0.12008   ,  0.06011742,  0.07610285,  0.16241968, -0.06534097,
           -0.0359082 ]),
    'split5_test_score': array([0.43575685, 0.48752286, 0.44087083, 0.40380316, 0.40979045,
           0.52963864]),
    'split6_test_score': array([-0.10016029, -0.04082145,  0.08764482, -0.08563733, -0.43530541,
           -0.40662807]),
    'split7_test_score': array([-0.14678241, -0.18827504, -0.38489583, -0.22930611, -0.81501321,
           -0.54359099]),
    'split8_test_score': array([0.15476144, 0.19766545, 0.12899651, 0.14406783, 0.10207731,
           0.11355994]),
    'split9_test_score': array([0.0060763 , 0.14466697, 0.01346488, 0.18500953, 0.10389089,
           0.22407262]),
    'mean_test_score': array([ 0.02523693,  0.02420469, -0.0039477 ,  0.01682672, -0.16791206,
           -0.09602293]),
    'std_test_score': array([0.23951238, 0.26010878, 0.26239516, 0.23674952, 0.36338368,
           0.35437609]),
    'rank_test_score': array([1, 2, 4, 3, 6, 5])},
   'multimetric_': False,
   'n_features_in_': 62,
   'n_splits_': 10,
   'refit_time_': 0.1588914394378662,
   'scorer_': <function sklearn.metrics._scorer._passthrough_scorer(estimator, *args, **kwargs)>},
  'paper': 'Paper or other resource for more information',
  'citation': 'Citation details',
  'license': 'License',
  'contact': 'Where to send questions or comments about the model'},
 'intended_use': {'primary_uses': 'Primary intended uses',
  'primary_users': 'Primary intended users',
  'out_of_scope': 'Out-of-scope use cases'},
 'factors': {'relevant': 'Relevant factors',
  'evaluation': 'Evaluation factors'},
 'metrics': {'performance_measures': 'Model performance measures',
  'thresholds': 'Decision thresholds',
  'variation_approaches': 'Variation approaches'},
 'evaluation_data': {'datasets': 'Datasets',
  'motivation': 'Motivation',
  'preprocessing': 'Preprocessing'},
 'training_data': {'training_data': 'Information on training data'},
 'quant_analyses': {'unitary': 'Unitary results',
  'intersectional': 'Intersectional results'},
 'ethical_considerations': {'sensitive_data': 'Does the model use any sensitive data (e.g., protected classes)?',
  'human_life': 'Is the model intended to inform decisions about matters central to human life or flourishing - e.g., health or safety? Or could it be used in such a way?',
  'mitigations': 'What risk mitigation strategies were used during model development?',
  'risks_and_harms': 'What risks may be present in model usage? Try to identify the potential recipients,likelihood, and magnitude of harms. If these cannot be determined, note that they were considered but remain unknown',
  'use_cases': 'Are there any known model use cases that are especially fraught?',
  'additional_information': 'If possible, this section should also include any additional ethical considerations that went into model development, for example, review by an external board, or testing with a specific community.'},
 'caveats_recommendations': {'caveats': 'For example, did the results suggest any further testing? Were there any relevant groups that were not represented in the evaluation dataset?',
  'recommendations': 'Are there additional recommendations for model use? What are the ideal characteristics of an evaluation dataset for this model?'}}

Since there is some information in the model documentation that was automatically added, we might want to concentrate on the fields in the model card that are yet to be answered. Actually, this is the purpose of gingado’s automatic documentation: to afford users more time so they can invest, if they want, on model documentation.

Code
benchmark.model_documentation.open_questions()
['model_details__developer',
 'model_details__version',
 'model_details__type',
 'model_details__paper',
 'model_details__citation',
 'model_details__license',
 'model_details__contact',
 'intended_use__primary_uses',
 'intended_use__primary_users',
 'intended_use__out_of_scope',
 'factors__relevant',
 'factors__evaluation',
 'metrics__performance_measures',
 'metrics__thresholds',
 'metrics__variation_approaches',
 'evaluation_data__datasets',
 'evaluation_data__motivation',
 'evaluation_data__preprocessing',
 'training_data__training_data',
 'quant_analyses__unitary',
 'quant_analyses__intersectional',
 'ethical_considerations__sensitive_data',
 'ethical_considerations__human_life',
 'ethical_considerations__mitigations',
 'ethical_considerations__risks_and_harms',
 'ethical_considerations__use_cases',
 'ethical_considerations__additional_information',
 'caveats_recommendations__caveats',
 'caveats_recommendations__recommendations']

Let’s fill some information:

Code
benchmark.model_documentation.fill_info({
    'intended_use': {
        'primary_uses': 'This model is trained for pedagogical uses only.',
        'primary_users': 'Everyone is welcome to follow the description showing the development of this benchmark.'
    }
})

Note the format, based on a Python dictionary. In particular, the open_questions method results include keys divided by double underscores. As seen above, these should be interpreted as different levels of the documentation template, leading to a nested dictionary.

Now when we confirm that the questions answered above are no longer “open questions”:

Code
benchmark.model_documentation.open_questions()
['model_details__developer',
 'model_details__version',
 'model_details__type',
 'model_details__paper',
 'model_details__citation',
 'model_details__license',
 'model_details__contact',
 'intended_use__out_of_scope',
 'factors__relevant',
 'factors__evaluation',
 'metrics__performance_measures',
 'metrics__thresholds',
 'metrics__variation_approaches',
 'evaluation_data__datasets',
 'evaluation_data__motivation',
 'evaluation_data__preprocessing',
 'training_data__training_data',
 'quant_analyses__unitary',
 'quant_analyses__intersectional',
 'ethical_considerations__sensitive_data',
 'ethical_considerations__human_life',
 'ethical_considerations__mitigations',
 'ethical_considerations__risks_and_harms',
 'ethical_considerations__use_cases',
 'ethical_considerations__additional_information',
 'caveats_recommendations__caveats',
 'caveats_recommendations__recommendations']

If we want, at any time we can save the documentation to a local JSON file, as well as read another document.

Trying out model alternatives

The benchmark model may be enough for some analyses, or maybe the user is interested in using the benchmark to explore the data and have an understanding of the importance of each regressor, to concentrate their work on data that can be meaningful for their purposes. But oftentimes a user will want to seek a machine learning model that performs as well as possible.

For users that want to manually create other models, gingado allows the possibility of comparing them with the benchmark. If the user model is better, it becomes the new benchmark!

For the following analyses, we will use K-fold as cross-validation, with 5 splits of the sample.

First candidate: a gradient boosting tree

Code
import numpy as np
from sklearn.model_selection import GridSearchCV
from sklearn.ensemble import GradientBoostingRegressor
Code
param_grid = {
    'learning_rate': [0.01, 0.1, 0.25],
    'max_depth': [3, 6, 9]
}

reg_gradbooster = GradientBoostingRegressor()

gradboosterg_grid = GridSearchCV(
    reg_gradbooster,
    param_grid,
    n_jobs=-1,
    verbose=2
).fit(X, y)
Fitting 5 folds for each of 9 candidates, totalling 45 fits
Code
y_pred = gradboosterg_grid.predict(X)
pd.DataFrame({
    'y': y,
    'y_pred': y_pred
    }).plot.scatter(x='y', y='y_pred', grid=True)
<Axes: xlabel='y', ylabel='y_pred'>

Code
pd.DataFrame(y - y_pred).plot.hist(bins=30)
<Axes: ylabel='Frequency'>

Second candidate: lasso

Code
from sklearn.linear_model import Lasso
Code
param_grid = {
    'alpha': [0.5, 1, 1.25],
}

reg_lasso = Lasso(fit_intercept=True)

lasso_grid = GridSearchCV(
    reg_lasso,
    param_grid,
    n_jobs=-1,
    verbose=2
).fit(X, y)
Fitting 5 folds for each of 3 candidates, totalling 15 fits
Code
y_pred = lasso_grid.predict(X)
pd.DataFrame({
    'y': y,
    'y_pred': y_pred
    }).plot.scatter(x='y', y='y_pred', grid=True)
<Axes: xlabel='y', ylabel='y_pred'>

Code
pd.DataFrame(y - y_pred).plot.hist(bins=30)
<Axes: ylabel='Frequency'>

Comparing the models with the benchmark

gingado allows users to compare different candidate models with the existing benchmark in a very simple way: using the compare method.

Code
candidates = [gradboosterg_grid, lasso_grid]
benchmark.compare(X, y, candidates)
Fitting 10 folds for each of 4 candidates, totalling 40 fits
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END candidate_estimator=GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
             estimator=RandomForestRegressor(oob_score=True),
             param_grid={'max_features': ['sqrt', 'log2', None],
                         'n_estimators': [100, 250]},
             verbose=2), candidate_estimator__cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None), candidate_estimator__error_score=nan, candidate_estimator__estimator=RandomForestRegressor(oob_score=True), candidate_estimator__estimator__bootstrap=True, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=squared_error, candidate_estimator__estimator__max_depth=None, candidate_estimator__estimator__max_features=1.0, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__max_samples=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_jobs=None, candidate_estimator__estimator__oob_score=True, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=None, candidate_estimator__param_grid={'n_estimators': [100, 250], 'max_features': ['sqrt', 'log2', None]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=  23.4s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.4s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END candidate_estimator=GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
             estimator=RandomForestRegressor(oob_score=True),
             param_grid={'max_features': ['sqrt', 'log2', None],
                         'n_estimators': [100, 250]},
             verbose=2), candidate_estimator__cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None), candidate_estimator__error_score=nan, candidate_estimator__estimator=RandomForestRegressor(oob_score=True), candidate_estimator__estimator__bootstrap=True, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=squared_error, candidate_estimator__estimator__max_depth=None, candidate_estimator__estimator__max_features=1.0, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__max_samples=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_jobs=None, candidate_estimator__estimator__oob_score=True, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=None, candidate_estimator__param_grid={'n_estimators': [100, 250], 'max_features': ['sqrt', 'log2', None]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=  23.5s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.5s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.9s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END candidate_estimator=GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
             estimator=RandomForestRegressor(oob_score=True),
             param_grid={'max_features': ['sqrt', 'log2', None],
                         'n_estimators': [100, 250]},
             verbose=2), candidate_estimator__cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None), candidate_estimator__error_score=nan, candidate_estimator__estimator=RandomForestRegressor(oob_score=True), candidate_estimator__estimator__bootstrap=True, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=squared_error, candidate_estimator__estimator__max_depth=None, candidate_estimator__estimator__max_features=1.0, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__max_samples=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_jobs=None, candidate_estimator__estimator__oob_score=True, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=None, candidate_estimator__param_grid={'n_estimators': [100, 250], 'max_features': ['sqrt', 'log2', None]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=  24.6s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.2s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.4s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.4s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.5s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.9s
[CV] END candidate_estimator=GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
             estimator=RandomForestRegressor(oob_score=True),
             param_grid={'max_features': ['sqrt', 'log2', None],
                         'n_estimators': [100, 250]},
             verbose=2), candidate_estimator__cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None), candidate_estimator__error_score=nan, candidate_estimator__estimator=RandomForestRegressor(oob_score=True), candidate_estimator__estimator__bootstrap=True, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=squared_error, candidate_estimator__estimator__max_depth=None, candidate_estimator__estimator__max_features=1.0, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__max_samples=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_jobs=None, candidate_estimator__estimator__oob_score=True, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=None, candidate_estimator__param_grid={'n_estimators': [100, 250], 'max_features': ['sqrt', 'log2', None]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=  25.4s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.2s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.2s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.6s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.4s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.4s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END candidate_estimator=GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
             estimator=RandomForestRegressor(oob_score=True),
             param_grid={'max_features': ['sqrt', 'log2', None],
                         'n_estimators': [100, 250]},
             verbose=2), candidate_estimator__cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None), candidate_estimator__error_score=nan, candidate_estimator__estimator=RandomForestRegressor(oob_score=True), candidate_estimator__estimator__bootstrap=True, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=squared_error, candidate_estimator__estimator__max_depth=None, candidate_estimator__estimator__max_features=1.0, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__max_samples=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_jobs=None, candidate_estimator__estimator__oob_score=True, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=None, candidate_estimator__param_grid={'n_estimators': [100, 250], 'max_features': ['sqrt', 'log2', None]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=  24.3s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END candidate_estimator=GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
             estimator=RandomForestRegressor(oob_score=True),
             param_grid={'max_features': ['sqrt', 'log2', None],
                         'n_estimators': [100, 250]},
             verbose=2), candidate_estimator__cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None), candidate_estimator__error_score=nan, candidate_estimator__estimator=RandomForestRegressor(oob_score=True), candidate_estimator__estimator__bootstrap=True, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=squared_error, candidate_estimator__estimator__max_depth=None, candidate_estimator__estimator__max_features=1.0, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__max_samples=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_jobs=None, candidate_estimator__estimator__oob_score=True, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=None, candidate_estimator__param_grid={'n_estimators': [100, 250], 'max_features': ['sqrt', 'log2', None]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=  23.7s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END candidate_estimator=GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
             estimator=RandomForestRegressor(oob_score=True),
             param_grid={'max_features': ['sqrt', 'log2', None],
                         'n_estimators': [100, 250]},
             verbose=2), candidate_estimator__cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None), candidate_estimator__error_score=nan, candidate_estimator__estimator=RandomForestRegressor(oob_score=True), candidate_estimator__estimator__bootstrap=True, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=squared_error, candidate_estimator__estimator__max_depth=None, candidate_estimator__estimator__max_features=1.0, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__max_samples=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_jobs=None, candidate_estimator__estimator__oob_score=True, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=None, candidate_estimator__param_grid={'n_estimators': [100, 250], 'max_features': ['sqrt', 'log2', None]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=  23.6s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END candidate_estimator=GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
             estimator=RandomForestRegressor(oob_score=True),
             param_grid={'max_features': ['sqrt', 'log2', None],
                         'n_estimators': [100, 250]},
             verbose=2), candidate_estimator__cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None), candidate_estimator__error_score=nan, candidate_estimator__estimator=RandomForestRegressor(oob_score=True), candidate_estimator__estimator__bootstrap=True, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=squared_error, candidate_estimator__estimator__max_depth=None, candidate_estimator__estimator__max_features=1.0, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__max_samples=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_jobs=None, candidate_estimator__estimator__oob_score=True, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=None, candidate_estimator__param_grid={'n_estimators': [100, 250], 'max_features': ['sqrt', 'log2', None]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=  24.1s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.0s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END candidate_estimator=GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
             estimator=RandomForestRegressor(oob_score=True),
             param_grid={'max_features': ['sqrt', 'log2', None],
                         'n_estimators': [100, 250]},
             verbose=2), candidate_estimator__cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None), candidate_estimator__error_score=nan, candidate_estimator__estimator=RandomForestRegressor(oob_score=True), candidate_estimator__estimator__bootstrap=True, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=squared_error, candidate_estimator__estimator__max_depth=None, candidate_estimator__estimator__max_features=1.0, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__max_samples=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_jobs=None, candidate_estimator__estimator__oob_score=True, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=None, candidate_estimator__param_grid={'n_estimators': [100, 250], 'max_features': ['sqrt', 'log2', None]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=  24.4s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END candidate_estimator=GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
             estimator=RandomForestRegressor(oob_score=True),
             param_grid={'max_features': ['sqrt', 'log2', None],
                         'n_estimators': [100, 250]},
             verbose=2), candidate_estimator__cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None), candidate_estimator__error_score=nan, candidate_estimator__estimator=RandomForestRegressor(oob_score=True), candidate_estimator__estimator__bootstrap=True, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=squared_error, candidate_estimator__estimator__max_depth=None, candidate_estimator__estimator__max_features=1.0, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__max_samples=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_jobs=None, candidate_estimator__estimator__oob_score=True, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=None, candidate_estimator__param_grid={'n_estimators': [100, 250], 'max_features': ['sqrt', 'log2', None]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=  24.2s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
[CV] END candidate_estimator=GridSearchCV(estimator=GradientBoostingRegressor(), n_jobs=-1,
             param_grid={'learning_rate': [0.01, 0.1, 0.25],
                         'max_depth': [3, 6, 9]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=GradientBoostingRegressor(), candidate_estimator__estimator__alpha=0.9, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=friedman_mse, candidate_estimator__estimator__init=None, candidate_estimator__estimator__learning_rate=0.1, candidate_estimator__estimator__loss=squared_error, candidate_estimator__estimator__max_depth=3, candidate_estimator__estimator__max_features=None, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_iter_no_change=None, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__subsample=1.0, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__validation_fraction=0.1, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'learning_rate': [0.01, 0.1, 0.25], 'max_depth': [3, 6, 9]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   2.2s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
[CV] END candidate_estimator=GridSearchCV(estimator=GradientBoostingRegressor(), n_jobs=-1,
             param_grid={'learning_rate': [0.01, 0.1, 0.25],
                         'max_depth': [3, 6, 9]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=GradientBoostingRegressor(), candidate_estimator__estimator__alpha=0.9, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=friedman_mse, candidate_estimator__estimator__init=None, candidate_estimator__estimator__learning_rate=0.1, candidate_estimator__estimator__loss=squared_error, candidate_estimator__estimator__max_depth=3, candidate_estimator__estimator__max_features=None, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_iter_no_change=None, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__subsample=1.0, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__validation_fraction=0.1, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'learning_rate': [0.01, 0.1, 0.25], 'max_depth': [3, 6, 9]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   2.1s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
[CV] END candidate_estimator=GridSearchCV(estimator=GradientBoostingRegressor(), n_jobs=-1,
             param_grid={'learning_rate': [0.01, 0.1, 0.25],
                         'max_depth': [3, 6, 9]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=GradientBoostingRegressor(), candidate_estimator__estimator__alpha=0.9, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=friedman_mse, candidate_estimator__estimator__init=None, candidate_estimator__estimator__learning_rate=0.1, candidate_estimator__estimator__loss=squared_error, candidate_estimator__estimator__max_depth=3, candidate_estimator__estimator__max_features=None, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_iter_no_change=None, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__subsample=1.0, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__validation_fraction=0.1, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'learning_rate': [0.01, 0.1, 0.25], 'max_depth': [3, 6, 9]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   2.1s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
[CV] END candidate_estimator=GridSearchCV(estimator=GradientBoostingRegressor(), n_jobs=-1,
             param_grid={'learning_rate': [0.01, 0.1, 0.25],
                         'max_depth': [3, 6, 9]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=GradientBoostingRegressor(), candidate_estimator__estimator__alpha=0.9, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=friedman_mse, candidate_estimator__estimator__init=None, candidate_estimator__estimator__learning_rate=0.1, candidate_estimator__estimator__loss=squared_error, candidate_estimator__estimator__max_depth=3, candidate_estimator__estimator__max_features=None, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_iter_no_change=None, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__subsample=1.0, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__validation_fraction=0.1, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'learning_rate': [0.01, 0.1, 0.25], 'max_depth': [3, 6, 9]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   2.1s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
[CV] END candidate_estimator=GridSearchCV(estimator=GradientBoostingRegressor(), n_jobs=-1,
             param_grid={'learning_rate': [0.01, 0.1, 0.25],
                         'max_depth': [3, 6, 9]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=GradientBoostingRegressor(), candidate_estimator__estimator__alpha=0.9, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=friedman_mse, candidate_estimator__estimator__init=None, candidate_estimator__estimator__learning_rate=0.1, candidate_estimator__estimator__loss=squared_error, candidate_estimator__estimator__max_depth=3, candidate_estimator__estimator__max_features=None, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_iter_no_change=None, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__subsample=1.0, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__validation_fraction=0.1, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'learning_rate': [0.01, 0.1, 0.25], 'max_depth': [3, 6, 9]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   2.1s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
[CV] END candidate_estimator=GridSearchCV(estimator=GradientBoostingRegressor(), n_jobs=-1,
             param_grid={'learning_rate': [0.01, 0.1, 0.25],
                         'max_depth': [3, 6, 9]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=GradientBoostingRegressor(), candidate_estimator__estimator__alpha=0.9, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=friedman_mse, candidate_estimator__estimator__init=None, candidate_estimator__estimator__learning_rate=0.1, candidate_estimator__estimator__loss=squared_error, candidate_estimator__estimator__max_depth=3, candidate_estimator__estimator__max_features=None, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_iter_no_change=None, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__subsample=1.0, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__validation_fraction=0.1, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'learning_rate': [0.01, 0.1, 0.25], 'max_depth': [3, 6, 9]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   2.2s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
[CV] END candidate_estimator=GridSearchCV(estimator=GradientBoostingRegressor(), n_jobs=-1,
             param_grid={'learning_rate': [0.01, 0.1, 0.25],
                         'max_depth': [3, 6, 9]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=GradientBoostingRegressor(), candidate_estimator__estimator__alpha=0.9, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=friedman_mse, candidate_estimator__estimator__init=None, candidate_estimator__estimator__learning_rate=0.1, candidate_estimator__estimator__loss=squared_error, candidate_estimator__estimator__max_depth=3, candidate_estimator__estimator__max_features=None, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_iter_no_change=None, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__subsample=1.0, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__validation_fraction=0.1, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'learning_rate': [0.01, 0.1, 0.25], 'max_depth': [3, 6, 9]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   2.6s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
[CV] END candidate_estimator=GridSearchCV(estimator=GradientBoostingRegressor(), n_jobs=-1,
             param_grid={'learning_rate': [0.01, 0.1, 0.25],
                         'max_depth': [3, 6, 9]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=GradientBoostingRegressor(), candidate_estimator__estimator__alpha=0.9, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=friedman_mse, candidate_estimator__estimator__init=None, candidate_estimator__estimator__learning_rate=0.1, candidate_estimator__estimator__loss=squared_error, candidate_estimator__estimator__max_depth=3, candidate_estimator__estimator__max_features=None, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_iter_no_change=None, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__subsample=1.0, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__validation_fraction=0.1, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'learning_rate': [0.01, 0.1, 0.25], 'max_depth': [3, 6, 9]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   2.5s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
[CV] END candidate_estimator=GridSearchCV(estimator=GradientBoostingRegressor(), n_jobs=-1,
             param_grid={'learning_rate': [0.01, 0.1, 0.25],
                         'max_depth': [3, 6, 9]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=GradientBoostingRegressor(), candidate_estimator__estimator__alpha=0.9, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=friedman_mse, candidate_estimator__estimator__init=None, candidate_estimator__estimator__learning_rate=0.1, candidate_estimator__estimator__loss=squared_error, candidate_estimator__estimator__max_depth=3, candidate_estimator__estimator__max_features=None, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_iter_no_change=None, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__subsample=1.0, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__validation_fraction=0.1, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'learning_rate': [0.01, 0.1, 0.25], 'max_depth': [3, 6, 9]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   3.1s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
[CV] END candidate_estimator=GridSearchCV(estimator=GradientBoostingRegressor(), n_jobs=-1,
             param_grid={'learning_rate': [0.01, 0.1, 0.25],
                         'max_depth': [3, 6, 9]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=GradientBoostingRegressor(), candidate_estimator__estimator__alpha=0.9, candidate_estimator__estimator__ccp_alpha=0.0, candidate_estimator__estimator__criterion=friedman_mse, candidate_estimator__estimator__init=None, candidate_estimator__estimator__learning_rate=0.1, candidate_estimator__estimator__loss=squared_error, candidate_estimator__estimator__max_depth=3, candidate_estimator__estimator__max_features=None, candidate_estimator__estimator__max_leaf_nodes=None, candidate_estimator__estimator__min_impurity_decrease=0.0, candidate_estimator__estimator__min_samples_leaf=1, candidate_estimator__estimator__min_samples_split=2, candidate_estimator__estimator__min_weight_fraction_leaf=0.0, candidate_estimator__estimator__n_estimators=100, candidate_estimator__estimator__n_iter_no_change=None, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__subsample=1.0, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__validation_fraction=0.1, candidate_estimator__estimator__verbose=0, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'learning_rate': [0.01, 0.1, 0.25], 'max_depth': [3, 6, 9]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   3.6s
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=GridSearchCV(estimator=Lasso(), n_jobs=-1, param_grid={'alpha': [0.5, 1, 1.25]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=Lasso(), candidate_estimator__estimator__alpha=1.0, candidate_estimator__estimator__copy_X=True, candidate_estimator__estimator__fit_intercept=True, candidate_estimator__estimator__max_iter=1000, candidate_estimator__estimator__positive=False, candidate_estimator__estimator__precompute=False, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__selection=cyclic, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'alpha': [0.5, 1, 1.25]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   0.0s
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=GridSearchCV(estimator=Lasso(), n_jobs=-1, param_grid={'alpha': [0.5, 1, 1.25]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=Lasso(), candidate_estimator__estimator__alpha=1.0, candidate_estimator__estimator__copy_X=True, candidate_estimator__estimator__fit_intercept=True, candidate_estimator__estimator__max_iter=1000, candidate_estimator__estimator__positive=False, candidate_estimator__estimator__precompute=False, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__selection=cyclic, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'alpha': [0.5, 1, 1.25]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   0.0s
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=GridSearchCV(estimator=Lasso(), n_jobs=-1, param_grid={'alpha': [0.5, 1, 1.25]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=Lasso(), candidate_estimator__estimator__alpha=1.0, candidate_estimator__estimator__copy_X=True, candidate_estimator__estimator__fit_intercept=True, candidate_estimator__estimator__max_iter=1000, candidate_estimator__estimator__positive=False, candidate_estimator__estimator__precompute=False, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__selection=cyclic, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'alpha': [0.5, 1, 1.25]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   0.0s
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=GridSearchCV(estimator=Lasso(), n_jobs=-1, param_grid={'alpha': [0.5, 1, 1.25]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=Lasso(), candidate_estimator__estimator__alpha=1.0, candidate_estimator__estimator__copy_X=True, candidate_estimator__estimator__fit_intercept=True, candidate_estimator__estimator__max_iter=1000, candidate_estimator__estimator__positive=False, candidate_estimator__estimator__precompute=False, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__selection=cyclic, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'alpha': [0.5, 1, 1.25]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   0.0s
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=GridSearchCV(estimator=Lasso(), n_jobs=-1, param_grid={'alpha': [0.5, 1, 1.25]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=Lasso(), candidate_estimator__estimator__alpha=1.0, candidate_estimator__estimator__copy_X=True, candidate_estimator__estimator__fit_intercept=True, candidate_estimator__estimator__max_iter=1000, candidate_estimator__estimator__positive=False, candidate_estimator__estimator__precompute=False, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__selection=cyclic, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'alpha': [0.5, 1, 1.25]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   0.0s
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=GridSearchCV(estimator=Lasso(), n_jobs=-1, param_grid={'alpha': [0.5, 1, 1.25]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=Lasso(), candidate_estimator__estimator__alpha=1.0, candidate_estimator__estimator__copy_X=True, candidate_estimator__estimator__fit_intercept=True, candidate_estimator__estimator__max_iter=1000, candidate_estimator__estimator__positive=False, candidate_estimator__estimator__precompute=False, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__selection=cyclic, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'alpha': [0.5, 1, 1.25]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   0.0s
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=GridSearchCV(estimator=Lasso(), n_jobs=-1, param_grid={'alpha': [0.5, 1, 1.25]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=Lasso(), candidate_estimator__estimator__alpha=1.0, candidate_estimator__estimator__copy_X=True, candidate_estimator__estimator__fit_intercept=True, candidate_estimator__estimator__max_iter=1000, candidate_estimator__estimator__positive=False, candidate_estimator__estimator__precompute=False, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__selection=cyclic, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'alpha': [0.5, 1, 1.25]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   0.0s
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=GridSearchCV(estimator=Lasso(), n_jobs=-1, param_grid={'alpha': [0.5, 1, 1.25]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=Lasso(), candidate_estimator__estimator__alpha=1.0, candidate_estimator__estimator__copy_X=True, candidate_estimator__estimator__fit_intercept=True, candidate_estimator__estimator__max_iter=1000, candidate_estimator__estimator__positive=False, candidate_estimator__estimator__precompute=False, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__selection=cyclic, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'alpha': [0.5, 1, 1.25]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   0.0s
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=GridSearchCV(estimator=Lasso(), n_jobs=-1, param_grid={'alpha': [0.5, 1, 1.25]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=Lasso(), candidate_estimator__estimator__alpha=1.0, candidate_estimator__estimator__copy_X=True, candidate_estimator__estimator__fit_intercept=True, candidate_estimator__estimator__max_iter=1000, candidate_estimator__estimator__positive=False, candidate_estimator__estimator__precompute=False, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__selection=cyclic, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'alpha': [0.5, 1, 1.25]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   0.0s
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=GridSearchCV(estimator=Lasso(), n_jobs=-1, param_grid={'alpha': [0.5, 1, 1.25]},
             verbose=2), candidate_estimator__cv=None, candidate_estimator__error_score=nan, candidate_estimator__estimator=Lasso(), candidate_estimator__estimator__alpha=1.0, candidate_estimator__estimator__copy_X=True, candidate_estimator__estimator__fit_intercept=True, candidate_estimator__estimator__max_iter=1000, candidate_estimator__estimator__positive=False, candidate_estimator__estimator__precompute=False, candidate_estimator__estimator__random_state=None, candidate_estimator__estimator__selection=cyclic, candidate_estimator__estimator__tol=0.0001, candidate_estimator__estimator__warm_start=False, candidate_estimator__n_jobs=-1, candidate_estimator__param_grid={'alpha': [0.5, 1, 1.25]}, candidate_estimator__pre_dispatch=2*n_jobs, candidate_estimator__refit=True, candidate_estimator__return_train_score=False, candidate_estimator__scoring=None, candidate_estimator__verbose=2; total time=   0.0s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.5s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=VotingRegressor(estimators=[('candidate_1',
                             GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                                          estimator=RandomForestRegressor(oob_score=True),
                                          param_grid={'max_features': ['sqrt',
                                                                       'log2',
                                                                       None],
                                                      'n_estimators': [100,
                                                                       250]},
                                          verbose=2)),
                            ('candidate_2',
                             GridSearchCV(estimator=GradientBoostingRegressor(),
                                          n_jobs=-1,
                                          param_grid={'learning_rate': [0.01,
                                                                        0.1,
                                                                        0.25],
                                                      'max_depth': [3, 6, 9]},
                                          verbose=2)),
                            ('candidate_3',
                             GridSearchCV(estimator=Lasso(), n_jobs=-1,
                                          param_grid={'alpha': [0.5, 1, 1.25]},
                                          verbose=2))]); total time=  26.3s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=VotingRegressor(estimators=[('candidate_1',
                             GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                                          estimator=RandomForestRegressor(oob_score=True),
                                          param_grid={'max_features': ['sqrt',
                                                                       'log2',
                                                                       None],
                                                      'n_estimators': [100,
                                                                       250]},
                                          verbose=2)),
                            ('candidate_2',
                             GridSearchCV(estimator=GradientBoostingRegressor(),
                                          n_jobs=-1,
                                          param_grid={'learning_rate': [0.01,
                                                                        0.1,
                                                                        0.25],
                                                      'max_depth': [3, 6, 9]},
                                          verbose=2)),
                            ('candidate_3',
                             GridSearchCV(estimator=Lasso(), n_jobs=-1,
                                          param_grid={'alpha': [0.5, 1, 1.25]},
                                          verbose=2))]); total time=  24.8s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=VotingRegressor(estimators=[('candidate_1',
                             GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                                          estimator=RandomForestRegressor(oob_score=True),
                                          param_grid={'max_features': ['sqrt',
                                                                       'log2',
                                                                       None],
                                                      'n_estimators': [100,
                                                                       250]},
                                          verbose=2)),
                            ('candidate_2',
                             GridSearchCV(estimator=GradientBoostingRegressor(),
                                          n_jobs=-1,
                                          param_grid={'learning_rate': [0.01,
                                                                        0.1,
                                                                        0.25],
                                                      'max_depth': [3, 6, 9]},
                                          verbose=2)),
                            ('candidate_3',
                             GridSearchCV(estimator=Lasso(), n_jobs=-1,
                                          param_grid={'alpha': [0.5, 1, 1.25]},
                                          verbose=2))]); total time=  24.9s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=VotingRegressor(estimators=[('candidate_1',
                             GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                                          estimator=RandomForestRegressor(oob_score=True),
                                          param_grid={'max_features': ['sqrt',
                                                                       'log2',
                                                                       None],
                                                      'n_estimators': [100,
                                                                       250]},
                                          verbose=2)),
                            ('candidate_2',
                             GridSearchCV(estimator=GradientBoostingRegressor(),
                                          n_jobs=-1,
                                          param_grid={'learning_rate': [0.01,
                                                                        0.1,
                                                                        0.25],
                                                      'max_depth': [3, 6, 9]},
                                          verbose=2)),
                            ('candidate_3',
                             GridSearchCV(estimator=Lasso(), n_jobs=-1,
                                          param_grid={'alpha': [0.5, 1, 1.25]},
                                          verbose=2))]); total time=  25.0s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=VotingRegressor(estimators=[('candidate_1',
                             GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                                          estimator=RandomForestRegressor(oob_score=True),
                                          param_grid={'max_features': ['sqrt',
                                                                       'log2',
                                                                       None],
                                                      'n_estimators': [100,
                                                                       250]},
                                          verbose=2)),
                            ('candidate_2',
                             GridSearchCV(estimator=GradientBoostingRegressor(),
                                          n_jobs=-1,
                                          param_grid={'learning_rate': [0.01,
                                                                        0.1,
                                                                        0.25],
                                                      'max_depth': [3, 6, 9]},
                                          verbose=2)),
                            ('candidate_3',
                             GridSearchCV(estimator=Lasso(), n_jobs=-1,
                                          param_grid={'alpha': [0.5, 1, 1.25]},
                                          verbose=2))]); total time=  24.8s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.5s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.5s
[CV] END ................max_features=None, n_estimators=100; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   1.8s
[CV] END ................max_features=None, n_estimators=250; total time=   1.2s
[CV] END ................max_features=None, n_estimators=250; total time=   0.9s
[CV] END ................max_features=None, n_estimators=250; total time=   1.0s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.9s
[CV] END ................max_features=None, n_estimators=250; total time=   0.9s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.9s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=VotingRegressor(estimators=[('candidate_1',
                             GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                                          estimator=RandomForestRegressor(oob_score=True),
                                          param_grid={'max_features': ['sqrt',
                                                                       'log2',
                                                                       None],
                                                      'n_estimators': [100,
                                                                       250]},
                                          verbose=2)),
                            ('candidate_2',
                             GridSearchCV(estimator=GradientBoostingRegressor(),
                                          n_jobs=-1,
                                          param_grid={'learning_rate': [0.01,
                                                                        0.1,
                                                                        0.25],
                                                      'max_depth': [3, 6, 9]},
                                          verbose=2)),
                            ('candidate_3',
                             GridSearchCV(estimator=Lasso(), n_jobs=-1,
                                          param_grid={'alpha': [0.5, 1, 1.25]},
                                          verbose=2))]); total time=  32.0s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.2s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.2s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.2s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.2s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.2s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.2s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.2s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.6s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.6s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.9s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.8s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.7s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.7s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.5s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.6s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.2s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.4s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.9s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.9s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=VotingRegressor(estimators=[('candidate_1',
                             GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                                          estimator=RandomForestRegressor(oob_score=True),
                                          param_grid={'max_features': ['sqrt',
                                                                       'log2',
                                                                       None],
                                                      'n_estimators': [100,
                                                                       250]},
                                          verbose=2)),
                            ('candidate_2',
                             GridSearchCV(estimator=GradientBoostingRegressor(),
                                          n_jobs=-1,
                                          param_grid={'learning_rate': [0.01,
                                                                        0.1,
                                                                        0.25],
                                                      'max_depth': [3, 6, 9]},
                                          verbose=2)),
                            ('candidate_3',
                             GridSearchCV(estimator=Lasso(), n_jobs=-1,
                                          param_grid={'alpha': [0.5, 1, 1.25]},
                                          verbose=2))]); total time=  32.1s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.4s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   1.1s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=VotingRegressor(estimators=[('candidate_1',
                             GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                                          estimator=RandomForestRegressor(oob_score=True),
                                          param_grid={'max_features': ['sqrt',
                                                                       'log2',
                                                                       None],
                                                      'n_estimators': [100,
                                                                       250]},
                                          verbose=2)),
                            ('candidate_2',
                             GridSearchCV(estimator=GradientBoostingRegressor(),
                                          n_jobs=-1,
                                          param_grid={'learning_rate': [0.01,
                                                                        0.1,
                                                                        0.25],
                                                      'max_depth': [3, 6, 9]},
                                          verbose=2)),
                            ('candidate_3',
                             GridSearchCV(estimator=Lasso(), n_jobs=-1,
                                          param_grid={'alpha': [0.5, 1, 1.25]},
                                          verbose=2))]); total time=  26.5s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.4s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.5s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=VotingRegressor(estimators=[('candidate_1',
                             GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                                          estimator=RandomForestRegressor(oob_score=True),
                                          param_grid={'max_features': ['sqrt',
                                                                       'log2',
                                                                       None],
                                                      'n_estimators': [100,
                                                                       250]},
                                          verbose=2)),
                            ('candidate_2',
                             GridSearchCV(estimator=GradientBoostingRegressor(),
                                          n_jobs=-1,
                                          param_grid={'learning_rate': [0.01,
                                                                        0.1,
                                                                        0.25],
                                                      'max_depth': [3, 6, 9]},
                                          verbose=2)),
                            ('candidate_3',
                             GridSearchCV(estimator=Lasso(), n_jobs=-1,
                                          param_grid={'alpha': [0.5, 1, 1.25]},
                                          verbose=2))]); total time=  25.5s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=100; total time=   0.2s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.7s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
Fitting 5 folds for each of 3 candidates, totalling 15 fits
[CV] END candidate_estimator=VotingRegressor(estimators=[('candidate_1',
                             GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                                          estimator=RandomForestRegressor(oob_score=True),
                                          param_grid={'max_features': ['sqrt',
                                                                       'log2',
                                                                       None],
                                                      'n_estimators': [100,
                                                                       250]},
                                          verbose=2)),
                            ('candidate_2',
                             GridSearchCV(estimator=GradientBoostingRegressor(),
                                          n_jobs=-1,
                                          param_grid={'learning_rate': [0.01,
                                                                        0.1,
                                                                        0.25],
                                                      'max_depth': [3, 6, 9]},
                                          verbose=2)),
                            ('candidate_3',
                             GridSearchCV(estimator=Lasso(), n_jobs=-1,
                                          param_grid={'alpha': [0.5, 1, 1.25]},
                                          verbose=2))]); total time=  25.6s
Fitting 10 folds for each of 6 candidates, totalling 60 fits
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=100; total time=   0.1s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=sqrt, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.0s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=100; total time=   0.1s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=log2, n_estimators=250; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=100; total time=   0.3s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
[CV] END ................max_features=None, n_estimators=250; total time=   0.8s
Fitting 5 folds for each of 9 candidates, totalling 45 fits
Fitting 5 folds for each of 3 candidates, totalling 15 fits
Benchmark updated!
New benchmark:
Pipeline(steps=[('candidate_estimator',
                 VotingRegressor(estimators=[('candidate_1',
                                              GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                                                           estimator=RandomForestRegressor(oob_score=True),
                                                           param_grid={'max_features': ['sqrt',
                                                                                        'log2',
                                                                                        None],
                                                                       'n_estimators': [100,
                                                                                        250]},
                                                           verbose=2)),
                                             ('candidate_2',
                                              GridSearchCV(estimator=GradientBoostingRegressor(),
                                                           n_jobs=-1,
                                                           param_grid={'learning_rate': [0.01,
                                                                                         0.1,
                                                                                         0.25],
                                                                       'max_depth': [3,
                                                                                     6,
                                                                                     9]},
                                                           verbose=2)),
                                             ('candidate_3',
                                              GridSearchCV(estimator=Lasso(),
                                                           n_jobs=-1,
                                                           param_grid={'alpha': [0.5,
                                                                                 1,
                                                                                 1.25]},
                                                           verbose=2))]))])

The output above clearly indicates that after evaluating the models - and their ensemble together with the existing benchmark - at least one of them was better than the current benchmark. Therefore, it will now be the new benchmark.

Code
y_pred = benchmark.predict(X)
pd.DataFrame({
    'y': y,
    'y_pred': y_pred
    }).plot.scatter(x='y', y='y_pred', grid=True)
<Axes: xlabel='y', ylabel='y_pred'>

Code
pd.DataFrame(y - y_pred).plot.hist(bins=30)
<Axes: ylabel='Frequency'>

Model documentation

After this process, we can now see how the model documentation was updated automatically:

Code
benchmark.model_documentation.show_json()
{'model_details': {'developer': 'Person or organisation developing the model',
  'datetime': '2024-02-27 15:46:05 ',
  'version': 'Model version',
  'type': 'Model type',
  'info': {'_estimator_type': 'regressor',
   'best_estimator_': Pipeline(steps=[('candidate_estimator',
                    VotingRegressor(estimators=[('candidate_1',
                                                 GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                                                              estimator=RandomForestRegressor(oob_score=True),
                                                              param_grid={'max_features': ['sqrt',
                                                                                           'log2',
                                                                                           None],
                                                                          'n_estimators': [100,
                                                                                           250]},
                                                              verbose=2)),
                                                ('candidate_2',
                                                 GridSearchCV(estimator=GradientBoostingRegressor(),
                                                              n_jobs=-1,
                                                              param_grid={'learning_rate': [0.01,
                                                                                            0.1,
                                                                                            0.25],
                                                                          'max_depth': [3,
                                                                                        6,
                                                                                        9]},
                                                              verbose=2)),
                                                ('candidate_3',
                                                 GridSearchCV(estimator=Lasso(),
                                                              n_jobs=-1,
                                                              param_grid={'alpha': [0.5,
                                                                                    1,
                                                                                    1.25]},
                                                              verbose=2))]))]),
   'best_index_': 3,
   'best_params_': {'candidate_estimator': VotingRegressor(estimators=[('candidate_1',
                                 GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                                              estimator=RandomForestRegressor(oob_score=True),
                                              param_grid={'max_features': ['sqrt',
                                                                           'log2',
                                                                           None],
                                                          'n_estimators': [100,
                                                                           250]},
                                              verbose=2)),
                                ('candidate_2',
                                 GridSearchCV(estimator=GradientBoostingRegressor(),
                                              n_jobs=-1,
                                              param_grid={'learning_rate': [0.01,
                                                                            0.1,
                                                                            0.25],
                                                          'max_depth': [3, 6, 9]},
                                              verbose=2)),
                                ('candidate_3',
                                 GridSearchCV(estimator=Lasso(), n_jobs=-1,
                                              param_grid={'alpha': [0.5, 1, 1.25]},
                                              verbose=2))])},
   'best_score_': -0.1334778603498503,
   'cv_results_': {'mean_fit_time': array([24.19174578,  2.54745026,  0.03593845, 26.83477316]),
    'std_fit_time': array([0.58680777, 0.49352541, 0.00715861, 2.69416102]),
    'mean_score_time': array([0.01762826, 0.00496593, 0.00156229, 0.01535699]),
    'std_score_time': array([0.00275886, 0.00700384, 0.00468686, 0.00620226]),
    'param_candidate_estimator': masked_array(data=[GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                                    estimator=RandomForestRegressor(oob_score=True),
                                    param_grid={'max_features': ['sqrt', 'log2', None],
                                                'n_estimators': [100, 250]},
                                    verbose=2)                                                                       ,
                       GridSearchCV(estimator=GradientBoostingRegressor(), n_jobs=-1,
                                    param_grid={'learning_rate': [0.01, 0.1, 0.25],
                                                'max_depth': [3, 6, 9]},
                                    verbose=2)                                       ,
                       GridSearchCV(estimator=Lasso(), n_jobs=-1, param_grid={'alpha': [0.5, 1, 1.25]},
                                    verbose=2)                                                         ,
                       VotingRegressor(estimators=[('candidate_1',
                                                    GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                                                                 estimator=RandomForestRegressor(oob_score=True),
                                                                 param_grid={'max_features': ['sqrt',
                                                                                              'log2',
                                                                                              None],
                                                                             'n_estimators': [100,
                                                                                              250]},
                                                                 verbose=2)),
                                                   ('candidate_2',
                                                    GridSearchCV(estimator=GradientBoostingRegressor(),
                                                                 n_jobs=-1,
                                                                 param_grid={'learning_rate': [0.01,
                                                                                               0.1,
                                                                                               0.25],
                                                                             'max_depth': [3, 6, 9]},
                                                                 verbose=2)),
                                                   ('candidate_3',
                                                    GridSearchCV(estimator=Lasso(), n_jobs=-1,
                                                                 param_grid={'alpha': [0.5, 1, 1.25]},
                                                                 verbose=2))])                                                                    ],
                 mask=[False, False, False, False],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__cv': masked_array(data=[ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                       None, None, --],
                 mask=[False, False, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__error_score': masked_array(data=[nan, nan, nan, --],
                 mask=[False, False, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator': masked_array(data=[RandomForestRegressor(oob_score=True),
                       GradientBoostingRegressor(), Lasso(), --],
                 mask=[False, False, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__bootstrap': masked_array(data=[True, --, --, --],
                 mask=[False,  True,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__ccp_alpha': masked_array(data=[0.0, 0.0, --, --],
                 mask=[False, False,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__criterion': masked_array(data=['squared_error', 'friedman_mse', --, --],
                 mask=[False, False,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__max_depth': masked_array(data=[None, 3, --, --],
                 mask=[False, False,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__max_features': masked_array(data=[1.0, None, --, --],
                 mask=[False, False,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__max_leaf_nodes': masked_array(data=[None, None, --, --],
                 mask=[False, False,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__max_samples': masked_array(data=[None, --, --, --],
                 mask=[False,  True,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__min_impurity_decrease': masked_array(data=[0.0, 0.0, --, --],
                 mask=[False, False,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__min_samples_leaf': masked_array(data=[1, 1, --, --],
                 mask=[False, False,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__min_samples_split': masked_array(data=[2, 2, --, --],
                 mask=[False, False,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__min_weight_fraction_leaf': masked_array(data=[0.0, 0.0, --, --],
                 mask=[False, False,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__n_estimators': masked_array(data=[100, 100, --, --],
                 mask=[False, False,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__n_jobs': masked_array(data=[None, --, --, --],
                 mask=[False,  True,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__oob_score': masked_array(data=[True, --, --, --],
                 mask=[False,  True,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__random_state': masked_array(data=[None, None, None, --],
                 mask=[False, False, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__verbose': masked_array(data=[0, 0, --, --],
                 mask=[False, False,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__warm_start': masked_array(data=[False, False, False, --],
                 mask=[False, False, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__n_jobs': masked_array(data=[None, -1, -1, --],
                 mask=[False, False, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__param_grid': masked_array(data=[{'n_estimators': [100, 250], 'max_features': ['sqrt', 'log2', None]},
                       {'learning_rate': [0.01, 0.1, 0.25], 'max_depth': [3, 6, 9]},
                       {'alpha': [0.5, 1, 1.25]}, --],
                 mask=[False, False, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__pre_dispatch': masked_array(data=['2*n_jobs', '2*n_jobs', '2*n_jobs', --],
                 mask=[False, False, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__refit': masked_array(data=[True, True, True, --],
                 mask=[False, False, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__return_train_score': masked_array(data=[False, False, False, --],
                 mask=[False, False, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__scoring': masked_array(data=[None, None, None, --],
                 mask=[False, False, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__verbose': masked_array(data=[2, 2, 2, --],
                 mask=[False, False, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__alpha': masked_array(data=[--, 0.9, 1.0, --],
                 mask=[ True, False, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__init': masked_array(data=[--, None, --, --],
                 mask=[ True, False,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__learning_rate': masked_array(data=[--, 0.1, --, --],
                 mask=[ True, False,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__loss': masked_array(data=[--, 'squared_error', --, --],
                 mask=[ True, False,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__n_iter_no_change': masked_array(data=[--, None, --, --],
                 mask=[ True, False,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__subsample': masked_array(data=[--, 1.0, --, --],
                 mask=[ True, False,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__tol': masked_array(data=[--, 0.0001, 0.0001, --],
                 mask=[ True, False, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__validation_fraction': masked_array(data=[--, 0.1, --, --],
                 mask=[ True, False,  True,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__copy_X': masked_array(data=[--, --, True, --],
                 mask=[ True,  True, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__fit_intercept': masked_array(data=[--, --, True, --],
                 mask=[ True,  True, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__max_iter': masked_array(data=[--, --, 1000, --],
                 mask=[ True,  True, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__positive': masked_array(data=[--, --, False, --],
                 mask=[ True,  True, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__precompute': masked_array(data=[--, --, False, --],
                 mask=[ True,  True, False,  True],
           fill_value='?',
                dtype=object),
    'param_candidate_estimator__estimator__selection': masked_array(data=[--, --, 'cyclic', --],
                 mask=[ True,  True, False,  True],
           fill_value='?',
                dtype=object),
    'params': [{'candidate_estimator': GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                   estimator=RandomForestRegressor(oob_score=True),
                   param_grid={'max_features': ['sqrt', 'log2', None],
                               'n_estimators': [100, 250]},
                   verbose=2),
      'candidate_estimator__cv': ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
      'candidate_estimator__error_score': nan,
      'candidate_estimator__estimator': RandomForestRegressor(oob_score=True),
      'candidate_estimator__estimator__bootstrap': True,
      'candidate_estimator__estimator__ccp_alpha': 0.0,
      'candidate_estimator__estimator__criterion': 'squared_error',
      'candidate_estimator__estimator__max_depth': None,
      'candidate_estimator__estimator__max_features': 1.0,
      'candidate_estimator__estimator__max_leaf_nodes': None,
      'candidate_estimator__estimator__max_samples': None,
      'candidate_estimator__estimator__min_impurity_decrease': 0.0,
      'candidate_estimator__estimator__min_samples_leaf': 1,
      'candidate_estimator__estimator__min_samples_split': 2,
      'candidate_estimator__estimator__min_weight_fraction_leaf': 0.0,
      'candidate_estimator__estimator__n_estimators': 100,
      'candidate_estimator__estimator__n_jobs': None,
      'candidate_estimator__estimator__oob_score': True,
      'candidate_estimator__estimator__random_state': None,
      'candidate_estimator__estimator__verbose': 0,
      'candidate_estimator__estimator__warm_start': False,
      'candidate_estimator__n_jobs': None,
      'candidate_estimator__param_grid': {'n_estimators': [100, 250],
       'max_features': ['sqrt', 'log2', None]},
      'candidate_estimator__pre_dispatch': '2*n_jobs',
      'candidate_estimator__refit': True,
      'candidate_estimator__return_train_score': False,
      'candidate_estimator__scoring': None,
      'candidate_estimator__verbose': 2},
     {'candidate_estimator': GridSearchCV(estimator=GradientBoostingRegressor(), n_jobs=-1,
                   param_grid={'learning_rate': [0.01, 0.1, 0.25],
                               'max_depth': [3, 6, 9]},
                   verbose=2),
      'candidate_estimator__cv': None,
      'candidate_estimator__error_score': nan,
      'candidate_estimator__estimator': GradientBoostingRegressor(),
      'candidate_estimator__estimator__alpha': 0.9,
      'candidate_estimator__estimator__ccp_alpha': 0.0,
      'candidate_estimator__estimator__criterion': 'friedman_mse',
      'candidate_estimator__estimator__init': None,
      'candidate_estimator__estimator__learning_rate': 0.1,
      'candidate_estimator__estimator__loss': 'squared_error',
      'candidate_estimator__estimator__max_depth': 3,
      'candidate_estimator__estimator__max_features': None,
      'candidate_estimator__estimator__max_leaf_nodes': None,
      'candidate_estimator__estimator__min_impurity_decrease': 0.0,
      'candidate_estimator__estimator__min_samples_leaf': 1,
      'candidate_estimator__estimator__min_samples_split': 2,
      'candidate_estimator__estimator__min_weight_fraction_leaf': 0.0,
      'candidate_estimator__estimator__n_estimators': 100,
      'candidate_estimator__estimator__n_iter_no_change': None,
      'candidate_estimator__estimator__random_state': None,
      'candidate_estimator__estimator__subsample': 1.0,
      'candidate_estimator__estimator__tol': 0.0001,
      'candidate_estimator__estimator__validation_fraction': 0.1,
      'candidate_estimator__estimator__verbose': 0,
      'candidate_estimator__estimator__warm_start': False,
      'candidate_estimator__n_jobs': -1,
      'candidate_estimator__param_grid': {'learning_rate': [0.01, 0.1, 0.25],
       'max_depth': [3, 6, 9]},
      'candidate_estimator__pre_dispatch': '2*n_jobs',
      'candidate_estimator__refit': True,
      'candidate_estimator__return_train_score': False,
      'candidate_estimator__scoring': None,
      'candidate_estimator__verbose': 2},
     {'candidate_estimator': GridSearchCV(estimator=Lasso(), n_jobs=-1, param_grid={'alpha': [0.5, 1, 1.25]},
                   verbose=2),
      'candidate_estimator__cv': None,
      'candidate_estimator__error_score': nan,
      'candidate_estimator__estimator': Lasso(),
      'candidate_estimator__estimator__alpha': 1.0,
      'candidate_estimator__estimator__copy_X': True,
      'candidate_estimator__estimator__fit_intercept': True,
      'candidate_estimator__estimator__max_iter': 1000,
      'candidate_estimator__estimator__positive': False,
      'candidate_estimator__estimator__precompute': False,
      'candidate_estimator__estimator__random_state': None,
      'candidate_estimator__estimator__selection': 'cyclic',
      'candidate_estimator__estimator__tol': 0.0001,
      'candidate_estimator__estimator__warm_start': False,
      'candidate_estimator__n_jobs': -1,
      'candidate_estimator__param_grid': {'alpha': [0.5, 1, 1.25]},
      'candidate_estimator__pre_dispatch': '2*n_jobs',
      'candidate_estimator__refit': True,
      'candidate_estimator__return_train_score': False,
      'candidate_estimator__scoring': None,
      'candidate_estimator__verbose': 2},
     {'candidate_estimator': VotingRegressor(estimators=[('candidate_1',
                                   GridSearchCV(cv=ShuffleSplit(n_splits=10, random_state=None, test_size=None, train_size=None),
                                                estimator=RandomForestRegressor(oob_score=True),
                                                param_grid={'max_features': ['sqrt',
                                                                             'log2',
                                                                             None],
                                                            'n_estimators': [100,
                                                                             250]},
                                                verbose=2)),
                                  ('candidate_2',
                                   GridSearchCV(estimator=GradientBoostingRegressor(),
                                                n_jobs=-1,
                                                param_grid={'learning_rate': [0.01,
                                                                              0.1,
                                                                              0.25],
                                                            'max_depth': [3, 6, 9]},
                                                verbose=2)),
                                  ('candidate_3',
                                   GridSearchCV(estimator=Lasso(), n_jobs=-1,
                                                param_grid={'alpha': [0.5, 1, 1.25]},
                                                verbose=2))])}],
    'split0_test_score': array([ 0.12804835,  0.53432154, -0.01859434,  0.38938267]),
    'split1_test_score': array([-0.49018834, -0.52804414, -0.42710614, -0.45019028]),
    'split2_test_score': array([ 0.15635849, -0.12348086, -0.21215091,  0.12855041]),
    'split3_test_score': array([ 0.04301892,  0.16881984, -1.04647184, -0.14303262]),
    'split4_test_score': array([-0.02858807, -0.36871869, -0.21840616, -0.16114508]),
    'split5_test_score': array([-0.08387596, -0.0008684 , -0.21229258,  0.06346885]),
    'split6_test_score': array([-1.09203873, -0.9983909 , -0.34868106, -0.67605684]),
    'split7_test_score': array([ 0.19154773,  0.04987919, -0.08547705,  0.06879344]),
    'split8_test_score': array([-0.46966373, -0.85738809, -0.6064379 , -0.4960824 ]),
    'split9_test_score': array([ 0.09348901, -0.28007549, -0.46966089, -0.05846674]),
    'mean_test_score': array([-0.15518923, -0.2403946 , -0.36452789, -0.13347786]),
    'std_test_score': array([0.38840165, 0.44485563, 0.28356188, 0.30893492]),
    'rank_test_score': array([2, 3, 4, 1])},
   'feature_names_in_': array(['Unnamed: 0', 'gdpsh465', 'bmp1l', 'freeop', 'freetar', 'h65',
          'hm65', 'hf65', 'p65', 'pm65', 'pf65', 's65', 'sm65', 'sf65',
          'fert65', 'mort65', 'lifee065', 'gpop1', 'fert1', 'mort1',
          'invsh41', 'geetot1', 'geerec1', 'gde1', 'govwb1', 'govsh41',
          'gvxdxe41', 'high65', 'highm65', 'highf65', 'highc65', 'highcm65',
          'highcf65', 'human65', 'humanm65', 'humanf65', 'hyr65', 'hyrm65',
          'hyrf65', 'no65', 'nom65', 'nof65', 'pinstab1', 'pop65',
          'worker65', 'pop1565', 'pop6565', 'sec65', 'secm65', 'secf65',
          'secc65', 'seccm65', 'seccf65', 'syr65', 'syrm65', 'syrf65',
          'teapri65', 'teasec65', 'ex1', 'im1', 'xr65', 'tot1'], dtype=object),
   'multimetric_': False,
   'n_features_in_': 62,
   'n_splits_': 10,
   'refit_time_': 27.246349573135376,
   'scorer_': <function sklearn.metrics._scorer._passthrough_scorer(estimator, *args, **kwargs)>},
  'paper': 'Paper or other resource for more information',
  'citation': 'Citation details',
  'license': 'License',
  'contact': 'Where to send questions or comments about the model'},
 'intended_use': {'primary_uses': 'Primary intended uses',
  'primary_users': 'Primary intended users',
  'out_of_scope': 'Out-of-scope use cases'},
 'factors': {'relevant': 'Relevant factors',
  'evaluation': 'Evaluation factors'},
 'metrics': {'performance_measures': 'Model performance measures',
  'thresholds': 'Decision thresholds',
  'variation_approaches': 'Variation approaches'},
 'evaluation_data': {'datasets': 'Datasets',
  'motivation': 'Motivation',
  'preprocessing': 'Preprocessing'},
 'training_data': {'training_data': 'Information on training data'},
 'quant_analyses': {'unitary': 'Unitary results',
  'intersectional': 'Intersectional results'},
 'ethical_considerations': {'sensitive_data': 'Does the model use any sensitive data (e.g., protected classes)?',
  'human_life': 'Is the model intended to inform decisions about matters central to human life or flourishing - e.g., health or safety? Or could it be used in such a way?',
  'mitigations': 'What risk mitigation strategies were used during model development?',
  'risks_and_harms': 'What risks may be present in model usage? Try to identify the potential recipients,likelihood, and magnitude of harms. If these cannot be determined, note that they were considered but remain unknown',
  'use_cases': 'Are there any known model use cases that are especially fraught?',
  'additional_information': 'If possible, this section should also include any additional ethical considerations that went into model development, for example, review by an external board, or testing with a specific community.'},
 'caveats_recommendations': {'caveats': 'For example, did the results suggest any further testing? Were there any relevant groups that were not represented in the evaluation dataset?',
  'recommendations': 'Are there additional recommendations for model use? What are the ideal characteristics of an evaluation dataset for this model?'}}

And as before, any remaining open questions can be viewed and answered using the same methods as above.

References

Barro, Robert J., and Jong-Wha Lee. 1994. “Sources of Economic Growth.” Carnegie-Rochester Conference Series on Public Policy 40: 1–46. https://doi.org/10.1016/0167-2231(94)90002-7.
Belloni, Alexandre, Victor Chernozhukov, and Christian Hansen. 2011. “Inference for High-Dimensional Sparse Econometric Models.” arXiv Preprint arXiv:1201.0220.
Giannone, Domenico, Michele Lenza, and Giorgio E Primiceri. 2021. “Economic Predictions with Big Data: The Illusion of Sparsity.” Econometrica 89 (5): 2409–37.