Created
August 18, 2019 16:26
-
-
Save sandys/d0acf13976bae81df253d0a09436cb2b to your computer and use it in GitHub Desktop.
Gridsearchcv with k-fold cross validation and early stopping
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
# | |
... import xgboost.sklearn as xgb | |
... from sklearn.model_selection import GridSearchCV | |
... from sklearn.model_selection import TimeSeriesSplit | |
... | |
... cv = 2 | |
... | |
... trainX= [[1], [2], [3], [4], [5]] | |
... trainY = [1, 2, 1, 2, 1] | |
... | |
... # these are the evaluation sets | |
... testX = trainX | |
... testY = trainY | |
... | |
... paramGrid = {"subsample" : [0.5, 0.8]} | |
... | |
... fit_params={"early_stopping_rounds":42, | |
... "eval_metric" : "mae", | |
... "eval_set" : [[testX, testY]]} | |
... | |
... model = xgb.XGBRegressor() | |
... | |
... from sklearn.model_selection import StratifiedKFold | |
... skf = StratifiedKFold(n_splits=cv, shuffle = True, random_state = 999) | |
... gridsearch = GridSearchCV(model, paramGrid, verbose=1, | |
... cv=skf.split(trainX, trainY)) | |
... | |
... g = gridsearch.fit(trainX, trainY, **fit_params) | |
... y_true, y_pred = testY, gridsearch.predict(testX) | |
... from sklearn.metrics import r2_score | |
... print(r2_score(y_true, y_pred)) | |
... print(g.cv_results_) |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
This would be biased as the early stopping is performed on the test set. The correct approach is to split each train fold into train and validation and use validation for early stopping. This cannot be achieved using
GridSearchCV
, instead the grid search must be performed manually, for example: https://xgboosting.com/xgboost-early-stopping-with-grid-search/