Skip to content

Instantly share code, notes, and snippets.

@liorshk
Created April 22, 2020 15:24
Show Gist options
  • Star 25 You must be signed in to star a gist
  • Fork 7 You must be signed in to fork a gist
  • Save liorshk/9dfcb4a8e744fc15650cbd4c2b0955e5 to your computer and use it in GitHub Desktop.
Save liorshk/9dfcb4a8e744fc15650cbd4c2b0955e5 to your computer and use it in GitHub Desktop.
Create MLFlow runs with Sklearn Gridsearch object
def log_run(gridsearch: sklearn.GridSearchCV, experiment_name: str, model_name: str, run_index: int, conda_env, tags={}):
"""Logging of cross validation results to mlflow tracking server
Args:
experiment_name (str): experiment name
model_name (str): Name of the model
run_index (int): Index of the run (in Gridsearch)
conda_env (str): A dictionary that describes the conda environment (MLFlow Format)
tags (dict): Dictionary of extra data and tags (usually features)
"""
cv_results = gridsearch.cv_results_
with mlflow.start_run(run_name=str(run_index)) as run:
mlflow.log_param("folds", gridsearch.cv)
print("Logging parameters")
params = list(gridsearch.param_grid.keys())
for param in params:
mlflow.log_param(param, cv_results["param_%s" % param][run_index])
print("Logging metrics")
for score_name in [score for score in cv_results if "mean_test" in score]:
mlflow.log_metric(score_name, cv_results[score_name][run_index])
mlflow.log_metric(score_name.replace("mean","std"), cv_results[score_name.replace("mean","std")][run_index])
print("Logging model")
mlflow.sklearn.log_model(gridsearch.best_estimator_, model_name, conda_env=conda_env)
print("Logging CV results matrix")
tempdir = tempfile.TemporaryDirectory().name
os.mkdir(tempdir)
timestamp = datetime.now().isoformat().split(".")[0].replace(":", ".")
filename = "%s-%s-cv_results.csv" % (model_name, timestamp)
csv = os.path.join(tempdir, filename)
with warnings.catch_warnings():
warnings.simplefilter("ignore")
pd.DataFrame(cv_results).to_csv(csv, index=False)
mlflow.log_artifact(csv, "cv_results")
print("Logging extra data related to the experiment")
mlflow.set_tags(tags)
run_id = run.info.run_uuid
experiment_id = run.info.experiment_id
mlflow.end_run()
print(mlflow.get_artifact_uri())
print("runID: %s" % run_id)
def log_results(gridsearch: sklearn.GridSearchCV, experiment_name, model_name, tags={}, log_only_best=False):
"""Logging of cross validation results to mlflow tracking server
Args:
experiment_name (str): experiment name
model_name (str): Name of the model
tags (dict): Dictionary of extra tags
log_only_best (bool): Whether to log only the best model in the gridsearch or all the other models as well
"""
conda_env = {
'name': 'mlflow-env',
'channels': ['defaults'],
'dependencies': [
'python=3.7.0',
'scikit-learn>=0.21.3',
{'pip': ['xgboost==1.0.1']}
]
}
best = gridsearch.best_index_
mlflow.set_tracking_uri("http://kubernetes.docker.internal:5000")
mlflow.set_experiment(experiment_name)
if(log_only_best):
log_run(gridsearch, experiment_name, model_name, best, conda_env, tags)
else:
for i in range(len(gridsearch.cv_results_['params'])):
log_run(gridsearch, experiment_name, model_name, i, conda_env, tags)
@AllieUbisse
Copy link

Thanks for sharing

@krushev36
Copy link

Great Job. Thanks for sharing

@kunamneni117
Copy link

kunamneni117 commented Oct 22, 2020

I can see the code provided above creates some extra runs along with gridsearchcv estimators trained,this is because of after ending the current mlflow run,starting the mlflow run again starting the mlflow server to print the runid and artifact URL.

	def log_run(gridsearch: sklearn.GridSearchCV, experiment_name: str, model_name: str, run_index: int, conda_env, tags={}):
		"""Logging of cross validation results to mlflow tracking server
		
		Args:
			experiment_name (str): experiment name
			model_name (str): Name of the model
			run_index (int): Index of the run (in Gridsearch)
			conda_env (str): A dictionary that describes the conda environment (MLFlow Format)
			tags (dict): Dictionary of extra data and tags (usually features)
		"""
		
		cv_results = gridsearch.cv_results_
		with mlflow.start_run(run_name=str(run_index)) as run:  

			mlflow.log_param("folds", gridsearch.cv)

			print("Logging parameters")
			params = list(gridsearch.param_grid.keys())
			for param in params:
				mlflow.log_param(param, cv_results["param_%s" % param][run_index])

			print("Logging metrics")
			for score_name in [score for score in cv_results if "mean_test" in score]:
				mlflow.log_metric(score_name, cv_results[score_name][run_index])
				mlflow.log_metric(score_name.replace("mean","std"), cv_results[score_name.replace("mean","std")][run_index])

			print("Logging model")        
			mlflow.sklearn.log_model(gridsearch.best_estimator_, model_name, conda_env=conda_env)

			print("Logging CV results matrix")
			tempdir = tempfile.TemporaryDirectory().name
			os.mkdir(tempdir)
			timestamp = datetime.now().isoformat().split(".")[0].replace(":", ".")
			filename = "%s-%s-cv_results.csv" % (model_name, timestamp)
			csv = os.path.join(tempdir, filename)
			with warnings.catch_warnings():
				warnings.simplefilter("ignore")
				pd.DataFrame(cv_results).to_csv(csv, index=False)
			
			mlflow.log_artifact(csv, "cv_results") 

			print("Logging extra data related to the experiment")
			mlflow.set_tags(tags) 

			run_id = run.info.run_uuid
			experiment_id = run.info.experiment_id
			print(mlflow.get_artifact_uri())
			print("runID: %s" % run_id)
			mlflow.end_run()

	def log_results(gridsearch: sklearn.GridSearchCV, experiment_name, model_name, tags={}, log_only_best=False):
		"""Logging of cross validation results to mlflow tracking server
		
		Args:
			experiment_name (str): experiment name
			model_name (str): Name of the model
			tags (dict): Dictionary of extra tags
			log_only_best (bool): Whether to log only the best model in the gridsearch or all the other models as well
		"""
		conda_env = {
				'name': 'mlflow-env',
				'channels': ['defaults'],
				'dependencies': [
					'python=3.7.0',
					'scikit-learn>=0.21.3',
					{'pip': ['xgboost==1.0.1']}
				]
			}


		best = gridsearch.best_index_

		mlflow.set_tracking_uri("http://kubernetes.docker.internal:5000")
		mlflow.set_experiment(experiment_name)

		if(log_only_best):
			log_run(gridsearch, experiment_name, model_name, best, conda_env, tags)
		else:
			for i in range(len(gridsearch.cv_results_['params'])):
				log_run(gridsearch, experiment_name, model_name, i, conda_env, tags)

@doaa-altarawy
Copy link

Thanks. You could also add the imports at the top

import mlflow
import sklearn
import warnings
import tempfile
from sklearn.model_selection import GridSearchCV
import os
from datetime import datetime
import pandas as pd

@john2408
Copy link

Thank you very much!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment