Skip to content

Instantly share code, notes, and snippets.

View sagarhowal's full-sized avatar
💭
commit-ment issues

Sagar Howal sagarhowal

💭
commit-ment issues
  • Dublin, Ireland
View GitHub Profile
@sagarhowal
sagarhowal / bagging-pasting.py
Created December 6, 2017 10:48
Bagging Pasting for Ensemble Learning
from sklearn.ensemble import BaggingClassifier
from sklearn.tree import DecisionTreeClassifier
bag_clf = BaggingClassifier(
DecisionTreeClassifier(random_state=42), n_estimators=500,
max_samples=100, bootstrap=True, n_jobs=-1, random_state=42)
bag_clf.fit(X_train, y_train)
y_pred = bag_clf.predict(X_test)
@sagarhowal
sagarhowal / outofbag.py
Created December 6, 2017 10:53
Out-of-Bag in Ensemble Learning
bag_clf = BaggingClassifier(
DecisionTreeClassifier(random_state=42), n_estimators=500,
max_samples=100, bootstrap=True, n_jobs=-1, random_state=42,
oob_score = True)
@sagarhowal
sagarhowal / randompatches.py
Created December 6, 2017 10:55
Random Patches in Ensemble Learning
bootstrap_features = True, max_samples = 0.6
#max_samples to be less than 1.0 or discard variable
@sagarhowal
sagarhowal / randomsubspaces.py
Created December 6, 2017 10:56
Random Subspaces in Ensemble Learning
Bootstrap = True, bootstrap_features = True, max_features = 0.6
#max_samples less than 1.0 or discarded
@sagarhowal
sagarhowal / randomforests.py
Created December 6, 2017 10:57
Random Forests in Ensemble Learning
from sklearn.ensemble import RandomForestClassifier
rnd_clf = RandomForestClassifier(n_estimators=500, max_leaf_nodes=16, n_jobs=-1, random_state=42)
rnd_clf.fit(X_train, y_train)
y_pred_rf = rnd_clf.predict(X_test)
@sagarhowal
sagarhowal / adaboost.py
Created December 6, 2017 10:58
AdaBoost in Ensemble Learning
from sklearn.ensemble import AdaBoostClassifier
ada_clf = AdaBoostClassifier(
DecisionTreeClassifier(max_depth=1), n_estimators=200,
algorithm="SAMME.R", learning_rate=0.5, random_state=42)
ada_clf.fit(X_train, y_train)
@sagarhowal
sagarhowal / gradientboosting.py
Created December 6, 2017 10:59
Gradient Boosting in Ensemble Learning
from sklearn.ensemble import GradientBoostingRegressor
gbrt = GradientBoostingRegressor(max_depth=2, n_estimators=3, learning_rate=1.0, random_state=42)
gbrt.fit(X, y)
@sagarhowal
sagarhowal / xgboost.py
Created December 6, 2017 10:59
XGBoost in Ensemble Learning
from xgboost import XGBoostClassifier
xgb_clf = XGBClassifier()
xgb_clf.fit(X, y)
@sagarhowal
sagarhowal / np_read_csv.py
Created December 25, 2017 10:41
Reading CSV File with NumPy
import numpy as np
import pandas as pd
#Importing dataset
dataset = pd.read_csv('breast_cancer.csv')
#Check the first 5 rows of the dataset.
dataset.head(5)
@sagarhowal
sagarhowal / X_and_y.py
Created December 25, 2017 10:42
Separating X and y
#Seperating dependent and independent variables.
X = dataset.iloc[:, 2:32].values #Note: Exclude Last column with all NaN values.
y = dataset.iloc[:, 1].values