I hereby claim:
- I am moritzkoerber on github.
- I am moritzkoerber (https://keybase.io/moritzkoerber) on keybase.
- I have a public key ASBagXNuNawc5COk1wSUH57zvRWiy4bM8o7ZeCxKWVv06Ao
To claim this, I am signing this object:
import pandas as pd | |
from sklearn.compose import ColumnTransformer | |
from sklearn.impute import SimpleImputer | |
from sklearn.linear_model import LogisticRegression | |
from sklearn.metrics import classification_report | |
from sklearn.model_selection import GridSearchCV, RepeatedStratifiedKFold | |
from sklearn.pipeline import Pipeline | |
from sklearn.preprocessing import OneHotEncoder, StandardScaler | |
titanic = pd.read_csv('./titanic.csv') |
import pandas as pd | |
from sklearn.linear_model import LogisticRegression | |
from sklearn.ensemble import RandomForestClassifier | |
from sklearn.model_selection import GridSearchCV, RepeatedStratifiedKFold | |
from sklearn.pipeline import Pipeline | |
from sklearn.compose import ColumnTransformer | |
from sklearn.preprocessing import OneHotEncoder, StandardScaler | |
from sklearn.metrics import f1_score, classification_report | |
from sklearn.impute import SimpleImputer | |
from sklearn.model_selection import train_test_split |
I hereby claim:
To claim this, I am signing this object:
import argparse | |
import pickle | |
import string | |
import sys | |
import nltk | |
import pandas as pd | |
from nltk.corpus import stopwords | |
from nltk.stem.wordnet import WordNetLemmatizer | |
from nltk.tokenize import word_tokenize |
import yaml | |
import great_expectations as ge | |
import os | |
from great_expectations.cli.datasource import sanitize_yaml_and_save_datasource | |
from great_expectations.core.batch import BatchRequest | |
from great_expectations.core.expectation_configuration import ExpectationConfiguration | |
from contextlib import suppress | |
project_dir = f"{os.getcwd()}/own_de_project/great_expectations" |