Skip to content

Instantly share code, notes, and snippets.

@f-rumblefish
f-rumblefish / library.csv
Last active April 5, 2020 04:52
Library
Library Descritption Usage in This Project
Librosa audio analysis
scikit-learn machine learning data split/Naive Bayes
imbalanced-learn ...
nlpaug data augmentation
@f-rumblefish
f-rumblefish / AudioModel.py
Last active April 5, 2020 02:48
Audio Model Selection
# algorithm 1 ------------------------------------------------------------------
print(" Naive Bayes ... ")
from sklearn import naive_bayes
classifier = naive_bayes.GaussianNB()
nb_model = classifier.fit(X, Y)
prediction = nb_model.predict(X_test)
print(" accuracy = ", accuracy_score(Y_test, prediction))
@f-rumblefish
f-rumblefish / AudioBalancing.py
Last active April 5, 2020 02:20
Audio Balancing
# import library
from imblearn.over_sampling import SMOTE
# define the environment variable
seed = 100
k = 1
# apply SMOTE to create the new dataset
sm = SMOTE(sampling_strategy='auto', k_neighbors=k, random_state=seed)
X_res, y_res = sm.fit_resample(pd_mfcc, pd_label)
@f-rumblefish
f-rumblefish / AudioAugmentation.py
Last active April 5, 2020 02:02
Audio Augmentation
# import library
import nlpaug
import nlpaug.augmenter.audio as naa
# loudness augmenter (where file_data is the output of librosa.load)
aug = naa.LoudnessAug(factor=(2, 5))
augmented_data = aug.augment(file_data)
# MFCC feature extraction for the new data ...
@f-rumblefish
f-rumblefish / mfcc_for_cat_dog.py
Created April 4, 2020 11:19
MFCC for Audio Cats and Dogs
# import library
import librosa
import numpy as np
# define the file name
wav_name = 'cat_1.wav'
# define the length of features
max_len = 20000
@f-rumblefish
f-rumblefish / performance.csv
Last active April 5, 2020 08:37
Audio Dog/Cat Classification
Model No SMOTE & No nlpaug SMOTE & No nlpaug SMOTE & nlpaug
Naive Bayes 73.1% 81.1% 81.9%
Random Forest 73.1% 87.8% 95.9%
Gradient Boosting 79.0% 89.5% 97.3%
XGBoost 88.3% 94.1% 97.3%
@f-rumblefish
f-rumblefish / template.csv
Last active February 23, 2020 03:51
course 101
topic data model software reference
classification
1 binary classification
2 multi-class classification MNIST
Fashion-MNIST
CIFAR-10/CIFAR-100
3 multi-label classifcation
@f-rumblefish
f-rumblefish / autoencoder.py
Last active March 3, 2021 07:10
Autoencoder
from keras.layers import Input, Dense, Conv2D, MaxPooling2D, UpSampling2D
from keras.models import Model
input_img = Input(shape=(28, 28, 1))
x = Conv2D(32, (3, 3), activation='relu', padding='same')(input_img)
x = MaxPooling2D((2, 2), padding='same')(x)
x = Conv2D(32, (3, 3), activation='relu', padding='same')(x)
x = MaxPooling2D((2, 2), padding='same')(x)
x = Conv2D(32, (3, 3), activation='relu', padding='same')(x)
@f-rumblefish
f-rumblefish / dataset.csv
Last active January 1, 2020 03:16
Dataset for Image Outlier Detecion
MNIST Fashion-MNIST Comment
Training Dataset 54000 0 data for training the autoencoder
Validation Dataset 6000 0 data for validating the autoencoder and defining the threshold
Testing Dataset 500 500 data for testing the solution
@f-rumblefish
f-rumblefish / Performance Summary.csv
Last active July 6, 2019 12:51
Multi-Input/Multi-Channel Performance
Approach Core Network Tail Network File Accuracy
Multi-Input 3 Conv2D/MaxPooling CNN Dense(1024/512/256) 101 65%
Multi-Input MobileNet ... 107
Multi-Channel 3 Conv2D/MaxPolling CNN ... 201 22%
Multi-Channel MobileNet GAP(0124)/Dense(256) 307 100%
Multi-Channel MobileNetV2 GAP(1024)/Dense(256) 308 2-->96%/10-->22%