Skip to content

Instantly share code, notes, and snippets.

@oscar-defelice
Created November 4, 2020 08:26
Show Gist options
  • Save oscar-defelice/ad6e2c946b5ae71e3eb88639f352cbd1 to your computer and use it in GitHub Desktop.
Save oscar-defelice/ad6e2c946b5ae71e3eb88639f352cbd1 to your computer and use it in GitHub Desktop.
from keras.preprocessing.text import Tokenizer
from keras.preprocessing.sequence import pad_sequences
tokeniser = Tokenizer()
tokeniser.fit_on_texts(train_data['Text'])
tokenised_text = tokeniser.texts_to_sequences(train_data['Text'])
tokenised_text = pad_sequences(tokenised_text, maxlen=max_len)
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment