Skip to content

Instantly share code, notes, and snippets.

@tikhonova
Created April 16, 2023 22:19
Show Gist options
  • Star 0 You must be signed in to star a gist
  • Fork 0 You must be signed in to fork a gist
  • Save tikhonova/71a80b85577ef3dbc6467013ccf96f8c to your computer and use it in GitHub Desktop.
Save tikhonova/71a80b85577ef3dbc6467013ccf96f8c to your computer and use it in GitHub Desktop.
# Attention parameters
attention_rnn_dim=1024, # sets the number of units in the RNN
attention_dim=128, # sets the number of units in the attention mechanism
# These two values are relatively large and may require a significant amount of GPU memory during training and inference.
# Location Layer parameters
attention_location_n_filters=32, # sets the number of filters in the CNN
attention_location_kernel_size=31, # sets the size of the filters
# This means that the CNN has 32 filters and each filter has a kernel size of 31.
# via https://github.com/tikhonova/what_would_alan_watts_say/blob/master/speech_synthesis/tacotron2_waveglow_model_files/tacotron_hparams.py
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment