Skip to content

Instantly share code, notes, and snippets.

@albertz
Created November 20, 2020 16:35
Show Gist options
  • Save albertz/81878e34eddeb247da48e50ee043d648 to your computer and use it in GitHub Desktop.
Save albertz/81878e34eddeb247da48e50ee043d648 to your computer and use it in GitHub Desktop.
pytorch to returnn, parallel wavegan, demo debug output
.tmp_root: (hidden)
data: None -> None
melgan: <ModuleEntry <Sequential>> -> <TensorEntry name:? tensor:? returnn_data:'layer23_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|4] axes {0:0,2:1,1:2}>
data: None -> None
layer0: <ModuleEntry <ReflectionPad1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer0_output' [B,F|80,T|'spatial:1:melgan/layer0'] axes id>
layer1: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer1_output' [B,F|384,T|'time:var:extern_data:data'] axes id>
layer2: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer2_output' [B,F|384,T|'time:var:extern_data:data'] axes id>
layer3: <ModuleEntry <ConvTranspose1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer3_output' [B,T|'spatial:0:melgan/layer3',F|192] axes {0:0,2:1,1:2}>
layer4: <ModuleEntry <ResidualStack>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
data: None -> None
stack: <ModuleEntry <Sequential>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
data: None -> None
layer0: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer0_output' [B,T|'spatial:0:melgan/layer3',F|192] axes {0:0,2:1,1:2}>
layer1: <ModuleEntry <ReflectionPad1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer1_output' [B,T|'spatial:0:melgan/layer4/stack/layer1',F|192] axes {0:0,2:1,1:2}>
layer2: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer2_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
layer3: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer3_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
layer4: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
skip_layer: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'skip_layer_output' [B,T|'spatial:0:melgan/layer4/skip_layer',F|192] axes {0:0,2:1,1:2}>
add: <ModuleEntry <BinaryOperator>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
layer4_add_ReturnnReinterpretSameSizeAs: <ModuleEntry <ReturnnReinterpretSameSizeAs>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_add_ReturnnReinterpretSameSizeAs_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
layer5: <ModuleEntry <ResidualStack>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
data: None -> None
stack: <ModuleEntry <Sequential>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
data: None -> None
layer0: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer0_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
layer1: <ModuleEntry <ReflectionPad1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer1_output' [B,T|'spatial:0:melgan/layer5/stack/layer1',F|192] axes {0:0,2:1,1:2}>
layer2: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer2_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
layer3: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer3_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
layer4: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
skip_layer: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'skip_layer_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
add: <ModuleEntry <BinaryOperator>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
layer6: <ModuleEntry <ResidualStack>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
data: None -> None
stack: <ModuleEntry <Sequential>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
data: None -> None
layer0: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer0_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
layer1: <ModuleEntry <ReflectionPad1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer1_output' [B,T|'spatial:0:melgan/layer6/stack/layer1',F|192] axes {0:0,2:1,1:2}>
layer2: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer2_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
layer3: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer3_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
layer4: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
skip_layer: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'skip_layer_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
add: <ModuleEntry <BinaryOperator>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
layer7: <ModuleEntry <ResidualStack>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
data: None -> None
stack: <ModuleEntry <Sequential>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
data: None -> None
layer0: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer0_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
layer1: <ModuleEntry <ReflectionPad1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer1_output' [B,T|'spatial:0:melgan/layer7/stack/layer1',F|192] axes {0:0,2:1,1:2}>
layer2: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer2_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
layer3: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer3_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
layer4: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
skip_layer: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'skip_layer_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
add: <ModuleEntry <BinaryOperator>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
layer8: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer8_output' [B,T|'spatial:0:melgan/layer4/stack/layer2',F|192] axes {0:0,2:1,1:2}>
layer9: <ModuleEntry <ConvTranspose1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer9_output' [B,T|'spatial:0:melgan/layer9',F|96] axes {0:0,2:1,1:2}>
layer10: <ModuleEntry <ResidualStack>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
data: None -> None
stack: <ModuleEntry <Sequential>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
data: None -> None
layer0: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer0_output' [B,T|'spatial:0:melgan/layer9',F|96] axes {0:0,2:1,1:2}>
layer1: <ModuleEntry <ReflectionPad1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer1_output' [B,T|'spatial:0:melgan/layer10/stack/layer1',F|96] axes {0:0,2:1,1:2}>
layer2: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer2_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
layer3: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer3_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
layer4: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
skip_layer: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'skip_layer_output' [B,T|'spatial:0:melgan/layer10/skip_layer',F|96] axes {0:0,2:1,1:2}>
add: <ModuleEntry <BinaryOperator>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
layer10_add_ReturnnReinterpretSameSizeAs: <ModuleEntry <ReturnnReinterpretSameSizeAs>> -> <TensorEntry name:? tensor:? returnn_data:'layer10_add_ReturnnReinterpretSameSizeAs_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
layer11: <ModuleEntry <ResidualStack>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
data: None -> None
stack: <ModuleEntry <Sequential>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
data: None -> None
layer0: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer0_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
layer1: <ModuleEntry <ReflectionPad1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer1_output' [B,T|'spatial:0:melgan/layer11/stack/layer1',F|96] axes {0:0,2:1,1:2}>
layer2: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer2_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
layer3: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer3_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
layer4: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
skip_layer: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'skip_layer_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
add: <ModuleEntry <BinaryOperator>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
layer12: <ModuleEntry <ResidualStack>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
data: None -> None
stack: <ModuleEntry <Sequential>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
data: None -> None
layer0: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer0_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
layer1: <ModuleEntry <ReflectionPad1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer1_output' [B,T|'spatial:0:melgan/layer12/stack/layer1',F|96] axes {0:0,2:1,1:2}>
layer2: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer2_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
layer3: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer3_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
layer4: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
skip_layer: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'skip_layer_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
add: <ModuleEntry <BinaryOperator>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
layer13: <ModuleEntry <ResidualStack>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
data: None -> None
stack: <ModuleEntry <Sequential>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
data: None -> None
layer0: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer0_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
layer1: <ModuleEntry <ReflectionPad1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer1_output' [B,T|'spatial:0:melgan/layer13/stack/layer1',F|96] axes {0:0,2:1,1:2}>
layer2: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer2_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
layer3: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer3_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
layer4: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
skip_layer: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'skip_layer_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
add: <ModuleEntry <BinaryOperator>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
layer14: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer14_output' [B,T|'spatial:0:melgan/layer10/stack/layer2',F|96] axes {0:0,2:1,1:2}>
layer15: <ModuleEntry <ConvTranspose1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer15_output' [B,T|'spatial:0:melgan/layer15',F|48] axes {0:0,2:1,1:2}>
layer16: <ModuleEntry <ResidualStack>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
data: None -> None
stack: <ModuleEntry <Sequential>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
data: None -> None
layer0: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer0_output' [B,T|'spatial:0:melgan/layer15',F|48] axes {0:0,2:1,1:2}>
layer1: <ModuleEntry <ReflectionPad1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer1_output' [B,T|'spatial:0:melgan/layer16/stack/layer1',F|48] axes {0:0,2:1,1:2}>
layer2: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer2_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
layer3: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer3_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
layer4: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
skip_layer: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'skip_layer_output' [B,T|'spatial:0:melgan/layer16/skip_layer',F|48] axes {0:0,2:1,1:2}>
add: <ModuleEntry <BinaryOperator>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
layer16_add_ReturnnReinterpretSameSizeAs: <ModuleEntry <ReturnnReinterpretSameSizeAs>> -> <TensorEntry name:? tensor:? returnn_data:'layer16_add_ReturnnReinterpretSameSizeAs_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
layer17: <ModuleEntry <ResidualStack>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
data: None -> None
stack: <ModuleEntry <Sequential>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
data: None -> None
layer0: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer0_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
layer1: <ModuleEntry <ReflectionPad1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer1_output' [B,T|'spatial:0:melgan/layer17/stack/layer1',F|48] axes {0:0,2:1,1:2}>
layer2: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer2_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
layer3: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer3_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
layer4: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
skip_layer: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'skip_layer_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
add: <ModuleEntry <BinaryOperator>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
layer18: <ModuleEntry <ResidualStack>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
data: None -> None
stack: <ModuleEntry <Sequential>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
data: None -> None
layer0: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer0_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
layer1: <ModuleEntry <ReflectionPad1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer1_output' [B,T|'spatial:0:melgan/layer18/stack/layer1',F|48] axes {0:0,2:1,1:2}>
layer2: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer2_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
layer3: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer3_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
layer4: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
skip_layer: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'skip_layer_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
add: <ModuleEntry <BinaryOperator>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
layer19: <ModuleEntry <ResidualStack>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
data: None -> None
stack: <ModuleEntry <Sequential>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
data: None -> None
layer0: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer0_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
layer1: <ModuleEntry <ReflectionPad1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer1_output' [B,T|'spatial:0:melgan/layer19/stack/layer1',F|48] axes {0:0,2:1,1:2}>
layer2: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer2_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
layer3: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer3_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
layer4: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer4_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
skip_layer: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'skip_layer_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
add: <ModuleEntry <BinaryOperator>> -> <TensorEntry name:? tensor:? returnn_data:'add_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
layer20: <ModuleEntry <LeakyReLU>> -> <TensorEntry name:? tensor:? returnn_data:'layer20_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|48] axes {0:0,2:1,1:2}>
layer21: <ModuleEntry <ReflectionPad1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer21_output' [B,T|'spatial:0:melgan/layer21',F|48] axes {0:0,2:1,1:2}>
layer22: <ModuleEntry <Conv1d>> -> <TensorEntry name:? tensor:? returnn_data:'layer22_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|4] axes {0:0,2:1,1:2}>
layer23: <ModuleEntry <Tanh>> -> <TensorEntry name:? tensor:? returnn_data:'layer23_output' [B,T|'spatial:0:melgan/layer16/stack/layer2',F|4] axes {0:0,2:1,1:2}>
PQMF_updown_filter: <ModuleEntry <Constant>> -> <TensorEntry name:'updown_filter' tensor:(4,4,4) returnn_data:'const:updown_filter' [4,4,F|4] axes id>
PQMF_mul_unnamed_const: <ModuleEntry <Constant>> -> <TensorEntry name:'value' tensor:() returnn_data:'const:unnamed_const' [] axes id>
PQMF_mul: <ModuleEntry <BinaryOperator>> -> <TensorEntry name:? tensor:? returnn_data:'PQMF_mul_output' [4,4,F|4] axes id>
PQMF_pad_fn: <ModuleEntry <ConstantPad1d>> -> <TensorEntry name:? tensor:(1,4,80) returnn_data:'PQMF_pad_fn_output' [B,T|'spatial:0:PQMF_pad_fn',F|4] axes {0:0,2:1,1:2}>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment