Skip to content

Instantly share code, notes, and snippets.

@manish-kumar-garg
Last active December 30, 2019 07:46
Show Gist options
  • Save manish-kumar-garg/7b017cfed5827a9083d5e16f7df43fda to your computer and use it in GitHub Desktop.
Save manish-kumar-garg/7b017cfed5827a9083d5e16f7df43fda to your computer and use it in GitHub Desktop.
asr local attention error
warning: unable to access '/home/ubuntu/.gitconfig': Is a directory
warning: unable to access '/home/ubuntu/.gitconfig': Is a directory
fatal: unknown error occurred while reading the configuration files
RETURNN starting up, version unknown(git exception: CalledProcessError(128, ('git', 'show', '-s', '--format=%ci', 'HEAD'))), date/time 2019-12-30-07-45-03 (UTC+0000), pid 61826, cwd /home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention, Python /home/ubuntu/tf1.13/bin/python3
RETURNN command line options: ['local_win05.config']
Hostname: ip-10-1-21-241
/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/framework/dtypes.py:526: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
_np_qint8 = np.dtype([("qint8", np.int8, 1)])
/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/framework/dtypes.py:527: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
_np_quint8 = np.dtype([("quint8", np.uint8, 1)])
/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/framework/dtypes.py:528: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
_np_qint16 = np.dtype([("qint16", np.int16, 1)])
/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/framework/dtypes.py:529: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
_np_quint16 = np.dtype([("quint16", np.uint16, 1)])
/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/framework/dtypes.py:530: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
_np_qint32 = np.dtype([("qint32", np.int32, 1)])
/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/framework/dtypes.py:535: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
np_resource = np.dtype([("resource", np.ubyte, 1)])
TensorFlow: 1.13.1 (b'v1.13.1-0-g6612da8951') (<site-package> in /home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow)
Setup TF inter and intra global thread pools, num_threads None, session opts {'log_device_placement': False, 'device_count': {'GPU': 0}}.
2019-12-30 07:45:04.884803: I tensorflow/core/platform/cpu_feature_guard.cc:141] Your CPU supports instructions that this TensorFlow binary was not compiled to use: AVX2 FMA
2019-12-30 07:45:05.081329: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:998] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2019-12-30 07:45:05.083163: I tensorflow/compiler/xla/service/service.cc:150] XLA service 0x55ef8d5e1630 executing computations on platform CUDA. Devices:
2019-12-30 07:45:05.083215: I tensorflow/compiler/xla/service/service.cc:158] StreamExecutor device (0): Tesla K80, Compute Capability 3.7
2019-12-30 07:45:05.106530: I tensorflow/core/platform/profile_utils/cpu_utils.cc:94] CPU Frequency: 2300090000 Hz
2019-12-30 07:45:05.109077: I tensorflow/compiler/xla/service/service.cc:150] XLA service 0x55ef8dca4bd0 executing computations on platform Host. Devices:
2019-12-30 07:45:05.109120: I tensorflow/compiler/xla/service/service.cc:158] StreamExecutor device (0): <undefined>, <undefined>
2019-12-30 07:45:05.109246: I tensorflow/core/common_runtime/gpu/gpu_device.cc:984] Device interconnect StreamExecutor with strength 1 edge matrix:
2019-12-30 07:45:05.109277: I tensorflow/core/common_runtime/gpu/gpu_device.cc:990]
CUDA_VISIBLE_DEVICES is set to '1'.
Collecting TensorFlow device list...
2019-12-30 07:45:05.112615: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1433] Found device 0 with properties:
name: Tesla K80 major: 3 minor: 7 memoryClockRate(GHz): 0.8235
pciBusID: 0000:00:18.0
totalMemory: 11.17GiB freeMemory: 11.11GiB
2019-12-30 07:45:05.112650: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1512] Adding visible gpu devices: 0
2019-12-30 07:45:05.114885: I tensorflow/core/common_runtime/gpu/gpu_device.cc:984] Device interconnect StreamExecutor with strength 1 edge matrix:
2019-12-30 07:45:05.114911: I tensorflow/core/common_runtime/gpu/gpu_device.cc:990] 0
2019-12-30 07:45:05.114928: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1003] 0: N
2019-12-30 07:45:05.115083: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1115] Created TensorFlow device (/device:GPU:0 with 10805 MB memory) -> physical GPU (device: 0, name: Tesla K80, pci bus id: 0000:00:18.0, compute capability: 3.7)
Local devices available to TensorFlow:
1/4: name: "/device:CPU:0"
device_type: "CPU"
memory_limit: 268435456
locality {
}
incarnation: 8403144321325319463
2/4: name: "/device:XLA_GPU:0"
device_type: "XLA_GPU"
memory_limit: 17179869184
locality {
}
incarnation: 10706382646999000709
physical_device_desc: "device: XLA_GPU device"
3/4: name: "/device:XLA_CPU:0"
device_type: "XLA_CPU"
memory_limit: 17179869184
locality {
}
incarnation: 1490385100323021608
physical_device_desc: "device: XLA_CPU device"
4/4: name: "/device:GPU:0"
device_type: "GPU"
memory_limit: 11330115994
locality {
bus_id: 1
links {
}
}
incarnation: 9325274229127908753
physical_device_desc: "device: 0, name: Tesla K80, pci bus id: 0000:00:18.0, compute capability: 3.7"
Using gpu device 1: Tesla K80
use local file: data/dataset/dev-clean.zip
use local file: data/dataset/dev-other.zip
use local file: data/dataset/train-clean-100.zip
use local file: data/dataset/train-clean-360.zip
use local file: data/dataset/train-other-500.zip
<LibriSpeechCorpus 'train' epoch=1>, epoch 1. Old mean seq len (transcription) is 183.267376, new is 63.708029, requested max is 75.000000. Old num seqs is 6575, new num seqs is 822.
<LibriSpeechCorpus 'train' epoch=1>, epoch 1. Old num seqs 14063, new num seqs 822.
<LibriSpeechCorpus 'train' epoch=1>, epoch 1. Old mean seq len (transcription) is 183.267376, new is 63.708029, requested max is 75.000000. Old num seqs is 6575, new num seqs is 822.
<LibriSpeechCorpus 'train' epoch=1>, epoch 1. Old num seqs 14063, new num seqs 822.
Train data:
input: 40 x 1
output: {'classes': [10025, 1], 'raw': {'dtype': 'string', 'shape': ()}, 'data': [40, 2]}
LibriSpeechCorpus, sequences: 822, frames: unknown
Dev data:
LibriSpeechCorpus, sequences: 3000, frames: unknown
Learning-rate-control: file data/exp-local_win05/train-scores.data does not exist yet
Update config key 'max_seq_length' for epoch 1: {'classes': 75} -> {'classes': 60}
Setup tf.Session with options {'log_device_placement': False, 'device_count': {'GPU': 1}} ...
2019-12-30 07:45:11.119403: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1512] Adding visible gpu devices: 0
2019-12-30 07:45:11.119463: I tensorflow/core/common_runtime/gpu/gpu_device.cc:984] Device interconnect StreamExecutor with strength 1 edge matrix:
2019-12-30 07:45:11.119479: I tensorflow/core/common_runtime/gpu/gpu_device.cc:990] 0
2019-12-30 07:45:11.119493: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1003] 0: N
2019-12-30 07:45:11.119659: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1115] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 10805 MB memory) -> physical GPU (device: 0, name: Tesla K80, pci bus id: 0000:00:18.0, compute capability: 3.7)
layer root/'data' output: Data(name='data', shape=(None, 40), batch_shape_meta=[B,T|'time:var:extern_data:data',F|40])
layer root/'source' output: Data(name='source_output', shape=(None, 40), batch_shape_meta=[B,T|'time:var:extern_data:data',F|40])
layer root/'lstm0_fw' output: Data(name='lstm0_fw_output', shape=(None, 1024), batch_dim_axis=1, batch_shape_meta=[T|'time:var:extern_data:data',B,F|1024])
layer root/'lstm0_bw' output: Data(name='lstm0_bw_output', shape=(None, 1024), batch_dim_axis=1, batch_shape_meta=[T|'time:var:extern_data:data',B,F|1024])
layer root/'lstm0_pool' output: Data(name='lstm0_pool_output', shape=(None, 2048), batch_shape_meta=[B,T|?,F|2048])
layer root/'lstm5_fw' output: Data(name='lstm5_fw_output', shape=(None, 1024), batch_dim_axis=1, batch_shape_meta=[T|'spatial:0:lstm0_pool',B,F|1024])
layer root/'lstm5_bw' output: Data(name='lstm5_bw_output', shape=(None, 1024), batch_dim_axis=1, batch_shape_meta=[T|'spatial:0:lstm0_pool',B,F|1024])
layer root/'encoder' output: Data(name='encoder_output', shape=(None, 2048), batch_dim_axis=1, batch_shape_meta=[T|'spatial:0:lstm0_pool',B,F|2048])
layer root/'ctc' output: Data(name='ctc_output', shape=(None, 10026), batch_dim_axis=1, batch_shape_meta=[T|'spatial:0:lstm0_pool',B,F|10026])
layer root/'enc_ctx' output: Data(name='enc_ctx_output', shape=(None, 1024), batch_dim_axis=1, batch_shape_meta=[T|'spatial:0:lstm0_pool',B,F|1024])
layer root/'inv_fertility' output: Data(name='inv_fertility_output', shape=(None, 1), batch_dim_axis=1, batch_shape_meta=[T|'spatial:0:lstm0_pool',B,F|1])
layer root/'enc_value' output: Data(name='enc_value_output', shape=(None, 1, 2048), batch_dim_axis=1, batch_shape_meta=[T|'spatial:0:lstm0_pool',B,1,F|2048])
layer root/'output' output: Data(name='output_output', shape=(None,), dtype='int32', sparse=True, dim=10025, batch_dim_axis=1, batch_shape_meta=[T|'time:var:extern_data:classes',B])
Rec layer 'output' (search False, train 'globals/train_flag:0') sub net:
Input layers moved out of loop: (#: 2)
output
target_embed
Output layers moved out of loop: (#: 3)
output_prob
readout
readout_in
Layers in loop: (#: 12)
s
att
att0
att_weights
energy_reinterpreted
energy
energy_tanh
energy_in
weight_feedback
accum_att_weights
s_transformed
p_t_in
Unused layers: (#: 2)
end
p_t
layer root/output:rec-subnet-input/'output' output: Data(name='output_output', shape=(None,), dtype='int32', sparse=True, dim=10025, batch_shape_meta=[B,T|'time:var:extern_data:classes'])
layer root/output:rec-subnet-input/'target_embed' output: Data(name='target_embed_output', shape=(None, 621), batch_shape_meta=[B,T|'time:var:extern_data:classes',F|621])
layer root/output:rec-subnet/'weight_feedback' output: Data(name='weight_feedback_output', shape=(None, 1024), batch_dim_axis=1, batch_shape_meta=[T|?,B,F|1024])
layer root/output:rec-subnet/'prev:target_embed' output: Data(name='target_embed_output', shape=(621,), time_dim_axis=None, batch_shape_meta=[B,F|621])
layer root/output:rec-subnet/'s' output: Data(name='s_output', shape=(1000,), time_dim_axis=None, batch_shape_meta=[B,F|1000])
layer root/output:rec-subnet/'s_transformed' output: Data(name='s_transformed_output', shape=(1024,), time_dim_axis=None, batch_shape_meta=[B,F|1024])
layer root/output:rec-subnet/'energy_in' output: Data(name='energy_in_output', shape=(None, 1024), batch_dim_axis=1, batch_shape_meta=[T|'spatial:0:lstm0_pool',B,F|1024])
layer root/output:rec-subnet/'energy_tanh' output: Data(name='energy_tanh_output', shape=(None, 1024), batch_dim_axis=1, batch_shape_meta=[T|'spatial:0:lstm0_pool',B,F|1024])
layer root/output:rec-subnet/'energy' output: Data(name='energy_output', shape=(None, 1), batch_dim_axis=1, batch_shape_meta=[T|'spatial:0:lstm0_pool',B,F|1])
layer root/output:rec-subnet/'energy_reinterpreted' output: Data(name='energy_reinterpreted_output', shape=(None, 1), batch_shape_meta=[B,T|'spatial:0:lstm0_pool',F|1])
layer root/output:rec-subnet/'p_t_in' output: Data(name='p_t_in_output', shape=(1,), dtype='int32', sparse=True, dim=None, time_dim_axis=None, feature_dim_axis=None, batch_shape_meta=[B,1])
layer root/output:rec-subnet/'att_weights' output: Data(name='att_weights_output', shape=(1, None), time_dim_axis=2, feature_dim_axis=1, batch_shape_meta=[B,F|1,T|'spatial:0:lstm0_pool'])
layer root/output:rec-subnet/'att0' output: Data(name='att0_output', shape=(1, 2048), time_dim_axis=None, batch_shape_meta=[B,1,F|2048])
layer root/output:rec-subnet/'att' output: Data(name='att_output', shape=(2048,), time_dim_axis=None, batch_shape_meta=[B,F|2048])
layer root/output:rec-subnet/'accum_att_weights' output: Data(name='accum_att_weights_output', shape=(None, 1), batch_dim_axis=1, batch_shape_meta=[T|'spatial:0:lstm0_pool',B,F|1])
layer root/output:rec-subnet-output/'s' output: Data(name='s_output', shape=(None, 1000), batch_dim_axis=1, batch_shape_meta=[T|'time:var:extern_data:classes',B,F|1000])
layer root/output:rec-subnet-output/'prev:target_embed' output: Data(name='target_embed_output', shape=(None, 621), batch_dim_axis=1, batch_shape_meta=[T|'time:var:extern_data:classes',B,F|621])
layer root/output:rec-subnet-output/'att' output: Data(name='att_output', shape=(None, 2048), batch_dim_axis=1, batch_shape_meta=[T|'time:var:extern_data:classes',B,F|2048])
layer root/output:rec-subnet-output/'readout_in' output: Data(name='readout_in_output', shape=(None, 1000), batch_dim_axis=1, batch_shape_meta=[T|'time:var:extern_data:classes',B,F|1000])
layer root/output:rec-subnet-output/'readout' output: Data(name='readout_output', shape=(None, 500), batch_dim_axis=1, batch_shape_meta=[T|'time:var:extern_data:classes',B,F|500])
layer root/output:rec-subnet-output/'output_prob' output: Data(name='output_prob_output', shape=(None, 10025), batch_dim_axis=1, batch_shape_meta=[T|'time:var:extern_data:classes',B,F|10025])
layer root/'decision' output: Data(name='output_output', shape=(None,), dtype='int32', sparse=True, dim=10025, batch_dim_axis=1, batch_shape_meta=[T|'time:var:extern_data:classes',B])
Network layer topology:
extern data: classes: Data(shape=(None,), dtype='int32', sparse=True, dim=10025, available_for_inference=False, batch_shape_meta=[B,T|'time:var:extern_data:classes']), data: Data(shape=(None, 40), batch_shape_meta=[B,T|'time:var:extern_data:data',F|40])
used data keys: ['classes', 'data']
layers:
layer softmax 'ctc' #: 10026
layer source 'data' #: 40
layer decide 'decision' #: 10025
layer linear 'enc_ctx' #: 1024
layer split_dims 'enc_value' #: 2048
layer copy 'encoder' #: 2048
layer linear 'inv_fertility' #: 1
layer rec 'lstm0_bw' #: 1024
layer rec 'lstm0_fw' #: 1024
layer pool 'lstm0_pool' #: 2048
layer rec 'lstm5_bw' #: 1024
layer rec 'lstm5_fw' #: 1024
layer rec 'output' #: 10025
layer eval 'source' #: 40
net params #: 87166092
net trainable params: [<tf.Variable 'ctc/W:0' shape=(2048, 10026) dtype=float32_ref>, <tf.Variable 'ctc/b:0' shape=(10026,) dtype=float32_ref>, <tf.Variable 'enc_ctx/W:0' shape=(2048, 1024) dtype=float32_ref>, <tf.Variable 'enc_ctx/b:0' shape=(1024,) dtype=float32_ref>, <tf.Variable 'inv_fertility/W:0' shape=(2048, 1) dtype=float32_ref>, <tf.Variable 'lstm0_bw/rec/W:0' shape=(40, 4096) dtype=float32_ref>, <tf.Variable 'lstm0_bw/rec/W_re:0' shape=(1024, 4096) dtype=float32_ref>, <tf.Variable 'lstm0_bw/rec/b:0' shape=(4096,) dtype=float32_ref>, <tf.Variable 'lstm0_fw/rec/W:0' shape=(40, 4096) dtype=float32_ref>, <tf.Variable 'lstm0_fw/rec/W_re:0' shape=(1024, 4096) dtype=float32_ref>, <tf.Variable 'lstm0_fw/rec/b:0' shape=(4096,) dtype=float32_ref>, <tf.Variable 'lstm5_bw/rec/W:0' shape=(2048, 4096) dtype=float32_ref>, <tf.Variable 'lstm5_bw/rec/W_re:0' shape=(1024, 4096) dtype=float32_ref>, <tf.Variable 'lstm5_bw/rec/b:0' shape=(4096,) dtype=float32_ref>, <tf.Variable 'lstm5_fw/rec/W:0' shape=(2048, 4096) dtype=float32_ref>, <tf.Variable 'lstm5_fw/rec/W_re:0' shape=(1024, 4096) dtype=float32_ref>, <tf.Variable 'lstm5_fw/rec/b:0' shape=(4096,) dtype=float32_ref>, <tf.Variable 'output/rec/energy/W:0' shape=(1024, 1) dtype=float32_ref>, <tf.Variable 'output/rec/output_prob/W:0' shape=(500, 10025) dtype=float32_ref>, <tf.Variable 'output/rec/output_prob/b:0' shape=(10025,) dtype=float32_ref>, <tf.Variable 'output/rec/readout_in/W:0' shape=(3669, 1000) dtype=float32_ref>, <tf.Variable 'output/rec/readout_in/b:0' shape=(1000,) dtype=float32_ref>, <tf.Variable 'output/rec/s/rec/lstm_cell/bias:0' shape=(4000,) dtype=float32_ref>, <tf.Variable 'output/rec/s/rec/lstm_cell/kernel:0' shape=(3669, 4000) dtype=float32_ref>, <tf.Variable 'output/rec/s_transformed/W:0' shape=(1000, 1024) dtype=float32_ref>, <tf.Variable 'output/rec/target_embed/W:0' shape=(10025, 621) dtype=float32_ref>, <tf.Variable 'output/rec/weight_feedback/W:0' shape=(1, 1024) dtype=float32_ref>]
start training at epoch 1 and step 0
using batch size: 20000, max seqs: 200
learning rate control: NewbobMultiEpoch(num_epochs=2, update_interval=1, relative_error_threshold=-0.01, learning_rate_decay_factor=0.9, learning_rate_growth_factor=1.0), epoch data: , error key: None
pretrain: Pretrain construction algo <function custom_construction_algo at 0x7fb3921b4ae8>, number of pretrain epochs: 45 (repetitions: [10, 5, 5, 5, 5, 5, 5, 5])
<LibriSpeechCorpus 'train' epoch=1>, epoch 1. Old mean seq len (transcription) is 183.267376, new is 63.708029, requested max is 75.000000. Old num seqs is 6575, new num seqs is 822.
<LibriSpeechCorpus 'train' epoch=1>, epoch 1. Old num seqs 14063, new num seqs 822.
Update config key 'max_seq_length' for epoch 1: {'classes': 75} -> {'classes': 60}
start pretrain epoch 1 with learning rate 1e-05 ...
TF: log_dir: data/exp-local_win05/train-2019-12-30-07-45-03
Create Adam optimizer.
/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/ops/gradients_impl.py:110: UserWarning: Converting sparse IndexedSlices to a dense Tensor of unknown shape. This may consume a large amount of memory.
"Converting sparse IndexedSlices to a dense Tensor of unknown shape. "
Initialize optimizer (default) with slots ['m', 'v'].
These additional variable were created by the optimizer: [<tf.Variable 'optimize/beta1_power:0' shape=() dtype=float32_ref>, <tf.Variable 'optimize/beta2_power:0' shape=() dtype=float32_ref>].
2019-12-30 07:45:27.482527: I tensorflow/stream_executor/dso_loader.cc:152] successfully opened CUDA library libcublas.so.10.0 locally
TensorFlow exception: Incompatible shapes: [14,1,45] vs. [45,1,45]
[[node output/rec/att_weights/LogicalAnd_1 (defined at /home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkLayer.py:3159) ]]
Caused by op 'output/rec/att_weights/LogicalAnd_1', defined at:
File "./returnn/rnn.py", line 654, in <module>
main(sys.argv)
File "./returnn/rnn.py", line 642, in main
execute_main_task()
File "./returnn/rnn.py", line 451, in execute_main_task
engine.init_train_from_config(config, train_data, dev_data, eval_data)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFEngine.py", line 891, in init_train_from_config
self.init_network_from_config(config)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFEngine.py", line 934, in init_network_from_config
self._init_network(net_desc=net_dict, epoch=self.epoch)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFEngine.py", line 1081, in _init_network
net_dict=net_desc)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFEngine.py", line 1113, in create_network
network.construct_from_dict(net_dict)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 460, in construct_from_dict
self.construct_layer(net_dict, name)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 652, in construct_layer
layer_class.transform_config_dict(layer_desc, network=self, get_layer=get_layer)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 4092, in transform_config_dict
super(BaseChoiceLayer, cls).transform_config_dict(d, network=network, get_layer=get_layer)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkLayer.py", line 448, in transform_config_dict
for src_name in src_names
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkLayer.py", line 449, in <listcomp>
if not src_name == "none"]
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 607, in get_layer
return self.construct_layer(net_dict=net_dict, name=src_name) # set get_layer to wrap construct_layer
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 655, in construct_layer
return add_layer(name=name, layer_class=layer_class, **layer_desc)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 760, in add_layer
layer = self._create_layer(name=name, layer_class=layer_class, **layer_desc)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 709, in _create_layer
layer = layer_class(**layer_desc)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 210, in __init__
y = self._get_output_subnet_unit(self.cell)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 834, in _get_output_subnet_unit
output = cell.get_output(rec_layer=self)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 2252, in get_output
shape_invariants=shape_invariants)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 1627, in _while_loop
back_prop=self.parent_rec_layer.back_prop)
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/ops/control_flow_ops.py", line 3556, in while_loop
return_same_structure)
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/ops/control_flow_ops.py", line 3087, in BuildLoop
pred, body, original_loop_vars, loop_vars, shape_invariants)
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/ops/control_flow_ops.py", line 3022, in _BuildLoop
body_result = body(*packed_vars_for_body)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 2102, in body
needed_outputs=needed_outputs)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 1422, in _construct
get_layer(layer_name)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 1398, in get_layer
layer = self.net.construct_layer(net_dict, name=name, get_layer=get_layer)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 652, in construct_layer
layer_class.transform_config_dict(layer_desc, network=self, get_layer=get_layer)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkLayer.py", line 448, in transform_config_dict
for src_name in src_names
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkLayer.py", line 449, in <listcomp>
if not src_name == "none"]
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 1398, in get_layer
layer = self.net.construct_layer(net_dict, name=name, get_layer=get_layer)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 652, in construct_layer
layer_class.transform_config_dict(layer_desc, network=self, get_layer=get_layer)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 5224, in transform_config_dict
d["weights"] = get_layer(d["weights"])
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 1398, in get_layer
layer = self.net.construct_layer(net_dict, name=name, get_layer=get_layer)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 655, in construct_layer
return add_layer(name=name, layer_class=layer_class, **layer_desc)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 760, in add_layer
layer = self._create_layer(name=name, layer_class=layer_class, **layer_desc)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 709, in _create_layer
layer = layer_class(**layer_desc)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkLayer.py", line 3003, in __init__
window_size=window_size.output if isinstance(window_size, LayerBase) else window_size)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkLayer.py", line 3159, in build_mask
tf.less(idxs, window_start.placeholder + window_size.placeholder)
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/ops/gen_math_ops.py", line 5095, in logical_and
"LogicalAnd", x=x, y=y, name=name)
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/framework/op_def_library.py", line 788, in _apply_op_helper
op_def=op_def)
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/util/deprecation.py", line 507, in new_func
return func(*args, **kwargs)
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/framework/ops.py", line 3300, in create_op
op_def=op_def)
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/framework/ops.py", line 1801, in __init__
self._traceback = tf_stack.extract_stack()
InvalidArgumentError (see above for traceback): Incompatible shapes: [14,1,45] vs. [45,1,45]
[[node output/rec/att_weights/LogicalAnd_1 (defined at /home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkLayer.py:3159) ]]
Exception InvalidArgumentError() in step 0.
Failing op: <tf.Operation 'output/rec/att_weights/LogicalAnd_1' type=LogicalAnd>
Execute again to debug the op inputs...
FetchHelper(0): <tf.Tensor 'output/rec/att_weights/LogicalAnd_2:0' shape=<unknown> dtype=bool> = shape (45, 1, 45), dtype bool, min/max False/True
FetchHelper(0): <tf.Tensor 'output/rec/att_weights/Reshape_2:0' shape=(?, 1, ?) dtype=bool> = shape (14, 1, 45), dtype bool, min/max False/True
Op inputs:
<tf.Tensor 'output/rec/att_weights/Reshape:0' shape=(?, 1, ?) dtype=bool>: shape (14, 1, 45), dtype bool, min/max False/True
<tf.Tensor 'output/rec/att_weights/LogicalAnd:0' shape=(?, 1, ?) dtype=bool>: shape (45, 1, 45), dtype bool, min/max False/True
Step meta information:
{'seq_idx': [0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44],
'seq_tag': ['train-clean-360-708-129393-0016',
'train-clean-360-4854-24732-0043',
'train-clean-360-3927-6557-0007',
'train-clean-360-4356-6506-0037',
'train-clean-360-3927-6556-0063',
'train-clean-100-3374-298026-0040',
'train-clean-360-8050-110238-0074',
'train-clean-360-8195-117382-0112',
'train-clean-360-671-131030-0024',
'train-clean-360-5448-19209-0024',
'train-clean-360-210-129396-0099',
'train-clean-360-8006-112593-0021',
'train-clean-360-210-129396-0126',
'train-clean-360-369-125882-0016',
'train-clean-360-2570-157243-0047',
'train-clean-360-7229-80665-0002',
'train-clean-360-209-4733-0022',
'train-clean-360-6918-47541-0068',
'train-clean-100-27-124992-0017',
'train-clean-360-210-129396-0078',
'train-clean-360-4363-14936-0014',
'train-clean-360-2285-163381-0010',
'train-clean-360-216-122451-0041',
'train-clean-100-1737-146161-0007',
'train-clean-100-2764-36616-0011',
'train-clean-360-1801-138032-0041',
'train-clean-360-2149-7239-0009',
'train-clean-360-2638-10172-0117',
'train-clean-360-8008-271812-0034',
'train-clean-100-328-129766-0017',
'train-clean-100-3259-158083-0039',
'train-clean-360-434-132649-0031',
'train-clean-360-434-132650-0013',
'train-clean-360-1974-139741-0052',
'train-clean-360-2368-157066-0001',
'train-clean-360-5133-30591-0042',
'train-clean-360-6497-234067-0018',
'train-clean-360-8050-110238-0068',
'train-clean-100-887-123290-0030',
'train-clean-360-5767-48579-0008',
'train-clean-360-6956-76046-0037',
'train-clean-360-525-126965-0115',
'train-clean-360-1705-142318-0072',
'train-clean-360-1943-138033-0065',
'train-clean-360-3448-5417-0009']}
Feed dict:
<tf.Tensor 'extern_data/placeholders/classes/classes:0' shape=(?, ?) dtype=int32>: shape (45, 12), dtype int32, min/max 0/9031, Data(name='classes', shape=(None,), dtype='int32', sparse=True, dim=10025, available_for_inference=False, batch_shape_meta=[B,T|'time:var:extern_data:classes'])
<tf.Tensor 'extern_data/placeholders/classes/classes_dim0_size:0' shape=(?,) dtype=int32>: shape (45,), dtype int32, min/max 2/12, ([ 2 2 2 3 3 3 2 3 4 5 4 4 6 5 5 6 5 8 5 7 6 9 6 5
6 6 7 7 7 6 7 6 6 7 8 7 6 6 7 6 8 12 9 7 8])
<tf.Tensor 'extern_data/placeholders/data/data:0' shape=(?, ?, 40) dtype=float32>: shape (45, 437, 40), dtype float32, min/max -5.747757/8.373278, mean/stddev 0.046555836/0.69089395, Data(name='data', shape=(None, 40), batch_shape_meta=[B,T|'time:var:extern_data:data',F|40])
<tf.Tensor 'extern_data/placeholders/data/data_dim0_size:0' shape=(?,) dtype=int32>: shape (45,), dtype int32, min/max 122/437, ([144 126 122 166 137 153 182 271 136 249 214 177 358 184 195 274 303 330
270 363 237 346 208 204 194 202 194 224 273 202 216 237 221 376 437 296
218 251 225 208 233 265 237 228 351])
<tf.Tensor 'globals/train_flag:0' shape=() dtype=bool>: bool(True)
EXCEPTION
Traceback (most recent call last):
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFEngine.py", line 585, in run
line: fetches_results = sess.run(
fetches_dict, feed_dict=feed_dict) # type: typing.Dict[str,typing.Union[numpy.ndarray,str]]
locals:
fetches_results = <not found>
sess = <local> <tensorflow.python.client.session.Session object at 0x7fb39a7ff630>
sess.run = <local> <bound method BaseSession.run of <tensorflow.python.client.session.Session object at 0x7fb39a7ff630>>
fetches_dict = <local> {'size:classes:0': <tf.Tensor 'extern_data/placeholders/classes/classes_dim0_size:0' shape=(?,) dtype=int32>, 'size:data:0': <tf.Tensor 'extern_data/placeholders/data/data_dim0_size:0' shape=(?,) dtype=int32>, 'loss': <tf.Tensor 'objective/add:0' shape=() dtype=float32>, 'cost:ctc': <tf.Tensor 'o..., len = 14
feed_dict = <local> {<tf.Tensor 'extern_data/placeholders/classes/classes:0' shape=(?, ?) dtype=int32>: array([[2085, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0],
[1203, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0],
[6137, 0, 0, 0, ...
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/client/session.py", line 929, in run
line: result = self._run(None, fetches, feed_dict, options_ptr,
run_metadata_ptr)
locals:
result = <not found>
self = <local> <tensorflow.python.client.session.Session object at 0x7fb39a7ff630>
self._run = <local> <bound method BaseSession._run of <tensorflow.python.client.session.Session object at 0x7fb39a7ff630>>
fetches = <local> {'size:classes:0': <tf.Tensor 'extern_data/placeholders/classes/classes_dim0_size:0' shape=(?,) dtype=int32>, 'size:data:0': <tf.Tensor 'extern_data/placeholders/data/data_dim0_size:0' shape=(?,) dtype=int32>, 'loss': <tf.Tensor 'objective/add:0' shape=() dtype=float32>, 'cost:ctc': <tf.Tensor 'o..., len = 14
feed_dict = <local> {<tf.Tensor 'extern_data/placeholders/classes/classes:0' shape=(?, ?) dtype=int32>: array([[2085, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0],
[1203, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0],
[6137, 0, 0, 0, ...
options_ptr = <local> None
run_metadata_ptr = <local> None
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/client/session.py", line 1152, in _run
line: results = self._do_run(handle, final_targets, final_fetches,
feed_dict_tensor, options, run_metadata)
locals:
results = <not found>
self = <local> <tensorflow.python.client.session.Session object at 0x7fb39a7ff630>
self._do_run = <local> <bound method BaseSession._do_run of <tensorflow.python.client.session.Session object at 0x7fb39a7ff630>>
handle = <local> None
final_targets = <local> [<tf.Operation 'optim_and_step_incr' type=NoOp>]
final_fetches = <local> [<tf.Tensor 'objective/add:0' shape=() dtype=float32>, <tf.Tensor 'objective/loss/loss/loss_ctc/Sum:0' shape=() dtype=float32>, <tf.Tensor 'objective/loss/error/loss_ctc_error/Sum:0' shape=() dtype=float32>, <tf.Tensor 'objective/loss/loss_init/truediv:0' shape=() dtype=float32>, <tf.Tensor 'obje..., len = 11
feed_dict_tensor = <local> {<tf.Tensor 'extern_data/placeholders/classes/classes:0' shape=(?, ?) dtype=int32>: array([[2085, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0],
[1203, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0],
[6137, 0, 0, 0, ...
options = <local> None
run_metadata = <local> None
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/client/session.py", line 1328, in _do_run
line: return self._do_call(_run_fn, feeds, fetches, targets, options,
run_metadata)
locals:
self = <local> <tensorflow.python.client.session.Session object at 0x7fb39a7ff630>
self._do_call = <local> <bound method BaseSession._do_call of <tensorflow.python.client.session.Session object at 0x7fb39a7ff630>>
_run_fn = <local> <function BaseSession._do_run.<locals>._run_fn at 0x7fb08df05378>
feeds = <local> {<tensorflow.python.pywrap_tensorflow_internal.TF_Output; proxy of <Swig Object of type 'TF_Output *' at 0x7fb0e8ad8b10> >: array([[2085, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0],
[1203, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
...
fetches = <local> [<tensorflow.python.pywrap_tensorflow_internal.TF_Output; proxy of <Swig Object of type 'TF_Output *' at 0x7fb0d93b6e10> >, <tensorflow.python.pywrap_tensorflow_internal.TF_Output; proxy of <Swig Object of type 'TF_Output *' at 0x7fb0d9a3e270> >, <tensorflow.python.pywrap_tensorflow_internal.TF_O..., len = 11
targets = <local> [<Swig Object of type 'TF_Operation *' at 0x7fb08dc24ba0>]
options = <local> None
run_metadata = <local> None
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/client/session.py", line 1348, in _do_call
line: raise type(e)(node_def, op, message)
locals:
type = <builtin> <class 'type'>
e = <not found>
node_def = <local> name: "output/rec/att_weights/LogicalAnd_1"
op: "LogicalAnd"
input: "output/rec/att_weights/Reshape"
input: "output/rec/att_weights/LogicalAnd"
op = <local> <tf.Operation 'output/rec/att_weights/LogicalAnd_1' type=LogicalAnd>
message = <local> 'Incompatible shapes: [14,1,45] vs. [45,1,45]\n\t [[node output/rec/att_weights/LogicalAnd_1 (defined at /home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkLayer.py:3159) ]]', len = 229
InvalidArgumentError: Incompatible shapes: [14,1,45] vs. [45,1,45]
[[node output/rec/att_weights/LogicalAnd_1 (defined at /home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkLayer.py:3159) ]]
Caused by op 'output/rec/att_weights/LogicalAnd_1', defined at:
File "./returnn/rnn.py", line 654, in <module>
main(sys.argv)
File "./returnn/rnn.py", line 642, in main
execute_main_task()
File "./returnn/rnn.py", line 451, in execute_main_task
engine.init_train_from_config(config, train_data, dev_data, eval_data)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFEngine.py", line 891, in init_train_from_config
self.init_network_from_config(config)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFEngine.py", line 934, in init_network_from_config
self._init_network(net_desc=net_dict, epoch=self.epoch)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFEngine.py", line 1081, in _init_network
net_dict=net_desc)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFEngine.py", line 1113, in create_network
network.construct_from_dict(net_dict)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 460, in construct_from_dict
self.construct_layer(net_dict, name)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 652, in construct_layer
layer_class.transform_config_dict(layer_desc, network=self, get_layer=get_layer)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 4092, in transform_config_dict
super(BaseChoiceLayer, cls).transform_config_dict(d, network=network, get_layer=get_layer)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkLayer.py", line 448, in transform_config_dict
for src_name in src_names
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkLayer.py", line 449, in <listcomp>
if not src_name == "none"]
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 607, in get_layer
return self.construct_layer(net_dict=net_dict, name=src_name) # set get_layer to wrap construct_layer
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 655, in construct_layer
return add_layer(name=name, layer_class=layer_class, **layer_desc)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 760, in add_layer
layer = self._create_layer(name=name, layer_class=layer_class, **layer_desc)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 709, in _create_layer
layer = layer_class(**layer_desc)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 210, in __init__
y = self._get_output_subnet_unit(self.cell)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 834, in _get_output_subnet_unit
output = cell.get_output(rec_layer=self)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 2252, in get_output
shape_invariants=shape_invariants)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 1627, in _while_loop
back_prop=self.parent_rec_layer.back_prop)
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/ops/control_flow_ops.py", line 3556, in while_loop
return_same_structure)
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/ops/control_flow_ops.py", line 3087, in BuildLoop
pred, body, original_loop_vars, loop_vars, shape_invariants)
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/ops/control_flow_ops.py", line 3022, in _BuildLoop
body_result = body(*packed_vars_for_body)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 2102, in body
needed_outputs=needed_outputs)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 1422, in _construct
get_layer(layer_name)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 1398, in get_layer
layer = self.net.construct_layer(net_dict, name=name, get_layer=get_layer)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 652, in construct_layer
layer_class.transform_config_dict(layer_desc, network=self, get_layer=get_layer)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkLayer.py", line 448, in transform_config_dict
for src_name in src_names
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkLayer.py", line 449, in <listcomp>
if not src_name == "none"]
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 1398, in get_layer
layer = self.net.construct_layer(net_dict, name=name, get_layer=get_layer)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 652, in construct_layer
layer_class.transform_config_dict(layer_desc, network=self, get_layer=get_layer)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 5224, in transform_config_dict
d["weights"] = get_layer(d["weights"])
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkRecLayer.py", line 1398, in get_layer
layer = self.net.construct_layer(net_dict, name=name, get_layer=get_layer)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 655, in construct_layer
return add_layer(name=name, layer_class=layer_class, **layer_desc)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 760, in add_layer
layer = self._create_layer(name=name, layer_class=layer_class, **layer_desc)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetwork.py", line 709, in _create_layer
layer = layer_class(**layer_desc)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkLayer.py", line 3003, in __init__
window_size=window_size.output if isinstance(window_size, LayerBase) else window_size)
File "/home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkLayer.py", line 3159, in build_mask
tf.less(idxs, window_start.placeholder + window_size.placeholder)
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/ops/gen_math_ops.py", line 5095, in logical_and
"LogicalAnd", x=x, y=y, name=name)
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/framework/op_def_library.py", line 788, in _apply_op_helper
op_def=op_def)
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/util/deprecation.py", line 507, in new_func
return func(*args, **kwargs)
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/framework/ops.py", line 3300, in create_op
op_def=op_def)
File "/home/ubuntu/tf1.13/lib/python3.6/site-packages/tensorflow/python/framework/ops.py", line 1801, in __init__
self._traceback = tf_stack.extract_stack()
InvalidArgumentError (see above for traceback): Incompatible shapes: [14,1,45] vs. [45,1,45]
[[node output/rec/att_weights/LogicalAnd_1 (defined at /home/ubuntu/rwth-i6/returnn-experiments/2018-asr-attention/librispeech/full-setup-attention/returnn/TFNetworkLayer.py:3159) ]]
Save model under data/exp-local_win05/model.pretrain.001.crash_0
Trainer not finalized, quitting.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment