Skip to content

Instantly share code, notes, and snippets.

@ezyang
Created November 20, 2022 21:35
Show Gist options
  • Star 0 You must be signed in to star a gist
  • Fork 0 You must be signed in to fork a gist
  • Save ezyang/3b3be6de1670829b311f24e09f7cc76e to your computer and use it in GitHub Desktop.
Save ezyang/3b3be6de1670829b311f24e09f7cc76e to your computer and use it in GitHub Desktop.
Sweep logs for symbolic-shapes --accuracy --backend aot_eager --training (TORCHDYNAMO_DYNAMIC_SHAPES=1) - 54c29d7f15911f7bd62456353611b8eab4ca42dd Sun Nov 20 11:10:13 PST 2022
Running torchbench.py BERT_pytorch...
cuda train BERT_pytorch PASS
Running torchbench.py Background_Matting...
cuda train Background_Matting PASS
WARNING:root:DALLE2_pytorch failed to load
Eager model failed to run
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1013, in validate_model
self.model_iter_fn(model, example_inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 338, in forward_and_backward_pass
self.grad_scaler.scale(loss).backward()
File "/data/users/ezyang/b/pytorch/torch/_tensor.py", line 473, in backward
torch.autograd.backward(
File "/data/users/ezyang/b/pytorch/torch/autograd/__init__.py", line 197, in backward
Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1961, in run
device, name, model, example_inputs, batch_size = runner.load_model(
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 283, in load_model
self.validate_model(model, example_inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1015, in validate_model
raise NotImplementedError("Eager model failed to run")
NotImplementedError: Eager model failed to run
Running torchbench.py LearningToPaint...
cuda train LearningToPaint PASS
Running torchbench.py Super_SloMo...
cuda train Super_SloMo FAIL (TIMEOUT)
Running torchbench.py alexnet...
cuda train alexnet PASS
Running torchbench.py attention_is_all_you_need_pytorch...
cuda train attention_is_all_you_need_pytorch PASS
Running torchbench.py dcgan...
cuda train dcgan PASS
Running torchbench.py densenet121...
cuda train densenet121 PASS
WARNING:root:detectron2_fcos_r_50_fpn failed to load
FCOS train is not supported by upstream detectron2. See GH Issue: https://github.com/facebookresearch/detectron2/issues/4369.
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1961, in run
device, name, model, example_inputs, batch_size = runner.load_model(
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 252, in load_model
benchmark = benchmark_cls(
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/util/model.py", line 18, in __call__
obj = type.__call__(cls, *args, **kwargs)
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/detectron2_fcos_r_50_fpn/__init__.py", line 15, in __init__
super().__init__(variant="COCO-Detection/fcos_R_50_FPN_1x.py", test=test, device=device,
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/util/framework/detectron2/model_factory.py", line 100, in __init__
loader = self.setup_train(cfg, args)
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/util/framework/detectron2/model_factory.py", line 110, in setup_train
raise NotImplementedError("FCOS train is not supported by upstream detectron2. " \
NotImplementedError: FCOS train is not supported by upstream detectron2. See GH Issue: https://github.com/facebookresearch/detectron2/issues/4369.
WARNING:root:detectron2_maskrcnn_r_50_c4 failed to load
Eager model failed to run
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1013, in validate_model
self.model_iter_fn(model, example_inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 337, in forward_and_backward_pass
loss = self.compute_loss(pred)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 327, in compute_loss
return reduce_to_scalar_loss(pred)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/testing.py", line 99, in reduce_to_scalar_loss
return sum([reduce_to_scalar_loss(x) for x in out]) / len(out)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/testing.py", line 99, in <listcomp>
return sum([reduce_to_scalar_loss(x) for x in out]) / len(out)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/testing.py", line 109, in reduce_to_scalar_loss
return sum([reduce_to_scalar_loss(value) for value in out.values()]) / len(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/testing.py", line 109, in <listcomp>
return sum([reduce_to_scalar_loss(value) for value in out.values()]) / len(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/testing.py", line 114, in reduce_to_scalar_loss
raise NotImplementedError("Don't know how to reduce", type(out))
NotImplementedError: ("Don't know how to reduce", <class 'detectron2.structures.instances.Instances'>)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1961, in run
device, name, model, example_inputs, batch_size = runner.load_model(
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 283, in load_model
self.validate_model(model, example_inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1015, in validate_model
raise NotImplementedError("Eager model failed to run")
NotImplementedError: Eager model failed to run
Running torchbench.py dlrm...
cuda train dlrm PASS
/data/users/ezyang/b/pytorch/torch/utils/tensorboard/__init__.py:4: DeprecationWarning: distutils Version classes are deprecated. Use packaging.version instead.
if not hasattr(tensorboard, "__version__") or LooseVersion(
/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/gym/core.py:317: DeprecationWarning: WARN: Initializing wrapper in old step API which returns one bool instead of two. It is recommended to set `new_step_api=True` to use new step API. This will be the default behaviour in future.
deprecation(
Running torchbench.py drq...
cuda train drq PASS
Running torchbench.py fastNLP_Bert...
cuda train fastNLP_Bert PASS
Running torchbench.py functorch_dp_cifar10...
cuda train functorch_dp_cifar10 PASS
Running torchbench.py functorch_maml_omniglot...
cuda train functorch_maml_omniglot PASS
Running torchbench.py hf_Albert...
cuda train hf_Albert PASS
Running torchbench.py hf_Bart...
cuda train hf_Bart PASS
Running torchbench.py hf_Bert...
cuda train hf_Bert PASS
Running torchbench.py hf_BigBird...
ERROR:common:Output 0 of CompiledFunctionBackward is a view and is being modified inplace. This view was created inside a custom Function (or because an input was returned as-is) and the autograd logic to handle view+inplace would override the custom backward associated with the custom Function, leading to incorrect gradients. This behavior is forbidden. You can fix this by cloning the output of the custom Function.
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 333, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 336, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2462, in forward
outputs = self.bert(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2148, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1641, in forward
layer_outputs = layer_module(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1493, in forward
self_attention_outputs = self.attention(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1406, in forward
self_outputs = self.self(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 475, in forward
context_layer, attention_probs = self.bigbird_block_sparse_attention(
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 573, in bigbird_block_sparse_attention
np.random.seed(seed)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 635, in <graph break in bigbird_block_sparse_attention>
first_context_layer.unsqueeze_(2)
RuntimeError: Output 0 of CompiledFunctionBackward is a view and is being modified inplace. This view was created inside a custom Function (or because an input was returned as-is) and the autograd logic to handle view+inplace would override the custom backward associated with the custom Function, leading to incorrect gradients. This behavior is forbidden. You can fix this by cloning the output of the custom Function.
TorchDynamo optimized model failed to run because of following error
cuda train hf_BigBird FAIL
Running torchbench.py hf_DistilBert...
cuda train hf_DistilBert PASS
Running torchbench.py hf_GPT2...
cuda train hf_GPT2 PASS
Running torchbench.py hf_GPT2_large...
cuda train hf_GPT2_large PASS
Running torchbench.py hf_Longformer...
ERROR:common:Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 333, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 336, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1813, in forward
outputs = self.longformer(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1696, in forward
padding_len, input_ids, attention_mask, token_type_ids, position_ids, inputs_embeds = self._pad_to_window_size(
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1715, in <graph break in forward>
encoder_outputs = self.encoder(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1265, in forward
is_global_attn = is_index_global_attn.flatten().any().item()
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1297, in <graph break in forward>
layer_outputs = layer_module(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1221, in forward
self_attn_outputs = self.attention(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1157, in forward
self_outputs = self.self(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 542, in forward
def forward(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1125, in compiled_function
fw_outs_including_aliases.append(input_alias.as_strided(out_tensor_meta.size(), out_tensor_meta.stride(), out_tensor_meta.storage_offset()))
RuntimeError: Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
TorchDynamo optimized model failed to run because of following error
cuda train hf_Longformer FAIL
Running torchbench.py hf_Reformer...
cuda train hf_Reformer PASS
Running torchbench.py hf_T5...
WARNING:common:fp64 golden ref were not generated for hf_T5
ERROR:common:
from user code:
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 529, in <graph break in forward>
scores += position_bias
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1091, in run_node
return node.target(*args, **kwargs)
RuntimeError: Output 0 of AsStridedBackward0 is a view of a view which was created in no_grad mode and is being modified inplace with grad mode enabled. Given that this use case is ambiguous and error-prone, it is forbidden. You can clarify your code by moving both the view and the inplace either both inside the no_grad block (if you don't want the inplace to be tracked) or both outside (if you want the inplace to be tracked).
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1057, in get_fake_value
return wrap_fake_exception(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 741, in wrap_fake_exception
return fn()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1058, in <lambda>
lambda: run_node(tx.output, node, args, kwargs, nnmodule)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1100, in run_node
raise RuntimeError(
RuntimeError: Failed running call_function <built-in function iadd>(*(FakeTensor(FakeTensor(..., device='meta', size=(s2, s1, s0, s0),
grad_fn=<AsStridedBackward0>), cuda:0), FakeTensor(FakeTensor(..., device='meta', size=(s2, s1, s0, s0), grad_fn=<AddBackward0>), cuda:0)), **{}):
Output 0 of AsStridedBackward0 is a view of a view which was created in no_grad mode and is being modified inplace with grad mode enabled. Given that this use case is ambiguous and error-prone, it is forbidden. You can clarify your code by moving both the view and the inplace either both inside the no_grad block (if you don't want the inplace to be tracked) or both outside (if you want the inplace to be tracked).
(scroll up for backtrace)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 333, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 336, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/util/framework/huggingface/model_factory.py", line 41, in forward
return self.model(input_ids=input_ids, decoder_input_ids=decoder_input_ids)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 945, in forward
attention_mask = torch.ones(batch_size, mask_seq_length).to(inputs_embeds.device)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in <graph break in forward>
layer_outputs = layer_module(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 286, in catch_errors
return callback(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 476, in _convert_frame
result = inner_convert(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 118, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 349, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 404, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 392, in transform
tracer.run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1617, in run
super().run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 483, in run
and self.step()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 453, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 120, in impl
self.push(fn_var.call_function(self, self.popn(nargs), {}))
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builtin.py", line 320, in call_function
return wrap_fx_proxy(tx, proxy, **options)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 652, in wrap_fx_proxy
return wrap_fx_proxy_cls(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 693, in wrap_fx_proxy_cls
example_value = get_fake_value(proxy.node, tx)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1070, in get_fake_value
raise TorchRuntimeError() from e
torch._dynamo.exc.TorchRuntimeError:
from user code:
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 529, in <graph break in forward>
scores += position_bias
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
cuda train hf_T5 FAIL
Running torchbench.py hf_T5_base...
WARNING:common:fp64 golden ref were not generated for hf_T5_base
ERROR:common:
from user code:
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 529, in <graph break in forward>
scores += position_bias
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1091, in run_node
return node.target(*args, **kwargs)
RuntimeError: Output 0 of AsStridedBackward0 is a view of a view which was created in no_grad mode and is being modified inplace with grad mode enabled. Given that this use case is ambiguous and error-prone, it is forbidden. You can clarify your code by moving both the view and the inplace either both inside the no_grad block (if you don't want the inplace to be tracked) or both outside (if you want the inplace to be tracked).
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1057, in get_fake_value
return wrap_fake_exception(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 741, in wrap_fake_exception
return fn()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1058, in <lambda>
lambda: run_node(tx.output, node, args, kwargs, nnmodule)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1100, in run_node
raise RuntimeError(
RuntimeError: Failed running call_function <built-in function iadd>(*(FakeTensor(FakeTensor(..., device='meta', size=(s2, s1, s0, s0),
grad_fn=<AsStridedBackward0>), cuda:0), FakeTensor(FakeTensor(..., device='meta', size=(s2, s1, s0, s0), grad_fn=<AddBackward0>), cuda:0)), **{}):
Output 0 of AsStridedBackward0 is a view of a view which was created in no_grad mode and is being modified inplace with grad mode enabled. Given that this use case is ambiguous and error-prone, it is forbidden. You can clarify your code by moving both the view and the inplace either both inside the no_grad block (if you don't want the inplace to be tracked) or both outside (if you want the inplace to be tracked).
(scroll up for backtrace)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 333, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 336, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/util/framework/huggingface/model_factory.py", line 41, in forward
return self.model(input_ids=input_ids, decoder_input_ids=decoder_input_ids)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 945, in forward
attention_mask = torch.ones(batch_size, mask_seq_length).to(inputs_embeds.device)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in <graph break in forward>
layer_outputs = layer_module(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 286, in catch_errors
return callback(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 476, in _convert_frame
result = inner_convert(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 118, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 349, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 404, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 392, in transform
tracer.run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1617, in run
super().run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 483, in run
and self.step()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 453, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 120, in impl
self.push(fn_var.call_function(self, self.popn(nargs), {}))
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builtin.py", line 320, in call_function
return wrap_fx_proxy(tx, proxy, **options)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 652, in wrap_fx_proxy
return wrap_fx_proxy_cls(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 693, in wrap_fx_proxy_cls
example_value = get_fake_value(proxy.node, tx)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1070, in get_fake_value
raise TorchRuntimeError() from e
torch._dynamo.exc.TorchRuntimeError:
from user code:
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 529, in <graph break in forward>
scores += position_bias
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
cuda train hf_T5_base FAIL
Running torchbench.py hf_T5_large...
cuda train hf_T5_large PASS
Running torchbench.py lennard_jones...
cuda train lennard_jones PASS
Running torchbench.py maml_omniglot...
cuda train maml_omniglot PASS
Running torchbench.py mnasnet1_0...
cuda train mnasnet1_0 PASS
Running torchbench.py mobilenet_v2...
cuda train mobilenet_v2 PASS
Running torchbench.py mobilenet_v2_quantized_qat...
WARNING:common:fp64 golden ref were not generated for mobilenet_v2_quantized_qat
ERROR:common:output 2: torch.Size([0]) != torch.Size([32])
While executing %_fused_moving_avg_obs_fq_helper_functional_1 : [#users=6] = call_function[target=torch.ops.aten._fused_moving_avg_obs_fq_helper_functional.default](args = (%mul : Tensor[size=[32, 3, 3, 3], stride=[27, 9, 3, 1]], %primals_170 : Tensor[size=[1], stride=[1]], %primals_169 : Tensor[size=[1], stride=[1]], %clone_6 : Tensor[size=[0], stride=[1]], %clone_7 : Tensor[size=[0], stride=[1]], %clone_4 : Tensor[size=[1], stride=[1]], %clone_5 : Tensor[size=[1], stride=[1]], 0.01, -128, 127, 0, True, True), kwargs = {})
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 333, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 336, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/fx/graph_module.py", line 660, in call_wrapped
return self._wrapped_call(self, *args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/fx/graph_module.py", line 279, in __call__
raise e
File "/data/users/ezyang/b/pytorch/torch/fx/graph_module.py", line 269, in __call__
return super(self.cls, obj).__call__(*args, **kwargs) # type: ignore[misc]
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "<eval_with_key>.8", line 4, in forward
def forward(self, x : torch.Tensor) -> torch.Tensor:
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1065, in compiled_function
outs = CompiledFunction.apply(*no_dupe_args_with_synthetic_bases)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 977, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 595, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 158, in run_node
check(nv, rv, lambda: f"output {i}")
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 126, in check
assert nv.size() == rv.size(), f"{desc()}: {nv.size()} != {rv.size()}"
AssertionError: output 2: torch.Size([0]) != torch.Size([32])
While executing %_fused_moving_avg_obs_fq_helper_functional_1 : [#users=6] = call_function[target=torch.ops.aten._fused_moving_avg_obs_fq_helper_functional.default](args = (%mul : Tensor[size=[32, 3, 3, 3], stride=[27, 9, 3, 1]], %primals_170 : Tensor[size=[1], stride=[1]], %primals_169 : Tensor[size=[1], stride=[1]], %clone_6 : Tensor[size=[0], stride=[1]], %clone_7 : Tensor[size=[0], stride=[1]], %clone_4 : Tensor[size=[1], stride=[1]], %clone_5 : Tensor[size=[1], stride=[1]], 0.01, -128, 127, 0, True, True), kwargs = {})
TorchDynamo optimized model failed to run because of following error
cuda train mobilenet_v2_quantized_qat FAIL
Running torchbench.py mobilenet_v3_large...
cuda train mobilenet_v3_large PASS
devgpu019:2092714:2092714 [0] NCCL INFO NCCL_SOCKET_IFNAME set by environment to eth0
devgpu019:2092714:2092714 [0] NCCL INFO NCCL_SOCKET_IFNAME set to eth0
devgpu019:2092714:2092714 [0] NCCL INFO Bootstrap : Using eth0:2803:6080:6188:70b4::1<0>
devgpu019:2092714:2092714 [0] NCCL INFO NET/Plugin : No plugin found (libnccl-net.so), using internal implementation
devgpu019:2092714:2092714 [0] NCCL INFO cudaDriverVersion 11040
NCCL version 2.14.3+cuda11.4
devgpu019:2092714:2093278 [0] NCCL INFO NCCL_IB_DISABLE set by environment to 1.
devgpu019:2092714:2093278 [0] NCCL INFO NCCL_SOCKET_IFNAME set by environment to eth0
devgpu019:2092714:2093278 [0] NCCL INFO NET/Socket : Using [0]eth0:2803:6080:6188:70b4::1<0>
devgpu019:2092714:2093278 [0] NCCL INFO Using network Socket
devgpu019:2092714:2093278 [0] NCCL INFO NET/Socket : GPU Direct RDMA Disabled for HCA 0 'eth0'
devgpu019:2092714:2093278 [0] NCCL INFO === System : maxBw 5000.0 totalBw 0.0 ===
devgpu019:2092714:2093278 [0] NCCL INFO CPU/0 (1/1/2)
devgpu019:2092714:2093278 [0] NCCL INFO + PCI[12.0] - PCI/D000 (11f840001d9bfbe1)
devgpu019:2092714:2093278 [0] NCCL INFO + PCI[24.0] - PCI/F000 (11f840001d9bfbe0)
devgpu019:2092714:2093278 [0] NCCL INFO + PCI[24.0] - GPU/11000 (0)
devgpu019:2092714:2093278 [0] NCCL INFO + PCI[12.0] - NIC/30000
devgpu019:2092714:2093278 [0] NCCL INFO ==========================================
devgpu019:2092714:2093278 [0] NCCL INFO GPU/11000 :GPU/11000 (0/5000.000000/LOC) CPU/0 (3/12.000000/PHB)
devgpu019:2092714:2093278 [0] NCCL INFO Setting affinity for GPU 0 to ffffff,00000000,00000000,00ffffff
devgpu019:2092714:2093278 [0] NCCL INFO Pattern 4, crossNic 0, nChannels 16, bw 44.000000/44.000000, type LOC/PIX, sameChannels 1
devgpu019:2092714:2093278 [0] NCCL INFO 0 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 1 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 2 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 3 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 4 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 5 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 6 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 7 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 8 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 9 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 10 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 11 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 12 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 13 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 14 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 15 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO Pattern 3, crossNic 0, nChannels 16, bw 44.000000/44.000000, type LOC/PIX, sameChannels 1
devgpu019:2092714:2093278 [0] NCCL INFO 0 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 1 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 2 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 3 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 4 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 5 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 6 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 7 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 8 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 9 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 10 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 11 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 12 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 13 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 14 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 15 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO Pattern 3, crossNic 0, nChannels 16, bw 44.000000/44.000000, type LOC/PIX, sameChannels 1
devgpu019:2092714:2093278 [0] NCCL INFO 0 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 1 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 2 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 3 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 4 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 5 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 6 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 7 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 8 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 9 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 10 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 11 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 12 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 13 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 14 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO 15 : GPU/0
devgpu019:2092714:2093278 [0] NCCL INFO Tree 0 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 16 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 1 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 17 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 2 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 18 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 3 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 19 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 4 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 20 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 5 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 21 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 6 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 22 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 7 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 23 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 8 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 24 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 9 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 25 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 10 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 26 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 11 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 27 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 12 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 28 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 13 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 29 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 14 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 30 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 15 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Tree 31 : -1 -> 0 -> -1/-1/-1
devgpu019:2092714:2093278 [0] NCCL INFO Channel 00/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 01/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 02/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 03/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 04/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 05/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 06/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 07/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 08/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 09/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 10/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 11/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 12/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 13/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 14/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 15/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 16/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 17/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 18/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 19/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 20/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 21/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 22/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 23/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 24/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 25/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 26/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 27/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 28/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 29/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 30/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Channel 31/32 : 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 00 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 01 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 02 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 03 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 04 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 05 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 06 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 07 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 08 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 09 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 10 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 11 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 12 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 13 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 14 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 15 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 16 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 17 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 18 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 19 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 20 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 21 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 22 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 23 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 24 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 25 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 26 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 27 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 28 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 29 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 30 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Ring 31 : 0 -> 0 -> 0
devgpu019:2092714:2093278 [0] NCCL INFO Trees [0] -1/-1/-1->0->-1 [1] -1/-1/-1->0->-1 [2] -1/-1/-1->0->-1 [3] -1/-1/-1->0->-1 [4] -1/-1/-1->0->-1 [5] -1/-1/-1->0->-1 [6] -1/-1/-1->0->-1 [7] -1/-1/-1->0->-1 [8] -1/-1/-1->0->-1 [9] -1/-1/-1->0->-1 [10] -1/-1/-1->0->-1 [11] -1/-1/-1->0->-1 [12] -1/-1/-1->0->-1 [13] -1/-1/-1->0->-1 [14] -1/-1/-1->0->-1 [15] -1/-1/-1->0->-1 [16] -1/-1/-1->0->-1 [17] -1/-1/-1->0->-1 [18] -1/-1/-1->0->-1 [19] -1/-1/-1->0->-1 [20] -1/-1/-1->0->-1 [21] -1/-1/-1->0->-1 [22] -1/-1/-1->0->-1 [23] -1/-1/-1->0->-1 [24] -1/-1/-1->0->-1 [25] -1/-1/-1->0->-1 [26] -1/-1/-1->0->-1 [27] -1/-1/-1->0->-1 [28] -1/-1/-1->0->-1 [29] -1/-1/-1->0->-1 [30] -1/-1/-1->0->-1 [31] -1/-1/-1->0->-1
devgpu019:2092714:2093278 [0] NCCL INFO Connected all rings
devgpu019:2092714:2093278 [0] NCCL INFO Connected all trees
devgpu019:2092714:2093278 [0] NCCL INFO 32 coll channels, 32 p2p channels, 32 p2p channels per peer
devgpu019:2092714:2093279 [0] NCCL INFO New proxy send connection 0 from local rank 0, transport 2
devgpu019:2092714:2093278 [0] NCCL INFO Connection to proxy localRank 0 -> connection 0x7fa910002e80
devgpu019:2092714:2093278 [0] NCCL INFO comm 0x6ae06c0 rank 0 nranks 1 cudaDev 0 busId 11000 - Init COMPLETE
Running torchbench.py moco...
ERROR:common:
from user code:
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/moco/moco/builder.py", line 172, in concat_all_gather
torch.distributed.all_gather(tensors_gather, tensor, async_op=False)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1091, in run_node
return node.target(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/distributed/distributed_c10d.py", line 1346, in wrapper
return func(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/distributed/distributed_c10d.py", line 2341, in all_gather
work = default_pg.allgather([tensor_list], [tensor])
File "/data/users/ezyang/b/pytorch/torch/_subclasses/fake_tensor.py", line 875, in __torch_dispatch__
r = func(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_ops.py", line 297, in __call__
return self._op(*args, **kwargs or {})
RuntimeError: Tensors must be CUDA and dense
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1057, in get_fake_value
return wrap_fake_exception(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 741, in wrap_fake_exception
return fn()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1058, in <lambda>
lambda: run_node(tx.output, node, args, kwargs, nnmodule)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1100, in run_node
raise RuntimeError(
RuntimeError: Failed running call_function <function all_gather at 0x7fb232f7fee0>(*([FakeTensor(FakeTensor(..., device='meta', size=(s0, s1, s2, s2)), cuda:0)], FakeTensor(FakeTensor(..., device='meta', size=(s0, s1, s2, s2)), cuda:0)), **{'async_op': False}):
Tensors must be CUDA and dense
(scroll up for backtrace)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 333, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 336, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/parallel/distributed.py", line 1098, in forward
output = self._run_ddp_forward(*inputs, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/parallel/distributed.py", line 1051, in _run_ddp_forward
return module_to_run(*inputs[0], **kwargs[0]) # type: ignore[index]
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/moco/moco/builder.py", line 130, in forward
self._momentum_update_key_encoder() # update the key encoder
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/moco/moco/builder.py", line 133, in <graph break in forward>
im_k, idx_unshuffle = self._batch_shuffle_ddp(im_k)
File "/data/users/ezyang/b/pytorch/torch/autograd/grad_mode.py", line 27, in decorate_context
return func(*args, **kwargs)
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/moco/moco/builder.py", line 76, in _batch_shuffle_ddp
x_gather = concat_all_gather(x)
File "/data/users/ezyang/b/pytorch/torch/autograd/grad_mode.py", line 27, in decorate_context
return func(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 286, in catch_errors
return callback(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 476, in _convert_frame
result = inner_convert(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 118, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 349, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 404, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 392, in transform
tracer.run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1617, in run
super().run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 483, in run
and self.step()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 453, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 287, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 959, in CALL_FUNCTION_KW
self.call_function(fn, args, kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 395, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/torch.py", line 430, in call_function
tensor_variable = wrap_fx_proxy(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 652, in wrap_fx_proxy
return wrap_fx_proxy_cls(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 693, in wrap_fx_proxy_cls
example_value = get_fake_value(proxy.node, tx)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1070, in get_fake_value
raise TorchRuntimeError() from e
torch._dynamo.exc.TorchRuntimeError:
from user code:
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/moco/moco/builder.py", line 172, in concat_all_gather
torch.distributed.all_gather(tensors_gather, tensor, async_op=False)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
cuda train moco FAIL
devgpu019:2092714:2093279 [0] NCCL INFO [Service thread] Connection closed by localRank 0
devgpu019:2092714:2092714 [0] NCCL INFO comm 0x6ae06c0 rank 0 nranks 1 cudaDev 0 busId 11000 - Abort COMPLETE
Running torchbench.py nvidia_deeprecommender...
cuda train nvidia_deeprecommender PASS
Running torchbench.py pytorch_CycleGAN_and_pix2pix...
--dataroot /data/users/ezyang/b/torchbenchmark/torchbenchmark/data/.data/pytorch_CycleGAN_and_pix2pix_inputs/datasets/horse2zebra --name horse2zebra --model cycle_gan --display_id 0 --n_epochs 3 --n_epochs_decay 3 --gpu_ids 0 --checkpoints_dir /data/users/ezyang/b/torchbenchmark/torchbenchmark/models/pytorch_CycleGAN_and_pix2pix/.data/checkpoints
cuda train pytorch_CycleGAN_and_pix2pix PASS
Running torchbench.py pytorch_stargan...
cuda train pytorch_stargan PASS
downloading en-ud-v2.zip
0%| | 0.00/688k [00:00<?, ?B/s] 7%|▋ | 49.2k/688k [00:00<00:01, 484kB/s] 14%|█▍ | 98.3k/688k [00:00<00:01, 395kB/s] 24%|██▍ | 164k/688k [00:00<00:01, 489kB/s] 33%|███▎ | 229k/688k [00:00<00:00, 539kB/s] 43%|████▎ | 295k/688k [00:00<00:00, 567kB/s] 55%|█████▍ | 377k/688k [00:00<00:00, 637kB/s] 64%|██████▍ | 442k/688k [00:00<00:00, 632kB/s] 79%|███████▊ | 541k/688k [00:00<00:00, 727kB/s] 93%|█████████▎| 639k/688k [00:00<00:00, 790kB/s] 100%|██████████| 688k/688k [00:01<00:00, 676kB/s]
Running torchbench.py pytorch_struct...
extracting
cuda train pytorch_struct PASS
Running torchbench.py pytorch_unet...
cuda train pytorch_unet PASS
Running torchbench.py resnet152...
cuda train resnet152 PASS
Running torchbench.py resnet18...
cuda train resnet18 PASS
Running torchbench.py resnet50...
cuda train resnet50 PASS
Running torchbench.py resnet50_quantized_qat...
WARNING:common:fp64 golden ref were not generated for resnet50_quantized_qat
ERROR:common:output 2: torch.Size([0]) != torch.Size([64])
While executing %_fused_moving_avg_obs_fq_helper_functional_1 : [#users=6] = call_function[target=torch.ops.aten._fused_moving_avg_obs_fq_helper_functional.default](args = (%mul : Tensor[size=[64, 3, 7, 7], stride=[147, 49, 7, 1]], %primals_173 : Tensor[size=[1], stride=[1]], %primals_172 : Tensor[size=[1], stride=[1]], %clone_6 : Tensor[size=[0], stride=[1]], %clone_7 : Tensor[size=[0], stride=[1]], %clone_4 : Tensor[size=[1], stride=[1]], %clone_5 : Tensor[size=[1], stride=[1]], 0.01, -128, 127, 0, True, True), kwargs = {})
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 333, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 336, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/fx/graph_module.py", line 660, in call_wrapped
return self._wrapped_call(self, *args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/fx/graph_module.py", line 279, in __call__
raise e
File "/data/users/ezyang/b/pytorch/torch/fx/graph_module.py", line 269, in __call__
return super(self.cls, obj).__call__(*args, **kwargs) # type: ignore[misc]
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "<eval_with_key>.8", line 4, in forward
def forward(self, x : torch.Tensor) -> torch.Tensor:
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1065, in compiled_function
outs = CompiledFunction.apply(*no_dupe_args_with_synthetic_bases)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 977, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 595, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 158, in run_node
check(nv, rv, lambda: f"output {i}")
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 126, in check
assert nv.size() == rv.size(), f"{desc()}: {nv.size()} != {rv.size()}"
AssertionError: output 2: torch.Size([0]) != torch.Size([64])
While executing %_fused_moving_avg_obs_fq_helper_functional_1 : [#users=6] = call_function[target=torch.ops.aten._fused_moving_avg_obs_fq_helper_functional.default](args = (%mul : Tensor[size=[64, 3, 7, 7], stride=[147, 49, 7, 1]], %primals_173 : Tensor[size=[1], stride=[1]], %primals_172 : Tensor[size=[1], stride=[1]], %clone_6 : Tensor[size=[0], stride=[1]], %clone_7 : Tensor[size=[0], stride=[1]], %clone_4 : Tensor[size=[1], stride=[1]], %clone_5 : Tensor[size=[1], stride=[1]], 0.01, -128, 127, 0, True, True), kwargs = {})
TorchDynamo optimized model failed to run because of following error
cuda train resnet50_quantized_qat FAIL
Running torchbench.py resnext50_32x4d...
cuda train resnext50_32x4d PASS
Running torchbench.py shufflenet_v2_x1_0...
cuda train shufflenet_v2_x1_0 PASS
/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/gym/core.py:317: DeprecationWarning: WARN: Initializing wrapper in old step API which returns one bool instead of two. It is recommended to set `new_step_api=True` to use new step API. This will be the default behaviour in future.
deprecation(
/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/gym/wrappers/step_api_compatibility.py:39: DeprecationWarning: WARN: Initializing environment in old step API which returns one bool instead of two. It is recommended to set `new_step_api=True` to use new step API. This will be the default behaviour in future.
deprecation(
Running torchbench.py soft_actor_critic...
cuda train soft_actor_critic PASS
Running torchbench.py speech_transformer...
ERROR:common:compile_fn raised AssertionError: While executing %self_layer_stack_0_slf_attn_attention_temperature : torch.Tensor [#users=1] = placeholder[target=self_layer_stack_0_slf_attn_attention_temperature]
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/output_graph.py", line 466, in call_user_compiler
compiled_fn = self.compiler_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/debug_utils.py", line 865, in debug_wrapper
compiled_gm = compiler_fn(gm, example_inputs, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/optimizations/training.py", line 94, in compile_fn
return cls(gm, example_inputs, fake_mode).verified_candidate()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/optimizations/training.py", line 117, in __init__
if not is_aot_autograd_safe_to_run(gm, example_inputs, fake_mode):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/optimizations/training.py", line 64, in is_aot_autograd_safe_to_run
mutated = has_mutation(gm, example_inputs, fake_mode, inputs_only=True)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/optimizations/analysis.py", line 138, in has_mutation
ShapeAliasingAndMutationProp(new_gm).run(*example_inputs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/optimizations/analysis.py", line 116, in run
super().run(*args)
File "/data/users/ezyang/b/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/optimizations/analysis.py", line 50, in run_node
result = getattr(self, n.op)(n.target, args, kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/optimizations/analysis.py", line 41, in placeholder
assert isinstance(value, torch.Tensor)
AssertionError: While executing %self_layer_stack_0_slf_attn_attention_temperature : torch.Tensor [#users=1] = placeholder[target=self_layer_stack_0_slf_attn_attention_temperature]
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 333, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 336, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/speech_transformer/speech_transformer/transformer/transformer.py", line 28, in forward
encoder_padded_outputs, *_ = self.encoder(padded_input, input_lengths)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/speech_transformer/speech_transformer/transformer/encoder.py", line 48, in forward
non_pad_mask = get_non_pad_mask(padded_input, input_lengths=input_lengths)
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/speech_transformer/speech_transformer/transformer/encoder.py", line 50, in <graph break in forward>
slf_attn_mask = get_attn_pad_mask(padded_input, input_lengths, length)
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/speech_transformer/speech_transformer/transformer/encoder.py", line 55, in <graph break in forward>
self.positional_encoding(padded_input))
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 286, in catch_errors
return callback(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 476, in _convert_frame
result = inner_convert(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 118, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 349, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 404, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 392, in transform
tracer.run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1617, in run
super().run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 483, in run
and self.step()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 453, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1679, in RETURN_VALUE
self.output.compile_subgraph(self)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/output_graph.py", line 386, in compile_subgraph
self.compile_and_call_fx_graph(tx, pass2.graph_output_vars(), root)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/output_graph.py", line 432, in compile_and_call_fx_graph
compiled_fn = self.call_user_compiler(gm)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/output_graph.py", line 475, in call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: compile_fn raised AssertionError: While executing %self_layer_stack_0_slf_attn_attention_temperature : torch.Tensor [#users=1] = placeholder[target=self_layer_stack_0_slf_attn_attention_temperature]
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
cuda train speech_transformer FAIL
Running torchbench.py squeezenet1_1...
cuda train squeezenet1_1 PASS
Running torchbench.py tacotron2...
ERROR:common:Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 333, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 336, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/tacotron2/model.py", line 505, in forward
encoder_outputs = self.encoder(embedded_inputs, text_lengths)
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/tacotron2/model.py", line 507, in <graph break in forward>
mel_outputs, gate_outputs, alignments = self.decoder(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/tacotron2/model.py", line 396, in forward
decoder_input = self.get_go_frame(memory).unsqueeze(0)
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/tacotron2/model.py", line 396, in <graph break in forward>
decoder_input = self.get_go_frame(memory).unsqueeze(0)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1125, in compiled_function
fw_outs_including_aliases.append(input_alias.as_strided(out_tensor_meta.size(), out_tensor_meta.stride(), out_tensor_meta.storage_offset()))
RuntimeError: Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
TorchDynamo optimized model failed to run because of following error
cuda train tacotron2 FAIL
Running torchbench.py timm_efficientdet...
ERROR:common:output 0: torch.Size([96, 1, 3, 3]) != torch.Size([144, 1, 5, 5])
While executing %primals_1 : [#users=0] = placeholder[target=primals_1]
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 333, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 336, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/effdet/bench.py", line 133, in forward
class_out, box_out = self.model(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/effdet/efficientdet.py", line 602, in forward
x = self.backbone(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/efficientnet.py", line 604, in forward
x = self.conv_stem(x)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/efficientnet.py", line 611, in <graph break in forward>
x = b(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/efficientnet_blocks.py", line 183, in forward
x = self.conv_dw(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/conv2d_same.py", line 30, in forward
return conv2d_same(x, self.weight, self.bias, self.stride, self.padding, self.dilation, self.groups)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/conv2d_same.py", line 13, in conv2d_same
def conv2d_same(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1065, in compiled_function
outs = CompiledFunction.apply(*no_dupe_args_with_synthetic_bases)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 977, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 595, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 158, in run_node
check(nv, rv, lambda: f"output {i}")
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 126, in check
assert nv.size() == rv.size(), f"{desc()}: {nv.size()} != {rv.size()}"
AssertionError: output 0: torch.Size([96, 1, 3, 3]) != torch.Size([144, 1, 5, 5])
While executing %primals_1 : [#users=0] = placeholder[target=primals_1]
TorchDynamo optimized model failed to run because of following error
cuda train timm_efficientdet FAIL
Running torchbench.py timm_efficientnet...
cuda train timm_efficientnet PASS
Running torchbench.py timm_regnet...
cuda train timm_regnet PASS
Running torchbench.py timm_resnest...
cuda train timm_resnest PASS
Running torchbench.py timm_vision_transformer...
cuda train timm_vision_transformer PASS
Running torchbench.py timm_vision_transformer_large...
cuda train timm_vision_transformer_large PASS
Running torchbench.py timm_vovnet...
cuda train timm_vovnet PASS
Running torchbench.py tts_angular...
ERROR:common:
from user code:
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/tts_angular/model.py", line 18, in <graph break in forward>
o, (_, _) = self.lstm(x)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1096, in run_node
return nnmodule(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/rnn.py", line 776, in forward
result = _VF.lstm(input, hx, self._flat_weights, self.bias, self.num_layers,
RuntimeError: Cannot call sizes() on tensor with symbolic sizes/strides
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1057, in get_fake_value
return wrap_fake_exception(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 741, in wrap_fake_exception
return fn()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1058, in <lambda>
lambda: run_node(tx.output, node, args, kwargs, nnmodule)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1100, in run_node
raise RuntimeError(
RuntimeError: Failed running call_module self_lstm(*(FakeTensor(FakeTensor(..., device='meta', size=(s0, s1, 40)), cuda:0),), **{}):
Cannot call sizes() on tensor with symbolic sizes/strides
(scroll up for backtrace)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 333, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 336, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/tts_angular/model.py", line 59, in forward
d = self.layers(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/tts_angular/model.py", line 17, in forward
self.lstm.flatten_parameters()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 286, in catch_errors
return callback(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 476, in _convert_frame
result = inner_convert(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 118, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 349, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 404, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 392, in transform
tracer.run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1617, in run
super().run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 483, in run
and self.step()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 453, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 287, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 910, in CALL_FUNCTION
self.call_function(fn, args, {})
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 395, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/nn_module.py", line 202, in call_function
return wrap_fx_proxy(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 652, in wrap_fx_proxy
return wrap_fx_proxy_cls(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 693, in wrap_fx_proxy_cls
example_value = get_fake_value(proxy.node, tx)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1070, in get_fake_value
raise TorchRuntimeError() from e
torch._dynamo.exc.TorchRuntimeError:
from user code:
File "/data/users/ezyang/b/torchbenchmark/torchbenchmark/models/tts_angular/model.py", line 18, in <graph break in forward>
o, (_, _) = self.lstm(x)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
cuda train tts_angular FAIL
Running torchbench.py vgg16...
cuda train vgg16 PASS
Running torchbench.py vision_maskrcnn...
ERROR:common:output 0: torch.Size([3, 427, 640]) != torch.Size([3, 612, 612])
While executing %arg0_1 : [#users=1] = placeholder[target=arg0_1]
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 333, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/torchbench.py", line 336, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/torchvision/torchvision/models/detection/generalized_rcnn.py", line 83, in forward
images, targets = self.transform(images, targets)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/torchvision/torchvision/models/detection/transform.py", line 129, in forward
image = self.normalize(image)
File "/data/users/ezyang/b/torchvision/torchvision/models/detection/transform.py", line 148, in normalize
def normalize(self, image: Tensor) -> Tensor:
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 639, in new_fn
return call_func_with_args(compiled_fw, args, disable_amp=disable_amp)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 595, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 158, in run_node
check(nv, rv, lambda: f"output {i}")
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 126, in check
assert nv.size() == rv.size(), f"{desc()}: {nv.size()} != {rv.size()}"
AssertionError: output 0: torch.Size([3, 427, 640]) != torch.Size([3, 612, 612])
While executing %arg0_1 : [#users=1] = placeholder[target=arg0_1]
TorchDynamo optimized model failed to run because of following error
cuda train vision_maskrcnn FAIL
Running torchbench.py yolov3...
[2022-11-20 11:54:00,366] torch._dynamo.convert_frame: [WARNING] torch._dynamo hit config.cache_size_limit (64)
function: 'forward' (/data/users/ezyang/b/pytorch/torch/nn/modules/container.py:202)
reasons: ['___check_obj_id(self, 140486728199232)']
to diagnose recompilation issues, see https://github.com/pytorch/torchdynamo/blob/main/TROUBLESHOOTING.md.
cuda train yolov3 PASS
Running huggingface.py AlbertForMaskedLM...
cuda train AlbertForMaskedLM PASS
Running huggingface.py AlbertForQuestionAnswering...
cuda train AlbertForQuestionAnswering PASS
Running huggingface.py AllenaiLongformerBase...
ERROR:common:Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/huggingface.py", line 482, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/huggingface.py", line 485, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1813, in forward
outputs = self.longformer(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1696, in forward
padding_len, input_ids, attention_mask, token_type_ids, position_ids, inputs_embeds = self._pad_to_window_size(
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1715, in <graph break in forward>
encoder_outputs = self.encoder(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1265, in forward
is_global_attn = is_index_global_attn.flatten().any().item()
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1297, in <graph break in forward>
layer_outputs = layer_module(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1221, in forward
self_attn_outputs = self.attention(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1157, in forward
self_outputs = self.self(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 542, in forward
def forward(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1125, in compiled_function
fw_outs_including_aliases.append(input_alias.as_strided(out_tensor_meta.size(), out_tensor_meta.stride(), out_tensor_meta.storage_offset()))
RuntimeError: Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
TorchDynamo optimized model failed to run because of following error
cuda train AllenaiLongformerBase FAIL
Running huggingface.py BartForCausalLM...
cuda train BartForCausalLM PASS
Running huggingface.py BartForConditionalGeneration...
cuda train BartForConditionalGeneration PASS
Running huggingface.py BertForMaskedLM...
cuda train BertForMaskedLM PASS
Running huggingface.py BertForQuestionAnswering...
cuda train BertForQuestionAnswering PASS
Running huggingface.py BlenderbotForCausalLM...
cuda train BlenderbotForCausalLM PASS
Running huggingface.py BlenderbotSmallForCausalLM...
cuda train BlenderbotSmallForCausalLM PASS
Running huggingface.py BlenderbotSmallForConditionalGeneration...
cuda train BlenderbotSmallForConditionalGeneration PASS
Running huggingface.py CamemBert...
cuda train CamemBert PASS
Running huggingface.py DebertaForMaskedLM...
cuda train DebertaForMaskedLM PASS
Running huggingface.py DebertaForQuestionAnswering...
cuda train DebertaForQuestionAnswering PASS
Running huggingface.py DebertaV2ForMaskedLM...
cuda train DebertaV2ForMaskedLM PASS
Running huggingface.py DebertaV2ForQuestionAnswering...
cuda train DebertaV2ForQuestionAnswering PASS
WARNING:__main__:Sequence Length not defined for DistilBertForMaskedLM. Choosing 128 arbitrarily
Running huggingface.py DistilBertForMaskedLM...
cuda train DistilBertForMaskedLM PASS
WARNING:__main__:Sequence Length not defined for DistilBertForQuestionAnswering. Choosing 128 arbitrarily
Running huggingface.py DistilBertForQuestionAnswering...
cuda train DistilBertForQuestionAnswering PASS
Running huggingface.py DistillGPT2...
cuda train DistillGPT2 PASS
If you want to use `ElectraForCausalLM` as a standalone, add `is_decoder=True.`
Running huggingface.py ElectraForCausalLM...
cuda train ElectraForCausalLM PASS
Running huggingface.py ElectraForQuestionAnswering...
cuda train ElectraForQuestionAnswering PASS
Running huggingface.py GPT2ForSequenceClassification...
cuda train GPT2ForSequenceClassification PASS
Running huggingface.py GoogleFnet...
cuda train GoogleFnet PASS
Running huggingface.py LayoutLMForMaskedLM...
cuda train LayoutLMForMaskedLM PASS
Running huggingface.py LayoutLMForSequenceClassification...
cuda train LayoutLMForSequenceClassification PASS
WARNING:__main__:Sequence Length not defined for M2M100ForConditionalGeneration. Choosing 128 arbitrarily
Running huggingface.py M2M100ForConditionalGeneration...
ERROR:common:Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/huggingface.py", line 482, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/huggingface.py", line 485, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1317, in forward
outputs = self.model(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 1190, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/m2m_100/modeling_m2m_100.py", line 716, in forward
def forward(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1125, in compiled_function
fw_outs_including_aliases.append(input_alias.as_strided(out_tensor_meta.size(), out_tensor_meta.stride(), out_tensor_meta.storage_offset()))
RuntimeError: Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
TorchDynamo optimized model failed to run because of following error
cuda train M2M100ForConditionalGeneration FAIL
Running huggingface.py MBartForCausalLM...
cuda train MBartForCausalLM PASS
Running huggingface.py MBartForConditionalGeneration...
cuda train MBartForConditionalGeneration PASS
WARNING:__main__:Sequence Length not defined for MT5ForConditionalGeneration. Choosing 128 arbitrarily
Running huggingface.py MT5ForConditionalGeneration...
WARNING:common:fp64 golden ref were not generated for MT5ForConditionalGeneration
ERROR:common:
from user code:
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 529, in <graph break in forward>
scores += position_bias
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1091, in run_node
return node.target(*args, **kwargs)
RuntimeError: Output 0 of AsStridedBackward0 is a view of a view which was created in no_grad mode and is being modified inplace with grad mode enabled. Given that this use case is ambiguous and error-prone, it is forbidden. You can clarify your code by moving both the view and the inplace either both inside the no_grad block (if you don't want the inplace to be tracked) or both outside (if you want the inplace to be tracked).
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1057, in get_fake_value
return wrap_fake_exception(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 741, in wrap_fake_exception
return fn()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1058, in <lambda>
lambda: run_node(tx.output, node, args, kwargs, nnmodule)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1100, in run_node
raise RuntimeError(
RuntimeError: Failed running call_function <built-in function iadd>(*(FakeTensor(FakeTensor(..., device='meta', size=(1, s1, s0, s0),
grad_fn=<AsStridedBackward0>), cuda:0), FakeTensor(FakeTensor(..., device='meta', size=(1, s1, s0, s0), grad_fn=<AddBackward0>), cuda:0)), **{}):
Output 0 of AsStridedBackward0 is a view of a view which was created in no_grad mode and is being modified inplace with grad mode enabled. Given that this use case is ambiguous and error-prone, it is forbidden. You can clarify your code by moving both the view and the inplace either both inside the no_grad block (if you don't want the inplace to be tracked) or both outside (if you want the inplace to be tracked).
(scroll up for backtrace)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/huggingface.py", line 482, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/huggingface.py", line 485, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 945, in forward
attention_mask = torch.ones(batch_size, mask_seq_length).to(inputs_embeds.device)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in <graph break in forward>
layer_outputs = layer_module(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 286, in catch_errors
return callback(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 476, in _convert_frame
result = inner_convert(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 118, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 349, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 404, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 392, in transform
tracer.run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1617, in run
super().run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 483, in run
and self.step()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 453, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 120, in impl
self.push(fn_var.call_function(self, self.popn(nargs), {}))
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builtin.py", line 320, in call_function
return wrap_fx_proxy(tx, proxy, **options)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 652, in wrap_fx_proxy
return wrap_fx_proxy_cls(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 693, in wrap_fx_proxy_cls
example_value = get_fake_value(proxy.node, tx)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1070, in get_fake_value
raise TorchRuntimeError() from e
torch._dynamo.exc.TorchRuntimeError:
from user code:
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 529, in <graph break in forward>
scores += position_bias
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
cuda train MT5ForConditionalGeneration FAIL
If you want to use `MegatronBertForCausalLM` as a standalone, add `is_decoder=True.`
Running huggingface.py MegatronBertForCausalLM...
cuda train MegatronBertForCausalLM PASS
Running huggingface.py MegatronBertForQuestionAnswering...
cuda train MegatronBertForQuestionAnswering PASS
Running huggingface.py MobileBertForMaskedLM...
cuda train MobileBertForMaskedLM PASS
Running huggingface.py MobileBertForQuestionAnswering...
cuda train MobileBertForQuestionAnswering PASS
Running huggingface.py OPTForCausalLM...
cuda train OPTForCausalLM PASS
Running huggingface.py PLBartForCausalLM...
cuda train PLBartForCausalLM PASS
Running huggingface.py PLBartForConditionalGeneration...
cuda train PLBartForConditionalGeneration PASS
WARNING:__main__:Sequence Length not defined for PegasusForCausalLM. Choosing 128 arbitrarily
Running huggingface.py PegasusForCausalLM...
cuda train PegasusForCausalLM PASS
WARNING:__main__:Sequence Length not defined for PegasusForConditionalGeneration. Choosing 128 arbitrarily
Running huggingface.py PegasusForConditionalGeneration...
cuda train PegasusForConditionalGeneration PASS
If you want to use `RobertaLMHeadModel` as a standalone, add `is_decoder=True.`
Running huggingface.py RobertaForCausalLM...
cuda train RobertaForCausalLM PASS
Running huggingface.py RobertaForQuestionAnswering...
cuda train RobertaForQuestionAnswering PASS
WARNING:__main__:Sequence Length not defined for Speech2Text2ForCausalLM. Choosing 128 arbitrarily
Running huggingface.py Speech2Text2ForCausalLM...
ERROR:common:Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/huggingface.py", line 482, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/huggingface.py", line 485, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/speech_to_text_2/modeling_speech_to_text_2.py", line 910, in forward
outputs = self.model.decoder(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/speech_to_text_2/modeling_speech_to_text_2.py", line 509, in forward
def forward(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1125, in compiled_function
fw_outs_including_aliases.append(input_alias.as_strided(out_tensor_meta.size(), out_tensor_meta.stride(), out_tensor_meta.storage_offset()))
RuntimeError: Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
TorchDynamo optimized model failed to run because of following error
cuda train Speech2Text2ForCausalLM FAIL
Running huggingface.py T5ForConditionalGeneration...
WARNING:common:fp64 golden ref were not generated for T5ForConditionalGeneration
ERROR:common:
from user code:
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 529, in <graph break in forward>
scores += position_bias
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1091, in run_node
return node.target(*args, **kwargs)
RuntimeError: Output 0 of AsStridedBackward0 is a view of a view which was created in no_grad mode and is being modified inplace with grad mode enabled. Given that this use case is ambiguous and error-prone, it is forbidden. You can clarify your code by moving both the view and the inplace either both inside the no_grad block (if you don't want the inplace to be tracked) or both outside (if you want the inplace to be tracked).
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1057, in get_fake_value
return wrap_fake_exception(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 741, in wrap_fake_exception
return fn()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1058, in <lambda>
lambda: run_node(tx.output, node, args, kwargs, nnmodule)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1100, in run_node
raise RuntimeError(
RuntimeError: Failed running call_function <built-in function iadd>(*(FakeTensor(FakeTensor(..., device='meta', size=(1, s1, s0, s0),
grad_fn=<AsStridedBackward0>), cuda:0), FakeTensor(FakeTensor(..., device='meta', size=(1, s1, s0, s0), grad_fn=<AddBackward0>), cuda:0)), **{}):
Output 0 of AsStridedBackward0 is a view of a view which was created in no_grad mode and is being modified inplace with grad mode enabled. Given that this use case is ambiguous and error-prone, it is forbidden. You can clarify your code by moving both the view and the inplace either both inside the no_grad block (if you don't want the inplace to be tracked) or both outside (if you want the inplace to be tracked).
(scroll up for backtrace)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/huggingface.py", line 482, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/huggingface.py", line 485, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 945, in forward
attention_mask = torch.ones(batch_size, mask_seq_length).to(inputs_embeds.device)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in <graph break in forward>
layer_outputs = layer_module(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 286, in catch_errors
return callback(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 476, in _convert_frame
result = inner_convert(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 118, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 349, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 404, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 392, in transform
tracer.run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1617, in run
super().run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 483, in run
and self.step()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 453, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 120, in impl
self.push(fn_var.call_function(self, self.popn(nargs), {}))
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builtin.py", line 320, in call_function
return wrap_fx_proxy(tx, proxy, **options)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 652, in wrap_fx_proxy
return wrap_fx_proxy_cls(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 693, in wrap_fx_proxy_cls
example_value = get_fake_value(proxy.node, tx)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1070, in get_fake_value
raise TorchRuntimeError() from e
torch._dynamo.exc.TorchRuntimeError:
from user code:
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 529, in <graph break in forward>
scores += position_bias
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
cuda train T5ForConditionalGeneration FAIL
Running huggingface.py T5Small...
WARNING:common:fp64 golden ref were not generated for T5Small
ERROR:common:
from user code:
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 529, in <graph break in forward>
scores += position_bias
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1091, in run_node
return node.target(*args, **kwargs)
RuntimeError: Output 0 of AsStridedBackward0 is a view of a view which was created in no_grad mode and is being modified inplace with grad mode enabled. Given that this use case is ambiguous and error-prone, it is forbidden. You can clarify your code by moving both the view and the inplace either both inside the no_grad block (if you don't want the inplace to be tracked) or both outside (if you want the inplace to be tracked).
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1057, in get_fake_value
return wrap_fake_exception(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 741, in wrap_fake_exception
return fn()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1058, in <lambda>
lambda: run_node(tx.output, node, args, kwargs, nnmodule)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1100, in run_node
raise RuntimeError(
RuntimeError: Failed running call_function <built-in function iadd>(*(FakeTensor(FakeTensor(..., device='meta', size=(1, s1, s0, s0),
grad_fn=<AsStridedBackward0>), cuda:0), FakeTensor(FakeTensor(..., device='meta', size=(1, s1, s0, s0), grad_fn=<AddBackward0>), cuda:0)), **{}):
Output 0 of AsStridedBackward0 is a view of a view which was created in no_grad mode and is being modified inplace with grad mode enabled. Given that this use case is ambiguous and error-prone, it is forbidden. You can clarify your code by moving both the view and the inplace either both inside the no_grad block (if you don't want the inplace to be tracked) or both outside (if you want the inplace to be tracked).
(scroll up for backtrace)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/huggingface.py", line 482, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/huggingface.py", line 485, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1601, in forward
encoder_outputs = self.encoder(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 945, in forward
attention_mask = torch.ones(batch_size, mask_seq_length).to(inputs_embeds.device)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1033, in <graph break in forward>
layer_outputs = layer_module(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 664, in forward
self_attention_outputs = self.layer[0](
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 570, in forward
attention_output = self.SelfAttention(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 519, in forward
position_bias = self.compute_bias(real_seq_length, key_length, device=scores.device)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 286, in catch_errors
return callback(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 476, in _convert_frame
result = inner_convert(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 118, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 349, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 404, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 392, in transform
tracer.run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1617, in run
super().run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 483, in run
and self.step()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 453, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 120, in impl
self.push(fn_var.call_function(self, self.popn(nargs), {}))
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builtin.py", line 320, in call_function
return wrap_fx_proxy(tx, proxy, **options)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 652, in wrap_fx_proxy
return wrap_fx_proxy_cls(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 693, in wrap_fx_proxy_cls
example_value = get_fake_value(proxy.node, tx)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1070, in get_fake_value
raise TorchRuntimeError() from e
torch._dynamo.exc.TorchRuntimeError:
from user code:
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 529, in <graph break in forward>
scores += position_bias
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
cuda train T5Small FAIL
Running huggingface.py TrOCRForCausalLM...
cuda train TrOCRForCausalLM PASS
WARNING:__main__:Sequence Length not defined for XGLMForCausalLM. Choosing 128 arbitrarily
Running huggingface.py XGLMForCausalLM...
ERROR:common:Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/huggingface.py", line 482, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/huggingface.py", line 485, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/xglm/modeling_xglm.py", line 889, in forward
outputs = self.model(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/xglm/modeling_xglm.py", line 591, in forward
@add_start_docstrings_to_model_forward(XGLM_INPUTS_DOCSTRING)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1125, in compiled_function
fw_outs_including_aliases.append(input_alias.as_strided(out_tensor_meta.size(), out_tensor_meta.stride(), out_tensor_meta.storage_offset()))
RuntimeError: Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
TorchDynamo optimized model failed to run because of following error
cuda train XGLMForCausalLM FAIL
Running huggingface.py XLNetLMHeadModel...
ERROR:common:Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/huggingface.py", line 482, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/huggingface.py", line 485, in <graph break in forward_and_backward_pass>
pred = mod(**cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1448, in forward
transformer_outputs = self.transformer(
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1207, in forward
pos_emb = self.relative_positional_encoding(qlen, klen, bsz=bsz)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 1237, in <graph break in forward>
new_mems = new_mems + (self.cache_mem(output_h, mems[i]),)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/transformers/models/xlnet/modeling_xlnet.py", line 991, in cache_mem
def cache_mem(self, curr_out, prev_mem):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1125, in compiled_function
fw_outs_including_aliases.append(input_alias.as_strided(out_tensor_meta.size(), out_tensor_meta.stride(), out_tensor_meta.storage_offset()))
RuntimeError: Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
TorchDynamo optimized model failed to run because of following error
cuda train XLNetLMHeadModel FAIL
Running huggingface.py YituTechConvBert...
cuda train YituTechConvBert PASS
Running timm_models.py adv_inception_v3...
cuda train adv_inception_v3 PASS
Running timm_models.py beit_base_patch16_224...
cuda train beit_base_patch16_224 PASS
Running timm_models.py botnet26t_256...
ERROR:common:Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 305, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 308, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/byobnet.py", line 1559, in forward
x = self.forward_features(x)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/byobnet.py", line 1551, in forward_features
x = self.stages(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/byobnet.py", line 1245, in forward
x = self.self_attn(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/bottleneck_attn.py", line 137, in forward
_assert(H == self.pos_embed.height, '')
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/bottleneck_attn.py", line 152, in <graph break in forward>
attn = (q @ k) * self.scale + self.pos_embed(q)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/bottleneck_attn.py", line 68, in forward
def forward(self, q):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1125, in compiled_function
fw_outs_including_aliases.append(input_alias.as_strided(out_tensor_meta.size(), out_tensor_meta.stride(), out_tensor_meta.storage_offset()))
RuntimeError: Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
TorchDynamo optimized model failed to run because of following error
cuda train botnet26t_256 FAIL
Running timm_models.py cait_m36_384...
cuda train cait_m36_384 PASS
Running timm_models.py coat_lite_mini...
cuda train coat_lite_mini PASS
Running timm_models.py convit_base...
WARNING:common:fp64 golden ref were not generated for convit_base
ERROR:common:
from user code:
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/convit.py", line 138, in get_rel_indices
indy = ind.repeat_interleave(img_size, dim=0).repeat_interleave(img_size, dim=1)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1093, in run_node
return getattr(args[0], node.target)(*args[1:], **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_subclasses/fake_tensor.py", line 867, in __torch_dispatch__
op_impl_out = op_impl(self, func, *args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_subclasses/fake_tensor.py", line 362, in dyn_shape
raise DynamicOutputShapeException(func)
torch._subclasses.fake_tensor.DynamicOutputShapeException: aten.repeat_interleave.Tensor
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1057, in get_fake_value
return wrap_fake_exception(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 741, in wrap_fake_exception
return fn()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1058, in <lambda>
lambda: run_node(tx.output, node, args, kwargs, nnmodule)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1100, in run_node
raise RuntimeError(
RuntimeError: Failed running call_method repeat_interleave(*(FakeTensor(FakeTensor(..., device='meta', size=(14, 14), dtype=torch.int64), cpu), 14), **{'dim': 0}):
aten.repeat_interleave.Tensor
(scroll up for backtrace)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 305, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 308, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/convit.py", line 333, in forward
x = self.forward_features(x)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/convit.py", line 315, in forward_features
x = self.patch_embed(x)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/convit.py", line 323, in <graph break in forward_features>
x = blk(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/convit.py", line 214, in forward
x = x + self.drop_path(self.attn(self.norm1(x)))
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/convit.py", line 86, in forward
self.rel_indices = self.get_rel_indices(N)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 286, in catch_errors
return callback(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 476, in _convert_frame
result = inner_convert(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 118, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 349, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 404, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 392, in transform
tracer.run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1617, in run
super().run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 483, in run
and self.step()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 453, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 287, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 959, in CALL_FUNCTION_KW
self.call_function(fn, args, kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 395, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/misc.py", line 582, in call_function
return self.obj.call_method(tx, self.name, args, kwargs).add_options(self)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/tensor.py", line 327, in call_method
return wrap_fx_proxy(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 652, in wrap_fx_proxy
return wrap_fx_proxy_cls(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 693, in wrap_fx_proxy_cls
example_value = get_fake_value(proxy.node, tx)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1070, in get_fake_value
raise TorchRuntimeError() from e
torch._dynamo.exc.TorchRuntimeError:
from user code:
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/convit.py", line 138, in get_rel_indices
indy = ind.repeat_interleave(img_size, dim=0).repeat_interleave(img_size, dim=1)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
cuda train convit_base FAIL
Running timm_models.py convmixer_768_32...
cuda train convmixer_768_32 PASS
Running timm_models.py convnext_base...
cuda train convnext_base PASS
Running timm_models.py crossvit_9_240...
ERROR:common:
from user code:
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/crossvit.py", line 281, in scale_image
x = torch.nn.functional.interpolate(x, size=ss, mode='bicubic', align_corners=False)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1091, in run_node
return node.target(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/functional.py", line 3966, in interpolate
return torch._C._nn.upsample_bicubic2d(input, output_size, align_corners, scale_factors)
RuntimeError: Cannot call sizes() on tensor with symbolic sizes/strides
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1057, in get_fake_value
return wrap_fake_exception(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 741, in wrap_fake_exception
return fn()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1058, in <lambda>
lambda: run_node(tx.output, node, args, kwargs, nnmodule)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1100, in run_node
raise RuntimeError(
RuntimeError: Failed running call_function <function interpolate at 0x7f786589a160>(*(FakeTensor(FakeTensor(..., device='meta', size=(s0, s1, s2, s2)), cuda:0),), **{'size': (224, 224), 'mode': 'bicubic', 'align_corners': False}):
Cannot call sizes() on tensor with symbolic sizes/strides
(scroll up for backtrace)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 305, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 308, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/crossvit.py", line 418, in forward
xs = self.forward_features(x)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/crossvit.py", line 394, in forward_features
x_ = scale_image(x_, ss, self.crop_scale)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 286, in catch_errors
return callback(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 476, in _convert_frame
result = inner_convert(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 118, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 349, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 404, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 392, in transform
tracer.run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1617, in run
super().run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 483, in run
and self.step()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 453, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 287, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 959, in CALL_FUNCTION_KW
self.call_function(fn, args, kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 395, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/torch.py", line 430, in call_function
tensor_variable = wrap_fx_proxy(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 652, in wrap_fx_proxy
return wrap_fx_proxy_cls(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 693, in wrap_fx_proxy_cls
example_value = get_fake_value(proxy.node, tx)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1070, in get_fake_value
raise TorchRuntimeError() from e
torch._dynamo.exc.TorchRuntimeError:
from user code:
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/crossvit.py", line 281, in scale_image
x = torch.nn.functional.interpolate(x, size=ss, mode='bicubic', align_corners=False)
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
cuda train crossvit_9_240 FAIL
Running timm_models.py cspdarknet53...
cuda train cspdarknet53 PASS
Running timm_models.py deit_base_distilled_patch16_224...
cuda train deit_base_distilled_patch16_224 PASS
Running timm_models.py dla102...
cuda train dla102 PASS
Running timm_models.py dm_nfnet_f0...
ERROR:common:output 0: torch.Size([2, 64, 96, 96]) != torch.Size([2, 256, 48, 48])
While executing %primals_1 : [#users=2] = placeholder[target=primals_1]
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 305, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 308, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/nfnet.py", line 583, in forward
x = self.forward_features(x)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/nfnet.py", line 570, in forward_features
x = self.stem(x)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/nfnet.py", line 574, in <graph break in forward_features>
x = self.stages(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/nfnet.py", line 354, in forward
out = self.conv2(self.act2(out))
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/std_conv.py", line 128, in forward
x = pad_same(x, self.kernel_size, self.stride, self.dilation)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/padding.py", line 28, in pad_same
def pad_same(x, k: List[int], s: List[int], d: List[int] = (1, 1), value: float = 0):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1065, in compiled_function
outs = CompiledFunction.apply(*no_dupe_args_with_synthetic_bases)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 977, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 595, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 158, in run_node
check(nv, rv, lambda: f"output {i}")
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 126, in check
assert nv.size() == rv.size(), f"{desc()}: {nv.size()} != {rv.size()}"
AssertionError: output 0: torch.Size([2, 64, 96, 96]) != torch.Size([2, 256, 48, 48])
While executing %primals_1 : [#users=2] = placeholder[target=primals_1]
TorchDynamo optimized model failed to run because of following error
cuda train dm_nfnet_f0 FAIL
Running timm_models.py dpn107...
cuda train dpn107 PASS
Running timm_models.py eca_botnext26ts_256...
ERROR:common:Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 305, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 308, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/byobnet.py", line 1559, in forward
x = self.forward_features(x)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/byobnet.py", line 1551, in forward_features
x = self.stages(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/byobnet.py", line 1245, in forward
x = self.self_attn(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/bottleneck_attn.py", line 137, in forward
_assert(H == self.pos_embed.height, '')
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/bottleneck_attn.py", line 152, in <graph break in forward>
attn = (q @ k) * self.scale + self.pos_embed(q)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/bottleneck_attn.py", line 68, in forward
def forward(self, q):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1125, in compiled_function
fw_outs_including_aliases.append(input_alias.as_strided(out_tensor_meta.size(), out_tensor_meta.stride(), out_tensor_meta.storage_offset()))
RuntimeError: Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
TorchDynamo optimized model failed to run because of following error
cuda train eca_botnext26ts_256 FAIL
Running timm_models.py eca_halonext26ts...
ERROR:common:output 0: torch.Size([64, 8, 8, 16]) != torch.Size([64, 4, 4, 16])
While executing %primals_1 : [#users=5] = placeholder[target=primals_1]
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 305, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 308, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/byobnet.py", line 1559, in forward
x = self.forward_features(x)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/byobnet.py", line 1551, in forward_features
x = self.stages(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/byobnet.py", line 1245, in forward
x = self.self_attn(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/halo_attn.py", line 171, in forward
_assert(H % self.block_size == 0, '')
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/halo_attn.py", line 199, in <graph break in forward>
attn = (q @ k.transpose(-1, -2)) * self.scale + self.pos_embed(q)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/halo_attn.py", line 86, in forward
rel_logits_w = rel_logits_1d(q, self.width_rel, permute_mask=(0, 1, 3, 2, 4))
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/halo_attn.py", line 30, in rel_logits_1d
def rel_logits_1d(q, rel_k, permute_mask: List[int]):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1065, in compiled_function
outs = CompiledFunction.apply(*no_dupe_args_with_synthetic_bases)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 977, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 595, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 158, in run_node
check(nv, rv, lambda: f"output {i}")
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 126, in check
assert nv.size() == rv.size(), f"{desc()}: {nv.size()} != {rv.size()}"
AssertionError: output 0: torch.Size([64, 8, 8, 16]) != torch.Size([64, 4, 4, 16])
While executing %primals_1 : [#users=5] = placeholder[target=primals_1]
TorchDynamo optimized model failed to run because of following error
cuda train eca_halonext26ts FAIL
Running timm_models.py ese_vovnet19b_dw...
cuda train ese_vovnet19b_dw PASS
Running timm_models.py fbnetc_100...
cuda train fbnetc_100 PASS
Running timm_models.py fbnetv3_b...
cuda train fbnetv3_b PASS
Running timm_models.py gernet_l...
cuda train gernet_l PASS
Running timm_models.py ghostnet_100...
cuda train ghostnet_100 PASS
Running timm_models.py gluon_inception_v3...
cuda train gluon_inception_v3 PASS
Running timm_models.py gluon_xception65...
cuda train gluon_xception65 PASS
Running timm_models.py gmixer_24_224...
cuda train gmixer_24_224 PASS
Running timm_models.py gmlp_s16_224...
cuda train gmlp_s16_224 PASS
Running timm_models.py hrnet_w18...
ERROR:common:
from user code:
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/hrnet.py", line 713, in stages
yl = self.stage2(xl)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/hrnet.py", line 495, in forward
y = y + fuse_outer[j](x[j])
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1096, in run_node
return nnmodule(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/upsampling.py", line 156, in forward
return F.interpolate(input, self.size, self.scale_factor, self.mode, self.align_corners,
File "/data/users/ezyang/b/pytorch/torch/nn/functional.py", line 3930, in interpolate
return torch._C._nn.upsample_nearest2d(input, output_size, scale_factors)
RuntimeError: Cannot call sizes() on tensor with symbolic sizes/strides
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1057, in get_fake_value
return wrap_fake_exception(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 741, in wrap_fake_exception
return fn()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1058, in <lambda>
lambda: run_node(tx.output, node, args, kwargs, nnmodule)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1100, in run_node
raise RuntimeError(
RuntimeError: Failed running call_module sub1_1_2(*(FakeTensor(FakeTensor(..., device='meta', size=(s0, 18, (s2 - 1)//8 + 1, (s2 - 1)//8 + 1),
grad_fn=<NativeBatchNormBackward0>), cuda:0),), **{}):
Cannot call sizes() on tensor with symbolic sizes/strides
(scroll up for backtrace)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 305, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 308, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 286, in catch_errors
return callback(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 476, in _convert_frame
result = inner_convert(frame, cache_size)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 118, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 90, in time_wrapper
r = func(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 349, in _convert_frame_assert
return _compile(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 404, in _compile
out_code = transform_code_object(code, transform)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/bytecode_transformation.py", line 341, in transform_code_object
transformations(instructions, code_options)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/convert_frame.py", line 392, in transform
tracer.run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1617, in run
super().run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 483, in run
and self.step()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 453, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 287, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 910, in CALL_FUNCTION
self.call_function(fn, args, {})
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 395, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/functions.py", line 226, in call_function
return super().call_function(tx, args, kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/functions.py", line 196, in call_function
return super(UserFunctionVariable, self).call_function(tx, args, kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/functions.py", line 67, in call_function
return tx.inline_user_function_return(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 424, in inline_user_function_return
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1689, in inline_call
return cls.inline_call_(parent, func, args, kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1743, in inline_call_
tracer.run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 483, in run
and self.step()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 453, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 287, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 910, in CALL_FUNCTION
self.call_function(fn, args, {})
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 395, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/functions.py", line 226, in call_function
return super().call_function(tx, args, kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/functions.py", line 196, in call_function
return super(UserFunctionVariable, self).call_function(tx, args, kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/functions.py", line 67, in call_function
return tx.inline_user_function_return(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 424, in inline_user_function_return
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1689, in inline_call
return cls.inline_call_(parent, func, args, kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1743, in inline_call_
tracer.run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 483, in run
and self.step()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 453, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 287, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 910, in CALL_FUNCTION
self.call_function(fn, args, {})
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 395, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/nn_module.py", line 183, in call_function
tx.call_function(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 395, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/nn_module.py", line 222, in call_function
return tx.inline_user_function_return(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 424, in inline_user_function_return
result = InliningInstructionTranslator.inline_call(self, fn, args, kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1689, in inline_call
return cls.inline_call_(parent, func, args, kwargs)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 1743, in inline_call_
tracer.run()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 483, in run
and self.step()
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 453, in step
getattr(self, inst.opname)(inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 287, in wrapper
return inner_fn(self, inst)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 910, in CALL_FUNCTION
self.call_function(fn, args, {})
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 395, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/nn_module.py", line 183, in call_function
tx.call_function(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/symbolic_convert.py", line 395, in call_function
self.push(fn.call_function(self, args, kwargs))
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/nn_module.py", line 202, in call_function
return wrap_fx_proxy(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 652, in wrap_fx_proxy
return wrap_fx_proxy_cls(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/variables/builder.py", line 693, in wrap_fx_proxy_cls
example_value = get_fake_value(proxy.node, tx)
File "/data/users/ezyang/b/pytorch/torch/_dynamo/utils.py", line 1070, in get_fake_value
raise TorchRuntimeError() from e
torch._dynamo.exc.TorchRuntimeError:
from user code:
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/hrnet.py", line 713, in stages
yl = self.stage2(xl)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/hrnet.py", line 495, in forward
y = y + fuse_outer[j](x[j])
Set torch._dynamo.config.verbose=True for more information
You can suppress this exception and fall back to eager by setting:
torch._dynamo.config.suppress_errors = True
TorchDynamo optimized model failed to run because of following error
cuda train hrnet_w18 FAIL
Running timm_models.py inception_v3...
cuda train inception_v3 PASS
Running timm_models.py jx_nest_base...
ERROR:common:Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 305, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 308, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/nest.py", line 372, in forward
x = self.forward_features(x)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/nest.py", line 359, in forward_features
x = self.patch_embed(x)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/nest.py", line 360, in <graph break in forward_features>
x = self.levels(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/nest.py", line 201, in forward
def forward(self, x):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1125, in compiled_function
fw_outs_including_aliases.append(input_alias.as_strided(out_tensor_meta.size(), out_tensor_meta.stride(), out_tensor_meta.storage_offset()))
RuntimeError: Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
TorchDynamo optimized model failed to run because of following error
cuda train jx_nest_base FAIL
Running timm_models.py lcnet_050...
cuda train lcnet_050 PASS
Running timm_models.py levit_128...
WARNING:common:fp64 golden ref were not generated for levit_128
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 322, in <module>
main(TimmRunnner())
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1651, in main
return maybe_fresh_cache(run, args.cold_start_latency and args.only)(
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 810, in inner
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1990, in run
runner.run_one_model(
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1336, in run_one_model
status = self.check_accuracy(
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1132, in check_accuracy
correct_result = self.run_n_iterations(
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1056, in run_n_iterations
return self.model_iter_fn(mod, inputs, collect_outputs=True)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 312, in forward_and_backward_pass
self.grad_scaler.scale(loss).backward()
File "/data/users/ezyang/b/pytorch/torch/_tensor.py", line 473, in backward
torch.autograd.backward(
File "/data/users/ezyang/b/pytorch/torch/autograd/__init__.py", line 197, in backward
Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
RuntimeError: Trying to backward through the graph a second time (or directly access saved tensors after they have already been freed). Saved intermediate values of the graph are freed when you call .backward() or autograd.grad(). Specify retain_graph=True if you need to backward through the graph a second time or if you need to access saved tensors after calling backward.
cuda train levit_128 FAIL
Running timm_models.py mixer_b16_224...
cuda train mixer_b16_224 PASS
Running timm_models.py mixnet_l...
cuda train mixnet_l PASS
Running timm_models.py mnasnet_100...
cuda train mnasnet_100 PASS
Running timm_models.py mobilenetv2_100...
cuda train mobilenetv2_100 PASS
Running timm_models.py mobilenetv3_large_100...
cuda train mobilenetv3_large_100 PASS
Running timm_models.py mobilevit_s...
cuda train mobilevit_s PASS
Running timm_models.py nfnet_l0...
cuda train nfnet_l0 PASS
Running timm_models.py pit_b_224...
cuda train pit_b_224 PASS
Running timm_models.py pnasnet5large...
ERROR:common:output 0: torch.Size([96, 1, 5, 5]) != torch.Size([54, 1, 7, 7])
While executing %primals_1 : [#users=0] = placeholder[target=primals_1]
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 305, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 308, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/pnasnet.py", line 342, in forward
x = self.forward_features(x)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/pnasnet.py", line 318, in forward_features
x_stem_0 = self.cell_stem_0(x_conv_0)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/pnasnet.py", line 182, in forward
x_out = self.cell_forward(x_left, x_right)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/pnasnet.py", line 121, in cell_forward
x_comb_iter_0_left = self.comb_iter_0_left(x_left)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/pnasnet.py", line 122, in <graph break in cell_forward>
x_comb_iter_0_right = self.comb_iter_0_right(x_left)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/pnasnet.py", line 125, in <graph break in cell_forward>
x_comb_iter_1_left = self.comb_iter_1_left(x_right)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/pnasnet.py", line 70, in forward
x = self.separable_1(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/pnasnet.py", line 49, in forward
x = self.depthwise_conv2d(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/conv2d_same.py", line 30, in forward
return conv2d_same(x, self.weight, self.bias, self.stride, self.padding, self.dilation, self.groups)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/conv2d_same.py", line 13, in conv2d_same
def conv2d_same(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1065, in compiled_function
outs = CompiledFunction.apply(*no_dupe_args_with_synthetic_bases)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 977, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 595, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 158, in run_node
check(nv, rv, lambda: f"output {i}")
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 126, in check
assert nv.size() == rv.size(), f"{desc()}: {nv.size()} != {rv.size()}"
AssertionError: output 0: torch.Size([96, 1, 5, 5]) != torch.Size([54, 1, 7, 7])
While executing %primals_1 : [#users=0] = placeholder[target=primals_1]
TorchDynamo optimized model failed to run because of following error
cuda train pnasnet5large FAIL
Running timm_models.py poolformer_m36...
cuda train poolformer_m36 PASS
Running timm_models.py regnety_002...
cuda train regnety_002 PASS
Running timm_models.py repvgg_a2...
cuda train repvgg_a2 PASS
Running timm_models.py res2net101_26w_4s...
cuda train res2net101_26w_4s PASS
Running timm_models.py res2net50_14w_8s...
cuda train res2net50_14w_8s PASS
Running timm_models.py res2next50...
cuda train res2next50 PASS
Running timm_models.py resmlp_12_224...
cuda train resmlp_12_224 PASS
Running timm_models.py resnest101e...
cuda train resnest101e PASS
Running timm_models.py rexnet_100...
cuda train rexnet_100 PASS
Running timm_models.py sebotnet33ts_256...
ERROR:common:Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 305, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 308, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/byobnet.py", line 1559, in forward
x = self.forward_features(x)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/byobnet.py", line 1551, in forward_features
x = self.stages(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/byobnet.py", line 1245, in forward
x = self.self_attn(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/bottleneck_attn.py", line 137, in forward
_assert(H == self.pos_embed.height, '')
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/bottleneck_attn.py", line 152, in <graph break in forward>
attn = (q @ k) * self.scale + self.pos_embed(q)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/bottleneck_attn.py", line 68, in forward
def forward(self, q):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1125, in compiled_function
fw_outs_including_aliases.append(input_alias.as_strided(out_tensor_meta.size(), out_tensor_meta.stride(), out_tensor_meta.storage_offset()))
RuntimeError: Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
TorchDynamo optimized model failed to run because of following error
cuda train sebotnet33ts_256 FAIL
Running timm_models.py selecsls42b...
cuda train selecsls42b PASS
Running timm_models.py spnasnet_100...
cuda train spnasnet_100 PASS
Running timm_models.py swin_base_patch4_window7_224...
cuda train swin_base_patch4_window7_224 PASS
Running timm_models.py swsl_resnext101_32x16d...
cuda train swsl_resnext101_32x16d PASS
Running timm_models.py tf_efficientnet_b0...
ERROR:common:output 0: torch.Size([96, 1, 3, 3]) != torch.Size([144, 1, 5, 5])
While executing %primals_1 : [#users=0] = placeholder[target=primals_1]
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 305, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 308, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/efficientnet.py", line 557, in forward
x = self.forward_features(x)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/efficientnet.py", line 540, in forward_features
x = self.conv_stem(x)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/efficientnet.py", line 545, in <graph break in forward_features>
x = self.blocks(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/efficientnet_blocks.py", line 183, in forward
x = self.conv_dw(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/conv2d_same.py", line 30, in forward
return conv2d_same(x, self.weight, self.bias, self.stride, self.padding, self.dilation, self.groups)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/conv2d_same.py", line 13, in conv2d_same
def conv2d_same(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1065, in compiled_function
outs = CompiledFunction.apply(*no_dupe_args_with_synthetic_bases)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 977, in forward
fw_outs = call_func_with_args(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 595, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 158, in run_node
check(nv, rv, lambda: f"output {i}")
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 126, in check
assert nv.size() == rv.size(), f"{desc()}: {nv.size()} != {rv.size()}"
AssertionError: output 0: torch.Size([96, 1, 3, 3]) != torch.Size([144, 1, 5, 5])
While executing %primals_1 : [#users=0] = placeholder[target=primals_1]
TorchDynamo optimized model failed to run because of following error
cuda train tf_efficientnet_b0 FAIL
Running timm_models.py tf_mixnet_l...
ERROR:common:Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 305, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 308, in <graph break in forward_and_backward_pass>
pred = mod(*cloned_inputs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/efficientnet.py", line 557, in forward
x = self.forward_features(x)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/efficientnet.py", line 540, in forward_features
x = self.conv_stem(x)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/efficientnet.py", line 545, in <graph break in forward_features>
x = self.blocks(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/container.py", line 204, in forward
input = module(input)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/efficientnet_blocks.py", line 181, in forward
x = self.conv_pw(x)
File "/data/users/ezyang/b/pytorch/torch/nn/modules/module.py", line 1427, in _call_impl
return forward_call(*input, **kwargs)
File "/home/ezyang/local/b/pytorch-env/lib/python3.9/site-packages/timm/models/layers/mixed_conv2d.py", line 47, in forward
def forward(self, x):
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1537, in forward
return compiled_function(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1507, in compiled_function
return aot_dispatcher_function(args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1125, in compiled_function
fw_outs_including_aliases.append(input_alias.as_strided(out_tensor_meta.size(), out_tensor_meta.stride(), out_tensor_meta.storage_offset()))
RuntimeError: Expected !is_symbolic() to be true, but got false. (Could this error message be improved? If so, please report an enhancement request to PyTorch.)
TorchDynamo optimized model failed to run because of following error
cuda train tf_mixnet_l FAIL
Running timm_models.py tinynet_a...
cuda train tinynet_a PASS
Running timm_models.py tnt_s_patch16_224...
cuda train tnt_s_patch16_224 PASS
Running timm_models.py twins_pcpvt_base...
cuda train twins_pcpvt_base FAIL (TIMEOUT)
Running timm_models.py visformer_small...
cuda train visformer_small PASS
Running timm_models.py vit_base_patch16_224...
cuda train vit_base_patch16_224 PASS
Running timm_models.py volo_d1_224...
cuda train volo_d1_224 PASS
Running timm_models.py xcit_large_24_p8_224...
WARNING:common:fp64 golden ref were not generated for xcit_large_24_p8_224
ERROR:common:output 0: (602112, 1, 21504, 768) != (602112, 784, 28, 1) (mismatch at index 1)
While executing %native_batch_norm_backward : [#users=3] = call_function[target=torch.ops.aten.native_batch_norm_backward.default](args = (%view_1 : Tensor[size=[2, 768, 28, 28], stride=[602112, 1, 21504, 768]], %convolution_2 : Tensor[size=[2, 768, 28, 28], stride=[602112, 784, 28, 1]], %primals_8 : Tensor[size=[768], stride=[1]], %primals_16 : Tensor[size=[768], stride=[1]], %primals_17 : Tensor[size=[768], stride=[1]], %getitem_7 : Tensor[size=[768], stride=[1]], %getitem_8 : Tensor[size=[768], stride=[1]], False, 1e-05, [True, True, True]), kwargs = {})
Traceback (most recent call last):
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1157, in check_accuracy
new_result = optimized_model_iter_fn(
File "/data/users/ezyang/b/pytorch/torch/_dynamo/eval_frame.py", line 174, in _fn
return fn(*args, **kwargs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/common.py", line 1055, in run_n_iterations
self.model_iter_fn(mod, inputs, collect_outputs=False)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 305, in forward_and_backward_pass
cloned_inputs = clone_inputs(inputs)
File "/data/users/ezyang/b/pytorch/benchmarks/dynamo/timm_models.py", line 312, in <graph break in forward_and_backward_pass>
self.grad_scaler.scale(loss).backward()
File "/data/users/ezyang/b/pytorch/torch/_tensor.py", line 473, in backward
torch.autograd.backward(
File "/data/users/ezyang/b/pytorch/torch/autograd/__init__.py", line 197, in backward
Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
File "/data/users/ezyang/b/pytorch/torch/autograd/function.py", line 270, in apply
return user_fn(self, *args)
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 1037, in backward
out = call_func_with_args(
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 595, in call_func_with_args
out = normalize_as_list(f(args))
File "/data/users/ezyang/b/pytorch/functorch/_src/aot_autograd.py", line 570, in g
return f(*args)
File "/data/users/ezyang/b/pytorch/torch/fx/interpreter.py", line 130, in run
self.env[node] = self.run_node(node)
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 158, in run_node
check(nv, rv, lambda: f"output {i}")
File "/data/users/ezyang/b/pytorch/functorch/_src/compilers.py", line 129, in check
assert same_strides, f"{desc()}: {nv.stride()} != {rv.stride()} (mismatch at index {idx})"
AssertionError: output 0: (602112, 1, 21504, 768) != (602112, 784, 28, 1) (mismatch at index 1)
While executing %native_batch_norm_backward : [#users=3] = call_function[target=torch.ops.aten.native_batch_norm_backward.default](args = (%view_1 : Tensor[size=[2, 768, 28, 28], stride=[602112, 1, 21504, 768]], %convolution_2 : Tensor[size=[2, 768, 28, 28], stride=[602112, 784, 28, 1]], %primals_8 : Tensor[size=[768], stride=[1]], %primals_16 : Tensor[size=[768], stride=[1]], %primals_17 : Tensor[size=[768], stride=[1]], %getitem_7 : Tensor[size=[768], stride=[1]], %getitem_8 : Tensor[size=[768], stride=[1]], False, 1e-05, [True, True, True]), kwargs = {})
TorchDynamo optimized model failed to run because of following error
cuda train xcit_large_24_p8_224 FAIL
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment